python main. py Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained. model. safetensors.index.json: 100%| | 13.5k/13.5k [00:00‹?, PB/s] model-00001-of-00002. safetensors: 100% | 4.95G/4.95G [07:27<00:00, 11. 1MB/s] model-00002-of-00002. safetensors: 100% 67. 1M/67.1M [00:05<00:00, 11.5MB/s] Downloading shards: 100% || | 2/2 [07:35‹00:00, 227.61s/it] Gemma's activation function should be approximate GeLU and not exact GeLU. Changing the activation function to 'gelu_pytorch_tanh.if you want to use the legacy "gelu', edit the "model.config to set hidden_activation=gelu* instead of todden act instead of hidden_act. See https://github.com/huggingface/transformers/pull/29402 for more details. Loading checkpoint shards: 100%| | 2/2 [00:03<00:00, 1.87s/itl generation_config json: 100%|| 137/137[00:00<?」3B/s] nexa model result: a pouto using the specified caea and resolutiou stones iption: rame rs a photo (cama a):) Captures - camera (str): Specifies the camera to use. Can be \'front\' or \'back\'. The default is \'back\'. \n\n Returns: \n - str: The string contains the file 2624 t 12 4a. Photo if nees at ay 96 83662387968t, ample: /storage/emulated/o/Pictures/NAPP/3N 123456.Jpg\'\n latency: 367.85967230796814