Mixtral_ether / main.py
jeduardogruiz's picture
Upload 22 files
516a027 verified
raw
history blame contribute delete
No virus
1.3 kB
python main. py
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
model. safetensors.index.json: 100%|
| 13.5k/13.5k [00:00‹?, PB/s]
model-00001-of-00002. safetensors: 100%
| 4.95G/4.95G [07:27<00:00, 11. 1MB/s]
model-00002-of-00002. safetensors: 100%
67. 1M/67.1M [00:05<00:00, 11.5MB/s]
Downloading shards: 100% ||
| 2/2 [07:3500:00, 227.61s/it]
Gemma's activation function should be approximate GeLU and not exact GeLU. Changing the activation function to 'gelu_pytorch_tanh.if you want to use the legacy "gelu', edit the "model.config to
set hidden_activation=gelu*
instead of todden act
instead of hidden_act. See https://github.com/huggingface/transformers/pull/29402 for
more details.
Loading checkpoint shards: 100%|
| 2/2 [00:03<00:00, 1.87s/itl
generation_config json: 100%||
137/13700:00<?」3B/s]
nexa model result:
a pouto using the specified caea and resolutiou stones iption: rame rs a photo (cama a):)
Captures
- camera (str): Specifies the camera
to use. Can be \'front\' or \'back\'. The default is \'back\'. \n\n
Returns: \n
- str: The string contains the file
2624 t 12 4a.
Photo if nees at ay 96 83662387968t, ample: /storage/emulated/o/Pictures/NAPP/3N
123456.Jpg\'\n latency: 367.85967230796814