ai-agi commited on
Commit
92ce212
1 Parent(s): a8f92fd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -6
README.md CHANGED
@@ -31,15 +31,17 @@ You can find more details in the [technical report](https://arxiv.org/abs/2310.1
31
 
32
 
33
  ## Use in Transformers
34
- ## Load model directly
35
  import torch \
36
- from transformers import AutoTokenizer, AutoModelForCausalLM, MistralForCausalLM
 
37
 
38
  model = MistralForCausalLM.from_pretrained("ai-agi/neural-zephyr", use_cache=False, torch_dtype=torch.bfloat16, device_map="auto") \
39
- state_dict = torch.load('model_weights.pth') \
40
- model.load_state_dict(state_dict)
 
41
 
42
  tokenizer = AutoTokenizer.from_pretrained("ai-agi/neural-zephyr", use_fast=True) \
43
  if tokenizer.pad_token is None: \
44
-     tokenizer.pad_token = tokenizer.eos_token)
45
-
 
31
 
32
 
33
  ## Use in Transformers
34
+ **Load model directly** \
35
  import torch \
36
+ from transformers import AutoTokenizer, AutoModelForCausalLM, MistralForCausalLM \
37
+ from huggingface_hub import hf_hub_download
38
 
39
  model = MistralForCausalLM.from_pretrained("ai-agi/neural-zephyr", use_cache=False, torch_dtype=torch.bfloat16, device_map="auto") \
40
+ model_weights = hf_hub_download(repo_id="ai-agi/neural-zephyr", filename="model_weights.pth") \
41
+ state_dict = torch.load(model_weights) \
42
+ model.load_state_dict(state_dict)
43
 
44
  tokenizer = AutoTokenizer.from_pretrained("ai-agi/neural-zephyr", use_fast=True) \
45
  if tokenizer.pad_token is None: \
46
+     tokenizer.pad_token = tokenizer.eos_token) \
47
+ **Manage your GPU/CPU memory for model and weights**