davda54 commited on
Commit
6344c0e
1 Parent(s): 8d33c81

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -161,7 +161,7 @@ For systems with limited VRAM, you can load the model in 8-bit or 4-bit quantiza
161
  import torch
162
  from transformers import AutoTokenizer, AutoModelForCausalLM
163
 
164
- tokenizer = AutoTokenizer.from_pretrained("norallm/normistral-11b")
165
 
166
  # Load in 8-bit mode (requires ~12GB VRAM)
167
  model = AutoModelForCausalLM.from_pretrained(
 
161
  import torch
162
  from transformers import AutoTokenizer, AutoModelForCausalLM
163
 
164
+ tokenizer = AutoTokenizer.from_pretrained("norallm/normistral-11b-warm")
165
 
166
  # Load in 8-bit mode (requires ~12GB VRAM)
167
  model = AutoModelForCausalLM.from_pretrained(