philschmid HF staff commited on
Commit
c8263f5
1 Parent(s): 50ac599

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -53,12 +53,12 @@ from transformers import AutoTokenizer, AutoModelModelForCausalLM
53
 
54
  # load base LLM model and tokenizer
55
  model = AutoModelModelForCausalLM.from_pretrained(
56
- "philschmid/llama-7b-instruction-generator",
57
  low_cpu_mem_usage=True,
58
  torch_dtype=torch.float16,
59
  load_in_4bit=True,
60
  )
61
- tokenizer = AutoTokenizer.from_pretrained("philschmid/llama-7b-instruction-generator")
62
 
63
  prompt = f"""### Instruction:
64
  Use the Input below to create an instruction, which could have been used to generate the input using an LLM.
 
53
 
54
  # load base LLM model and tokenizer
55
  model = AutoModelModelForCausalLM.from_pretrained(
56
+ "philschmid/llama-2-7b-instruction-generator",
57
  low_cpu_mem_usage=True,
58
  torch_dtype=torch.float16,
59
  load_in_4bit=True,
60
  )
61
+ tokenizer = AutoTokenizer.from_pretrained("philschmid/llama-2-7b-instruction-generator")
62
 
63
  prompt = f"""### Instruction:
64
  Use the Input below to create an instruction, which could have been used to generate the input using an LLM.