aekpic877 commited on
Commit
57c7d18
1 Parent(s): 4514961

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +15 -10
app.py CHANGED
@@ -1,10 +1,15 @@
1
- from transformers import AutoTokenizer, AutoModelForCausalLM
2
- tokenizer = AutoTokenizer.from_pretrained("deepseek-ai/deepseek-coder-1.3b-instruct", trust_remote_code=True)
3
- model = AutoModelForCausalLM.from_pretrained("deepseek-ai/deepseek-coder-1.3b-instruct", trust_remote_code=True, torch_dtype=torch.bfloat16).cuda()
4
- messages=[
5
- { 'role': 'user', 'content': "write a quick sort algorithm in python."}
6
- ]
7
- inputs = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to(model.device)
8
- # tokenizer.eos_token_id is the id of <|EOT|> token
9
- outputs = model.generate(inputs, max_new_tokens=512, do_sample=False, top_k=50, top_p=0.95, num_return_sequences=1, eos_token_id=tokenizer.eos_token_id)
10
- print(tokenizer.decode(outputs[0][len(inputs[0]):], skip_special_tokens=True))
 
 
 
 
 
 
1
+ import torch
2
+ from transformers import AutoTokenizer, AutoModelForCausalLM, GenerationConfig
3
+
4
+ model_name = "deepseek-ai/deepseek-math-7b-base"
5
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
6
+ model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16, device_map="auto")
7
+ model.generation_config = GenerationConfig.from_pretrained(model_name)
8
+ model.generation_config.pad_token_id = model.generation_config.eos_token_id
9
+
10
+ text = "The integral of x^2 from 0 to 2 is"
11
+ inputs = tokenizer(text, return_tensors="pt")
12
+ outputs = model.generate(**inputs.to(model.device), max_new_tokens=100)
13
+
14
+ result = tokenizer.decode(outputs[0], skip_special_tokens=True)
15
+ print(result)