Distill-Qwen-1.5B-GGUF-Transformers / transformers_gguf.py
Davidqian123's picture
Upload transformers_gguf.py
53d2d48 verified
raw
history blame contribute delete
639 Bytes
from transformers import AutoTokenizer, AutoModelForCausalLM
model_id = "lmstudio-community/DeepSeek-R1-Distill-Qwen-1.5B-GGUF"
filename = "DeepSeek-R1-Distill-Qwen-1.5B-Q4_K_M.gguf"
tokenizer = AutoTokenizer.from_pretrained(model_id, gguf_file=filename)
model = AutoModelForCausalLM.from_pretrained(model_id, gguf_file=filename)
model.save_pretrained('./converted_model')
tokenizer.save_pretrained('./converted_model')
# text = "What is 1+2+3+4+5+6+7+8+9+10?"
# inputs = tokenizer(text, return_tensors="pt")
# outputs = model.generate(**inputs, max_new_tokens=512)
# print(tokenizer.decode(outputs[0], skip_special_tokens=False))