from transformers import AutoTokenizer, AutoModelForCausalLM | |
model_id = "lmstudio-community/DeepSeek-R1-Distill-Qwen-1.5B-GGUF" | |
filename = "DeepSeek-R1-Distill-Qwen-1.5B-Q4_K_M.gguf" | |
tokenizer = AutoTokenizer.from_pretrained(model_id, gguf_file=filename) | |
model = AutoModelForCausalLM.from_pretrained(model_id, gguf_file=filename) | |
model.save_pretrained('./converted_model') | |
tokenizer.save_pretrained('./converted_model') | |
# text = "What is 1+2+3+4+5+6+7+8+9+10?" | |
# inputs = tokenizer(text, return_tensors="pt") | |
# outputs = model.generate(**inputs, max_new_tokens=512) | |
# print(tokenizer.decode(outputs[0], skip_special_tokens=False)) |