smolLM-arena / load_models.py
as-cle-bert's picture
Update load_models.py
f869026 verified
raw
history blame contribute delete
No virus
675 Bytes
from transformers import AutoModelForCausalLM, AutoTokenizer
# Model checkpoints
models_checkpoints = [
"HuggingFaceTB/SmolLM-135M-Instruct",
"HuggingFaceTB/SmolLM-360M",
"HuggingFaceTB/SmolLM-360M-Instruct",
"HuggingFaceTB/SmolLM-1.7B",
"HuggingFaceTB/SmolLM-1.7B-Instruct",
"Qwen/Qwen2-1.5B-Instruct",
"Qwen/Qwen2-0.5B"
]
# Dictionary to store models and tokenizers
models_and_tokenizers = {}
# Loading models and tokenizers
for checkpoint in models_checkpoints:
model = AutoModelForCausalLM.from_pretrained(checkpoint)
tokenizer = AutoTokenizer.from_pretrained(checkpoint)
models_and_tokenizers[checkpoint] = (model, tokenizer)