Text Generation
Transformers
Safetensors
qwen2
logic
argumentation
critical-thinking
argument-mapping
trl
sft
conversational
text-generation-inference

Model Card for Qwen2.5-Argunaut-1-1.5B-SFT

🧪 Experimental, not recommended for use in teaching.

This model is a fine-tuned version of Qwen/Qwen2.5-1.5B-Instruct. It has been trained using TRL.

📘 HF Blog Article

Quick start

from transformers import pipeline
question = "Are you familiar with Argdown syntax? What's its purpose?"
generator = pipeline("text-generation", model="DebateLabKIT/Llama-3.1-Argunaut-1-8B-SFT", device="cuda")
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
print(output["generated_text"])

Evaluation

Chat Experience

coming soon

Metrics

coming soon

SFT dataset mixture

Dataset Weight (examples) Weight (tokens)
DebateLabKIT/deepa2-conversations 25% 49%
DebateLabKIT/deep-argmap-conversations 25% 18%
allenai/tulu-3-sft-mixture 50% 33%

Training procedure

Trained with SFT on 1M examples and for 1 epoch with

  • context length 8196
  • packing (trl implementation)
# Training parameters
num_train_epochs: 1
per_device_train_batch_size: 32
gradient_accumulation_steps: 1
gradient_checkpointing: true
gradient_checkpointing_kwargs:
  use_reentrant: false
learning_rate: 5.0e-6
lr_scheduler_type: cosine
warmup_ratio: 0.1

Hardware: 4 x H100 GPUs.

This work was performed on the HoreKa supercomputer funded by the Ministry of Science, Research and the Arts Baden-Württemberg and by the Federal Ministry of Education and Research.

Framework versions

  • TRL: 0.14.0
  • Transformers: 4.46.3
  • Pytorch: 2.4.1
  • Datasets: 3.1.0
  • Tokenizers: 0.20.3

Credits

This work wouldn't be possible without all the great contributions from the open LLM community. Thank you! Special kudos go to

Downloads last month
38
Safetensors
Model size
1.54B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DebateLabKIT/Qwen2.5-Argunaut-1-1.5B-SFT

Base model

Qwen/Qwen2.5-1.5B
Finetuned
(584)
this model
Quantizations
1 model

Datasets used to train DebateLabKIT/Qwen2.5-Argunaut-1-1.5B-SFT