Draft GGUFs
Collection
5 items
•
Updated
•
1
This model is meant to be used as draft model for speculative decoding with mistralai/Mistral-Small-3.1-24B-Instruct-2503 or mistralai/Mistral-Small-24B-Instruct-2501
The data are Mistral's outputs and includes all kind of tasks from various datasets in English, French, German, Spanish, Italian and Portuguese. It has been trained for 2 epochs on 20k unique examples, for a total of 12 million tokens per epoch.
4-bit
5-bit
6-bit
8-bit
16-bit
Base model
Qwen/Qwen2.5-0.5B