Falcon3
Collection
Falcon3 family of Open Foundation Models is a set of pretrained and instruct LLMs ranging from 1B to 10B parameters.
•
40 items
•
Updated
•
70
The model has been trained following the training strategies from the recent 1-bit LLM HF blogpost and 1-bit LLM paper. For more details about the training protocol of this model, please refer to the Falcon-3 technical report, section Compression.
Currently to use this model you can on BitNet library. You can also play with the model using the falcon-1.58bit playground (only for the 7B instruct version).
git clone https://github.com/microsoft/BitNet && cd BitNet
pip install -r requirements.txt
huggingface-cli download tiiuae/Falcon3-7B-Instruct-1.58bit-GGUF ggml-model-i2_s.gguf --local-dir models/Falcon3-7B-1.58bit/
python run_inference.py -m models/Falcon3-7B-1.58bit/ggml-model-i2_s.gguf -p "You are a helpful assistant" -cnv
We report in the following table our internal pipeline benchmarks:
Note evaluation results are normalized score from v2 leaderboard tasks - reported results of original models in the blogpost are raw scores
Benchmark | Llama3-8B-1.58-100B-tokens | Falcon3-7B-Instruct-1.58bit |
---|---|---|
IFEval | 17.91 | 59.24 |
MUSR | 4.87 | 1.76 |
GPQA | 1.83 | 5.25 |
BBH | 5.36 | 8.54 |
MMLU-PRO | 2.78 | 8.44 |
MATH | 0.26 | 2.93 |
Average | 5.5 | 14.36 |
If the Falcon3 family of models were helpful to your work, feel free to give us a cite.
@misc{Falcon3,
title = {The Falcon 3 Family of Open Models},
author = {Falcon-LLM Team},
month = {December},
year = {2024}
}