--- base_model: tiiuae/Falcon3-10B-Instruct tags: - fluently-lm - fluently-sets - demo - reasoning - thinking - text-generation-inference - transformers - unsloth - falcon3 - falcon - llama - trl - sft license: apache-2.0 language: - en datasets: - fluently-sets/ultrathink pipeline_tag: text-generation --- # FalconThink3-10B Demo (Finetune of Falcon3-10B-IT on Ultrathink dataset) ***Q4_K_M GGUF-quant available [here](https://huggingface.co/fluently-sets/FalconThink3-10B-IT-Q4_K_M-GGUF)*** This is SFT-finetune Falcon3-10B-IT on Ultrathink dataset. This is far from a perfect model, its main purpose is to show an example of using the dataset. - **Base model**: [tiiuae/Falcon3-10B-Instruct](https://huggingface.co/tiiuae/Falcon3-10B-Instruct) - **Model type**: [LlamaForCausalLM](https://huggingface.co/models?other=llama) - **Number of parameters**: 10.3B - **Precision**: FP16 - **Training method**: SFT - **Training dataset**: [fluently-sets/ultrathink](https://huggingface.co/datasets/fluently-sets/ultrathink) - **Languages**: English (mostly) *Trained by Fluently Team ([@ehristoforu](https://huggingface.co/ehristoforu)) with [Unsloth AI](https://github.com/unslothai/unsloth) with lovešŸ„°* [](https://github.com/unslothai/unsloth)