Falcon Mamba Playground
Generate chat responses using FalconMamba-7b model
This collection features the FalconMamba 7B base model, the instruction-tuned version, their 4-bit and GGUF variants, and the demo.
Generate chat responses using FalconMamba-7b model
Note FalconMamba technical report
Note First strong attention free model for general purpose usage, based on Mamba1 architecture
Note FalconMamba-7B fine-tuned on instruction data, for chat-like interaction with the model
Note FalconMamba-7B quantized in 4bit precision using `bitsandbytes` library for lighter memory requirements and smaller GPU hardwares
Note FalconMamba-7B-instruct quantized in 4bit precision using `bitsandbytes` library for lighter memory requirements and smaller GPU hardwares
Note Falcon Mamba 7b-instruct in GGUF format (compatible with llama.cpp) in BF16 format
Note Falcon Mamba 7b-instruct in GGUF format (compatible with llama.cpp) in F16 format
Note Falcon Mamba 7b-instruct in GGUF format (compatible with llama.cpp) in quantized Q8_0 format
Note Falcon Mamba 7b-instruct in GGUF format (compatible with llama.cpp) in quantized Q4_K_M format
Note Falcon Mamba 7b in GGUF format (compatible with llama.cpp) in BF16 format
Note Falcon Mamba 7b in GGUF format (compatible with llama.cpp) in F16 format
Note Falcon Mamba 7b in GGUF format (compatible with llama.cpp) in quantized Q8_0 format
Note Falcon Mamba 7b in GGUF format (compatible with llama.cpp) in quantized Q4_K_M format
Note Pre-decay stage checkpoint useful for continuous pretraining