metadata
license: apache-2.0
base_model: v2ray/Mixtral-8x22B-v0.1
inference: false
model_creator: MaziyarPanahi
model_name: Mixtral-8x22B-v0.1-GGUF
pipeline_tag: text-generation
quantized_by: MaziyarPanahi
tags:
- quantized
- 2-bit
- 3-bit
- 4-bit
- 5-bit
- 6-bit
- 8-bit
- 16-bit
- GGUF
- mixtral
- moe
Mixtral-8x22B-v0.1-GGUF
in progress ...
Load sharded model
llama_load_model_from_file
will detect the number of files and will load additional tensors from the rest of files.
main --model Mixtral-8x22B-v0.1.fp16-00001-of-00005.gguf -ngl 64