ybelkada commited on
Commit
0c02dd3
·
1 Parent(s): 5d08940

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -0
README.md ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ inference: false
3
+ language:
4
+ - en
5
+ library_name: transformers
6
+ license: apache-2.0
7
+ model_name: Mixtral 8X7B - bnb 4-bit
8
+ model_type: mixtral
9
+ pipeline_tag: text-generation
10
+ '
11
+ quantized_by: ybelkada
12
+ tags:
13
+ - mistral
14
+ - mixtral
15
+ ---
16
+
17
+ # Mixtral 8x7B Instruct-v0.1 - `bitsandbytes` 4-bit
18
+
19
+ This repository contains the bitsandbytes 4-bit quantized version of [`mistralai/Mixtral-8x7B-Instruct-v0.1`](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1). To use it, make sure to have the latest version of `bitsandbytes` and `transformers` installed from source:
20
+
21
+ Loading this model as such: will directly load the quantized model in 4-bit precision.
22
+ ```python
23
+ from transformers import AutoModelForCausalLM, AutoTokenizer
24
+
25
+ model_id = "ybelkada/Mixtral-8x7B-Instruct-v0.1-bnb-4bit"
26
+ model = AutoModelForCausalLM.from_pretrained(model_id)
27
+ ```
28
+ Note you need a CUDA-compatible GPU device to run low-bit precision models with `bitsandbytes`