fbaldassarri
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -49,21 +49,16 @@ Note: this INT8 version of Llama-3.2-1B has been quantized to run inference thro
|
|
49 |
I suggest to install requirements into a dedicated python-virtualenv or a conda enviroment.
|
50 |
|
51 |
```
|
52 |
-
|
|
|
|
|
|
|
53 |
```
|
54 |
|
55 |
-
|
56 |
-
- auto_gptq==0.7.1
|
57 |
-
- neural_compressor==3.1
|
58 |
-
- torch==2.3.0+cpu
|
59 |
-
- torchaudio==2.5.0+cpu
|
60 |
-
- torchvision==0.18.0+cpu
|
61 |
-
- transformers==4.45.2
|
62 |
-
|
63 |
-
### Step 2 Build Intel Autoround wheel from sources
|
64 |
|
65 |
```
|
66 |
-
|
67 |
```
|
68 |
|
69 |
### Step 3 Script for Quantization
|
|
|
49 |
I suggest to install requirements into a dedicated python-virtualenv or a conda enviroment.
|
50 |
|
51 |
```
|
52 |
+
wget https://github.com/intel/auto-round/archive/refs/tags/v0.4.3.tar.gz
|
53 |
+
tar -xvzf v0.4.3.tar.gz
|
54 |
+
cd auto-round-0.4.3
|
55 |
+
pip install -r requirements-cpu.txt --upgrade
|
56 |
```
|
57 |
|
58 |
+
### Step 2 Build Intel AutoRound wheel from sources
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
59 |
|
60 |
```
|
61 |
+
pip install -vvv --no-build-isolation -e .[cpu]
|
62 |
```
|
63 |
|
64 |
### Step 3 Script for Quantization
|