sigmareaver commited on
Commit
bea966a
·
1 Parent(s): 3d290ac

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +38 -0
README.md CHANGED
@@ -1,3 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
 
1
+ # flan-ul2 4-bit 128-groupsize GPTQ
2
+ Quantized using qwopqwop200's GPTQ-for-Llama repo on the t5 branch.
3
+
4
+ Quantization command:
5
+
6
+ `PYTORCH_CUDA_ALLOC_CONF=max_split_size_mb:512 python t5.py ../full-models/flan-ul2 wikitext2 --nsamples 256 --wbits 4 --act-order --groupsize 128 --save ../gptq-models/flan-ul2-gptq/flan-ul2-4bit-128g-gptq.pt`
7
+
8
+ Benchmark command:
9
+
10
+ `python t5.py ../full-models/flan-ul2 wikitext2 --load ../gptq-models/flan-ul2-gptq/flan-ul2-4bit-128g-gptq2.pt --wbits 4 --groupsize 128 --benchmark --benchmark_mode mmlu`
11
+
12
+ Results :
13
+ ```
14
+ Average accuracy 0.289 - math
15
+ Average accuracy 0.562 - health
16
+ Average accuracy 0.416 - physics
17
+ Average accuracy 0.780 - business
18
+ Average accuracy 0.610 - biology
19
+ Average accuracy 0.446 - chemistry
20
+ Average accuracy 0.461 - computer science
21
+ Average accuracy 0.513 - economics
22
+ Average accuracy 0.538 - engineering
23
+ Average accuracy 0.455 - philosophy
24
+ Average accuracy 0.622 - other
25
+ Average accuracy 0.703 - history
26
+ Average accuracy 0.707 - geography
27
+ Average accuracy 0.718 - politics
28
+ Average accuracy 0.653 - psychology
29
+ Average accuracy 0.711 - culture
30
+ Average accuracy 0.447 - law
31
+ Average accuracy 0.416 - STEM
32
+ Average accuracy 0.501 - humanities
33
+ Average accuracy 0.643 - social sciences
34
+ Average accuracy 0.613 - other (business, health, misc.)
35
+ MMLU Average accuracy: 0.540
36
+
37
+ ```
38
+
39
  ---
40
  license: apache-2.0
41
  ---