wenhuach commited on
Commit
cf5a1db
·
1 Parent(s): 36f5657

align with gptq format

Browse files

Signed-off-by: wenhuach <[email protected]>

Files changed (2) hide show
  1. config.json +3 -1
  2. quantization_config.json +5 -3
config.json CHANGED
@@ -40,7 +40,6 @@
40
  "quantization_config": {
41
  "amp": true,
42
  "autoround_version": "0.4.5",
43
- "backend": "auto_round:gptq:exllamav2",
44
  "batch_size": 4,
45
  "bits": 4,
46
  "data_type": "int",
@@ -55,6 +54,9 @@
55
  "minmax_lr": 0.005,
56
  "nsamples": 512,
57
  "quant_method": "gptq",
 
 
 
58
  "scale_dtype": "torch.float16",
59
  "seqlen": 2048,
60
  "sym": true,
 
40
  "quantization_config": {
41
  "amp": true,
42
  "autoround_version": "0.4.5",
 
43
  "batch_size": 4,
44
  "bits": 4,
45
  "data_type": "int",
 
54
  "minmax_lr": 0.005,
55
  "nsamples": 512,
56
  "quant_method": "gptq",
57
+ "desc_act": false,
58
+ "true_sequential": false,
59
+ "damp_percent": 0.01,
60
  "scale_dtype": "torch.float16",
61
  "seqlen": 2048,
62
  "sym": true,
quantization_config.json CHANGED
@@ -19,6 +19,8 @@
19
  "enable_norm_bias_tuning": false,
20
  "dataset": "NeelNanda/pile-10k",
21
  "autoround_version": "0.4.5",
22
- "quant_method": "intel/auto-round",
23
- "backend": "auto_round:gptq:exllamav2"
24
- }
 
 
 
19
  "enable_norm_bias_tuning": false,
20
  "dataset": "NeelNanda/pile-10k",
21
  "autoround_version": "0.4.5",
22
+ "quant_method": "gptq",
23
+ "desc_act": false,
24
+ "true_sequential": false,
25
+ "damp_percent": 0.01
26
+ }