wejoncy commited on
Commit
fc13c9a
1 Parent(s): f67154a

update config

Browse files
Files changed (1) hide show
  1. config.json +13 -12
config.json CHANGED
@@ -17,7 +17,15 @@
17
  "num_hidden_layers": 80,
18
  "num_key_value_heads": 8,
19
  "pretraining_tp": 1,
20
- "quant_config": {
 
 
 
 
 
 
 
 
21
  "model.layers.0.mlp.down_proj": {
22
  "bias": false,
23
  "enable_norm": true,
@@ -13457,14 +13465,7 @@
13457
  -1,
13458
  6
13459
  ]
13460
- }
13461
- },
13462
- "rms_norm_eps": 1e-05,
13463
- "rope_scaling": null,
13464
- "rope_theta": 10000.0,
13465
- "tie_word_embeddings": false,
13466
- "torch_dtype": "float16",
13467
- "transformers_version": "4.37.2",
13468
- "use_cache": true,
13469
- "vocab_size": 32000
13470
- }
 
17
  "num_hidden_layers": 80,
18
  "num_key_value_heads": 8,
19
  "pretraining_tp": 1,
20
+ "rms_norm_eps": 1e-05,
21
+ "rope_scaling": null,
22
+ "rope_theta": 10000.0,
23
+ "tie_word_embeddings": false,
24
+ "torch_dtype": "float16",
25
+ "transformers_version": "4.37.2",
26
+ "use_cache": true,
27
+ "vocab_size": 32000,
28
+ "quantization_config": {
29
  "model.layers.0.mlp.down_proj": {
30
  "bias": false,
31
  "enable_norm": true,
 
13465
  -1,
13466
  6
13467
  ]
13468
+ },
13469
+ "quant_method": "vptq"
13470
+ }
13471
+ }