Almheiri commited on
Commit
ed14860
1 Parent(s): 5f690e5

Upload of AutoGPTQ quantized model

Browse files
config.json CHANGED
@@ -1,5 +1,6 @@
1
  {
2
- "_name_or_path": "meta-llama/Llama-3.2-1B-Instruct",
 
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -23,6 +24,19 @@
23
  "num_hidden_layers": 16,
24
  "num_key_value_heads": 8,
25
  "pretraining_tp": 1,
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  "rms_norm_eps": 1e-05,
27
  "rope_scaling": {
28
  "factor": 32.0,
@@ -33,7 +47,7 @@
33
  },
34
  "rope_theta": 500000.0,
35
  "tie_word_embeddings": true,
36
- "torch_dtype": "bfloat16",
37
  "transformers_version": "4.47.0",
38
  "use_cache": true,
39
  "vocab_size": 128256
 
1
  {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "sparse_llama/Llama-3.2-1B-Instruct-SparseGPT-INT4",
4
  "architectures": [
5
  "LlamaForCausalLM"
6
  ],
 
24
  "num_hidden_layers": 16,
25
  "num_key_value_heads": 8,
26
  "pretraining_tp": 1,
27
+ "quantization_config": {
28
+ "bits": 4,
29
+ "damp_percent": 0.1,
30
+ "desc_act": true,
31
+ "group_size": 128,
32
+ "is_marlin_format": false,
33
+ "model_file_base_name": null,
34
+ "model_name_or_path": null,
35
+ "quant_method": "gptq",
36
+ "static_groups": false,
37
+ "sym": true,
38
+ "true_sequential": true
39
+ },
40
  "rms_norm_eps": 1e-05,
41
  "rope_scaling": {
42
  "factor": 32.0,
 
47
  },
48
  "rope_theta": 500000.0,
49
  "tie_word_embeddings": true,
50
+ "torch_dtype": "float16",
51
  "transformers_version": "4.47.0",
52
  "use_cache": true,
53
  "vocab_size": 128256
gptq_model-4bit-128g.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb30078d3b42663d1da0742d8ef6ea1ae40f7a38d9de6d19a27f3a37a2edbfb1
3
+ size 1558482936
quantize_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 4,
3
+ "group_size": 128,
4
+ "damp_percent": 0.1,
5
+ "desc_act": true,
6
+ "static_groups": false,
7
+ "sym": true,
8
+ "true_sequential": true,
9
+ "model_name_or_path": null,
10
+ "model_file_base_name": null,
11
+ "is_marlin_format": false,
12
+ "quant_method": "gptq"
13
+ }