Xu-Ouyang commited on
Commit
7a365e8
·
verified ·
1 Parent(s): ae12256

Upload GPTNeoXForCausalLM

Browse files
config.json CHANGED
@@ -19,22 +19,31 @@
19
  "num_attention_heads": 40,
20
  "num_hidden_layers": 36,
21
  "quantization_config": {
 
22
  "bits": 8,
 
23
  "damp_percent": 0.1,
24
  "dataset": "wikitext2",
25
  "desc_act": false,
 
26
  "group_size": 128,
27
- "modules_in_block_to_quantize": null,
 
 
 
 
 
28
  "quant_method": "gptq",
29
  "sym": true,
30
- "true_sequential": true
 
31
  },
32
  "rope_scaling": null,
33
  "rotary_emb_base": 10000,
34
  "rotary_pct": 0.25,
35
  "tie_word_embeddings": false,
36
  "torch_dtype": "float16",
37
- "transformers_version": "4.41.2",
38
  "use_cache": true,
39
  "use_parallel_residual": true,
40
  "vocab_size": 50688
 
19
  "num_attention_heads": 40,
20
  "num_hidden_layers": 36,
21
  "quantization_config": {
22
+ "batch_size": 1,
23
  "bits": 8,
24
+ "block_name_to_quantize": "gpt_neox.layers",
25
  "damp_percent": 0.1,
26
  "dataset": "wikitext2",
27
  "desc_act": false,
28
+ "disable_exllama": false,
29
  "group_size": 128,
30
+ "model_seqlen": 4096,
31
+ "module_name_preceding_first_block": [
32
+ "gpt_neox.embed_in",
33
+ "gpt_neox.emb_dropout"
34
+ ],
35
+ "pad_token_id": null,
36
  "quant_method": "gptq",
37
  "sym": true,
38
+ "true_sequential": true,
39
+ "use_cuda_fp16": true
40
  },
41
  "rope_scaling": null,
42
  "rotary_emb_base": 10000,
43
  "rotary_pct": 0.25,
44
  "tie_word_embeddings": false,
45
  "torch_dtype": "float16",
46
+ "transformers_version": "4.44.0",
47
  "use_cache": true,
48
  "use_parallel_residual": true,
49
  "vocab_size": 50688
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
  "eos_token_id": 0,
5
- "transformers_version": "4.41.2"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
  "eos_token_id": 0,
5
+ "transformers_version": "4.44.0"
6
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa7470885ffa9dbbb9a01fa51c1ae1457765ced161802566bc9895b2384c53ed
3
  size 4922816920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70c354bc14f4761970d18ccd01b5ec920c752a0db1c0e5a93e803dafd872c86d
3
  size 4922816920
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84244c9feeeecc4f3d1fb600200790b7cff5b62ac05acd59925813d2e891b46c
3
  size 4967652888
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d7dee582f749f051ce22455a04c9ed1e1e5687e3993b44bda4d1bf9761778db
3
  size 4967652888
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:321265c3103abff9935c8797f90f9a5c16914510ff84c0390d177d8bfa21f482
3
  size 2747738936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f84df2ee41f096c75d4a55cb4f8b5b5fb4c55aeef492d6f600f1cfa8793e3d21
3
  size 2747738936