devanand7800 commited on
Commit
cf12c23
·
1 Parent(s): c5596e4

Upload GPTNeoXForCausalLM

Browse files
Files changed (3) hide show
  1. config.json +103 -0
  2. generation_config.json +77 -0
  3. pytorch_model.bin +3 -0
config.json ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "PygmalionAI/pygmalion-1.3b",
3
+ "architectures": [
4
+ "GPTNeoXForCausalLM"
5
+ ],
6
+ "bad_words_ids": [
7
+ [
8
+ 434,
9
+ 15694,
10
+ 66,
11
+ 27,
12
+ 209
13
+ ],
14
+ [
15
+ 15362
16
+ ],
17
+ [
18
+ 1713
19
+ ],
20
+ [
21
+ 1713,
22
+ 64
23
+ ],
24
+ [
25
+ 1713,
26
+ 876
27
+ ],
28
+ [
29
+ 2016,
30
+ 251,
31
+ 857,
32
+ 75,
33
+ 9194,
34
+ 35478
35
+ ],
36
+ [
37
+ 2391
38
+ ],
39
+ [
40
+ 20340
41
+ ],
42
+ [
43
+ 33021
44
+ ],
45
+ [
46
+ 2391,
47
+ 1051
48
+ ],
49
+ [
50
+ 5638
51
+ ],
52
+ [
53
+ 2391,
54
+ 20340
55
+ ],
56
+ [
57
+ 5638,
58
+ 537
59
+ ],
60
+ [
61
+ 1559,
62
+ 2345
63
+ ],
64
+ [
65
+ 1559,
66
+ 7849
67
+ ],
68
+ [
69
+ 1559,
70
+ 17379
71
+ ],
72
+ [
73
+ 25321,
74
+ 4611
75
+ ]
76
+ ],
77
+ "bos_token_id": 0,
78
+ "classifier_dropout": 0.1,
79
+ "eos_token_id": 0,
80
+ "hidden_act": "gelu",
81
+ "hidden_size": 2048,
82
+ "initializer_range": 0.02,
83
+ "intermediate_size": 8192,
84
+ "layer_norm_eps": 1e-05,
85
+ "max_position_embeddings": 2048,
86
+ "model_type": "gpt_neox",
87
+ "num_attention_heads": 16,
88
+ "num_hidden_layers": 24,
89
+ "quantization_config": {
90
+ "llm_int8_enable_fp32_cpu_offload": false,
91
+ "llm_int8_skip_modules": null,
92
+ "llm_int8_threshold": 6.0,
93
+ "load_in_8bit": true
94
+ },
95
+ "rotary_emb_base": 10000,
96
+ "rotary_pct": 0.25,
97
+ "tie_word_embeddings": false,
98
+ "torch_dtype": "float16",
99
+ "transformers_version": "4.29.1",
100
+ "use_cache": true,
101
+ "use_parallel_residual": true,
102
+ "vocab_size": 50304
103
+ }
generation_config.json ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bad_words_ids": [
4
+ [
5
+ 434,
6
+ 15694,
7
+ 66,
8
+ 27,
9
+ 209
10
+ ],
11
+ [
12
+ 15362
13
+ ],
14
+ [
15
+ 1713
16
+ ],
17
+ [
18
+ 1713,
19
+ 64
20
+ ],
21
+ [
22
+ 1713,
23
+ 876
24
+ ],
25
+ [
26
+ 2016,
27
+ 251,
28
+ 857,
29
+ 75,
30
+ 9194,
31
+ 35478
32
+ ],
33
+ [
34
+ 2391
35
+ ],
36
+ [
37
+ 20340
38
+ ],
39
+ [
40
+ 33021
41
+ ],
42
+ [
43
+ 2391,
44
+ 1051
45
+ ],
46
+ [
47
+ 5638
48
+ ],
49
+ [
50
+ 2391,
51
+ 20340
52
+ ],
53
+ [
54
+ 5638,
55
+ 537
56
+ ],
57
+ [
58
+ 1559,
59
+ 2345
60
+ ],
61
+ [
62
+ 1559,
63
+ 7849
64
+ ],
65
+ [
66
+ 1559,
67
+ 17379
68
+ ],
69
+ [
70
+ 25321,
71
+ 4611
72
+ ]
73
+ ],
74
+ "bos_token_id": 0,
75
+ "eos_token_id": 0,
76
+ "transformers_version": "4.29.1"
77
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69c88d37407ee62f722f755532272d340205c8b11c7f214dacc027931d412355
3
+ size 1723924029