Qusaiiii commited on
Commit
a02bc5a
·
verified ·
1 Parent(s): c0ac419

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,55 +1,39 @@
1
  {
2
- "_name_or_path": "EleutherAI/gpt-neo-125M",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
- "GPTNeoForCausalLM"
6
- ],
7
- "attention_dropout": 0,
8
- "attention_layers": [
9
- "global",
10
- "local",
11
- "global",
12
- "local",
13
- "global",
14
- "local",
15
- "global",
16
- "local",
17
- "global",
18
- "local",
19
- "global",
20
- "local"
21
- ],
22
- "attention_types": [
23
- [
24
- [
25
- "global",
26
- "local"
27
- ],
28
- 6
29
- ]
30
  ],
 
31
  "bos_token_id": 50256,
32
- "classifier_dropout": 0.1,
33
- "embed_dropout": 0,
34
  "eos_token_id": 50256,
35
- "gradient_checkpointing": false,
36
- "hidden_size": 768,
37
  "initializer_range": 0.02,
38
- "intermediate_size": null,
39
  "layer_norm_epsilon": 1e-05,
40
- "max_position_embeddings": 2048,
41
- "model_type": "gpt_neo",
42
- "num_heads": 12,
43
- "num_layers": 12,
44
- "resid_dropout": 0,
 
 
 
 
 
 
45
  "summary_activation": null,
46
  "summary_first_dropout": 0.1,
47
  "summary_proj_to_labels": true,
48
  "summary_type": "cls_index",
49
  "summary_use_proj": true,
 
 
 
 
 
 
50
  "torch_dtype": "float32",
51
  "transformers_version": "4.46.3",
52
  "use_cache": true,
53
- "vocab_size": 50257,
54
- "window_size": 256
55
  }
 
1
  {
2
+ "_name_or_path": "gpt2",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
+ "GPT2LMHeadModel"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  ],
7
+ "attn_pdrop": 0.1,
8
  "bos_token_id": 50256,
9
+ "embd_pdrop": 0.1,
 
10
  "eos_token_id": 50256,
 
 
11
  "initializer_range": 0.02,
 
12
  "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
+ "n_embd": 768,
16
+ "n_head": 12,
17
+ "n_inner": null,
18
+ "n_layer": 12,
19
+ "n_positions": 1024,
20
+ "reorder_and_upcast_attn": false,
21
+ "resid_pdrop": 0.1,
22
+ "scale_attn_by_inverse_layer_idx": false,
23
+ "scale_attn_weights": true,
24
  "summary_activation": null,
25
  "summary_first_dropout": 0.1,
26
  "summary_proj_to_labels": true,
27
  "summary_type": "cls_index",
28
  "summary_use_proj": true,
29
+ "task_specific_params": {
30
+ "text-generation": {
31
+ "do_sample": true,
32
+ "max_length": 50
33
+ }
34
+ },
35
  "torch_dtype": "float32",
36
  "transformers_version": "4.46.3",
37
  "use_cache": true,
38
+ "vocab_size": 50257
 
39
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:27c049d05b3c1c6bddb1fa2657e01fa4ada43be88573ef660c2503e8d6c81c43
3
- size 500811336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31cbf7d7a629bfaed724bfd6ec69dbfa161e3dc8a7b27362bf866976889b20ef
3
+ size 497774208
special_tokens_map.json CHANGED
@@ -1,24 +1,6 @@
1
  {
2
- "bos_token": {
3
- "content": "<|endoftext|>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "<|endoftext|>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
  "pad_token": "<|endoftext|>",
17
- "unk_token": {
18
- "content": "<|endoftext|>",
19
- "lstrip": false,
20
- "normalized": true,
21
- "rstrip": false,
22
- "single_word": false
23
- }
24
  }
 
1
  {
2
+ "bos_token": "<|endoftext|>",
3
+ "eos_token": "<|endoftext|>",
 
 
 
 
 
 
 
 
 
 
 
 
4
  "pad_token": "<|endoftext|>",
5
+ "unk_token": "<|endoftext|>"
 
 
 
 
 
 
6
  }
tokenizer_config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "add_bos_token": false,
3
  "add_prefix_space": false,
4
  "added_tokens_decoder": {
5
  "50256": {
@@ -12,10 +11,9 @@
12
  }
13
  },
14
  "bos_token": "<|endoftext|>",
15
- "clean_up_tokenization_spaces": true,
16
  "eos_token": "<|endoftext|>",
17
- "errors": "replace",
18
- "model_max_length": 2048,
19
  "pad_token": "<|endoftext|>",
20
  "tokenizer_class": "GPT2Tokenizer",
21
  "unk_token": "<|endoftext|>"
 
1
  {
 
2
  "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
  "50256": {
 
11
  }
12
  },
13
  "bos_token": "<|endoftext|>",
14
+ "clean_up_tokenization_spaces": false,
15
  "eos_token": "<|endoftext|>",
16
+ "model_max_length": 1024,
 
17
  "pad_token": "<|endoftext|>",
18
  "tokenizer_class": "GPT2Tokenizer",
19
  "unk_token": "<|endoftext|>"
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f4c666ebe9ca6ef8728385a3d3333c19c27e0a5ded14defdd850b5af2faae7ee
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:813062ba8003f1b08fe674315d9f8b4a9929c631d53749d8e4d1bc9b3599ca40
3
  size 5240