Eugene Cheah (picocreator)
commited on
Commit
•
194fd29
1
Parent(s):
ec429f3
Changed to v5 Eagle 7B model
Browse files- config.json +3 -2
- pytorch_model.bin +2 -2
- tokenizer_config.json +5 -1
config.json
CHANGED
@@ -6,15 +6,16 @@
|
|
6 |
"AutoConfig": "configuration_rwkv5.Rwkv5Config",
|
7 |
"AutoModelForCausalLM": "modeling_rwkv5.Rwkv5ForCausalLM"
|
8 |
},
|
9 |
-
"attention_hidden_size":
|
10 |
"bos_token_id": 0,
|
11 |
"context_length": 4096,
|
12 |
"eos_token_id": 0,
|
13 |
"head_size": 64,
|
14 |
-
"hidden_size":
|
15 |
"intermediate_size": null,
|
16 |
"layer_norm_epsilon": 1e-05,
|
17 |
"model_type": "rwkv5",
|
|
|
18 |
"num_hidden_layers": 32,
|
19 |
"rescale_every": 6,
|
20 |
"tie_word_embeddings": false,
|
|
|
6 |
"AutoConfig": "configuration_rwkv5.Rwkv5Config",
|
7 |
"AutoModelForCausalLM": "modeling_rwkv5.Rwkv5ForCausalLM"
|
8 |
},
|
9 |
+
"attention_hidden_size": 4096,
|
10 |
"bos_token_id": 0,
|
11 |
"context_length": 4096,
|
12 |
"eos_token_id": 0,
|
13 |
"head_size": 64,
|
14 |
+
"hidden_size": 4096,
|
15 |
"intermediate_size": null,
|
16 |
"layer_norm_epsilon": 1e-05,
|
17 |
"model_type": "rwkv5",
|
18 |
+
"model_version": "5_2",
|
19 |
"num_hidden_layers": 32,
|
20 |
"rescale_every": 6,
|
21 |
"tie_word_embeddings": false,
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a0739fe6d2d69945f4780c020f6d6649f4183131e1b18c78add00d2d06ce293e
|
3 |
+
size 15036174346
|
tokenizer_config.json
CHANGED
@@ -8,5 +8,9 @@
|
|
8 |
"tokenization_rwkv_world.RWKVWorldTokenizer",
|
9 |
null
|
10 |
]
|
11 |
-
}
|
|
|
|
|
|
|
|
|
12 |
}
|
|
|
8 |
"tokenization_rwkv_world.RWKVWorldTokenizer",
|
9 |
null
|
10 |
]
|
11 |
+
},
|
12 |
+
|
13 |
+
"clean_up_tokenization_spaces": true,
|
14 |
+
"errors": "replace",
|
15 |
+
"model_max_length": 1000000000000000019884624838656
|
16 |
}
|