leoh commited on
Commit
c7b13ee
·
1 Parent(s): 789796f
Files changed (1) hide show
  1. config.json +23 -5
config.json CHANGED
@@ -1,26 +1,44 @@
1
  {
2
- "_name_or_path": "leoh/roberta-base-8196",
3
  "architectures": [
4
- "RobertaModel"
5
  ],
 
6
  "attention_probs_dropout_prob": 0.1,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  "bos_token_id": 0,
8
  "classifier_dropout": null,
9
  "eos_token_id": 2,
 
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
  "hidden_size": 768,
 
13
  "initializer_range": 0.02,
14
  "intermediate_size": 3072,
15
  "layer_norm_eps": 1e-05,
16
- "max_position_embeddings": 8224,
17
- "model_type": "roberta",
18
  "num_attention_heads": 12,
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
 
22
  "torch_dtype": "float32",
23
- "transformers_version": "4.30.0.dev0",
24
  "type_vocab_size": 1,
25
  "use_cache": true,
26
  "vocab_size": 50265
 
1
  {
2
+ "_name_or_path": "allenai/longformer-base-4096",
3
  "architectures": [
4
+ "LongformerModel"
5
  ],
6
+ "attention_mode": "longformer",
7
  "attention_probs_dropout_prob": 0.1,
8
+ "attention_window": [
9
+ 512,
10
+ 512,
11
+ 512,
12
+ 512,
13
+ 512,
14
+ 512,
15
+ 512,
16
+ 512,
17
+ 512,
18
+ 512,
19
+ 512,
20
+ 512
21
+ ],
22
  "bos_token_id": 0,
23
  "classifier_dropout": null,
24
  "eos_token_id": 2,
25
+ "gradient_checkpointing": false,
26
  "hidden_act": "gelu",
27
  "hidden_dropout_prob": 0.1,
28
  "hidden_size": 768,
29
+ "ignore_attention_mask": false,
30
  "initializer_range": 0.02,
31
  "intermediate_size": 3072,
32
  "layer_norm_eps": 1e-05,
33
+ "max_position_embeddings": 16392,
34
+ "model_type": "longformer",
35
  "num_attention_heads": 12,
36
  "num_hidden_layers": 12,
37
  "pad_token_id": 1,
38
  "position_embedding_type": "absolute",
39
+ "sep_token_id": 2,
40
  "torch_dtype": "float32",
41
+ "transformers_version": "4.18.0",
42
  "type_vocab_size": 1,
43
  "use_cache": true,
44
  "vocab_size": 50265