qinglinf commited on
Commit
d52f373
1 Parent(s): 363ecdc

Upload 5 files

Browse files
Files changed (5) hide show
  1. config.json +2 -5
  2. merges.txt +0 -0
  3. tokenizer.json +0 -0
  4. tokenizer_config.json +1 -0
  5. vocab.json +0 -0
config.json CHANGED
@@ -1,11 +1,10 @@
1
  {
2
- "_name_or_path": "shahrukhx01/roberta-base-boolq",
3
  "architectures": [
4
  "RobertaForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
8
- "classifier_dropout": null,
9
  "eos_token_id": 2,
10
  "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
@@ -14,7 +13,6 @@
14
  "initializer_range": 0.02,
15
  "intermediate_size": 3072,
16
  "layer_norm_eps": 1e-05,
17
- "max_length": 512,
18
  "max_position_embeddings": 514,
19
  "model_type": "roberta",
20
  "num_attention_heads": 12,
@@ -22,8 +20,7 @@
22
  "pad_token_id": 1,
23
  "position_embedding_type": "absolute",
24
  "problem_type": "single_label_classification",
25
- "torch_dtype": "float32",
26
- "transformers_version": "4.27.3",
27
  "type_vocab_size": 1,
28
  "use_cache": true,
29
  "vocab_size": 50265
 
1
  {
2
+ "_name_or_path": "roberta-base",
3
  "architectures": [
4
  "RobertaForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
 
8
  "eos_token_id": 2,
9
  "gradient_checkpointing": false,
10
  "hidden_act": "gelu",
 
13
  "initializer_range": 0.02,
14
  "intermediate_size": 3072,
15
  "layer_norm_eps": 1e-05,
 
16
  "max_position_embeddings": 514,
17
  "model_type": "roberta",
18
  "num_attention_heads": 12,
 
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
  "problem_type": "single_label_classification",
23
+ "transformers_version": "4.6.1",
 
24
  "type_vocab_size": 1,
25
  "use_cache": true,
26
  "vocab_size": 50265
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "roberta-base"}
vocab.json ADDED
The diff for this file is too large to render. See raw diff