asahi417 commited on
Commit
5853909
1 Parent(s): f730872
Files changed (2) hide show
  1. config.json +4 -2
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -1,10 +1,11 @@
1
  {
2
- "_name_or_path": "roberta-large",
3
  "architectures": [
4
  "RobertaModel"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
 
8
  "eos_token_id": 2,
9
  "gradient_checkpointing": false,
10
  "hidden_act": "gelu",
@@ -24,7 +25,8 @@
24
  "template": "Today, I finally discovered the relation between <subj> and <obj> : <obj> is <subj>'s <mask>",
25
  "template_mode": "manual"
26
  },
27
- "transformers_version": "4.6.1",
 
28
  "type_vocab_size": 1,
29
  "use_cache": true,
30
  "vocab_size": 50265
 
1
  {
2
+ "_name_or_path": "relbert_output/models/b.nce_logout.average_no_mask.roberta-large.0.000005.8.0.05.640/best_model",
3
  "architectures": [
4
  "RobertaModel"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
  "eos_token_id": 2,
10
  "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
 
25
  "template": "Today, I finally discovered the relation between <subj> and <obj> : <obj> is <subj>'s <mask>",
26
  "template_mode": "manual"
27
  },
28
+ "torch_dtype": "float32",
29
+ "transformers_version": "4.20.1",
30
  "type_vocab_size": 1,
31
  "use_cache": true,
32
  "vocab_size": 50265
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:96362bcf259a4ea0806733423df4f45b7f1305c7eed5899aa5e7024b8daa92a0
3
- size 1421595889
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1870e7722d2b5b828fd8e3370db6feb6567d0c7b967ae2c4db2bae0d3611d676
3
+ size 1421566897