Christopher Klamm commited on
Commit
8c418e3
1 Parent(s): f6af94a

updated model

Browse files
.DS_Store ADDED
Binary file (6.15 kB). View file
 
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/data/My Drive/data/ArgumenText_2018/roberta_model/checkpoint-2000",
3
  "architectures": [
4
  "RobertaForSequenceClassification"
5
  ],
 
1
  {
2
+ "_name_or_path": "roberta-base",
3
  "architectures": [
4
  "RobertaForSequenceClassification"
5
  ],
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a2d4f42ee41eb99e2e1b86bac4d53e5dabc89fc37a2df290d87b1b8694418fb
3
  size 498679241
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95774bf144e6794e6dd49c2bbfafbb7de79658f6e3cf3fe8e10b4da1890210d9
3
  size 498679241
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "model_max_length": 512, "name_or_path": "roberta-base"}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f88d802fe15c89b2671a1e07922c61649bac586fb328c8aebeb76f051a8b8cf2
3
  size 2159
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d5305fc7342164c6e61013b2d0a219fae993dfdd1b809091447a2e1516be6e5
3
  size 2159
vocab.json ADDED
The diff for this file is too large to render. See raw diff