sciarrilli commited on
Commit
6aa823d
1 Parent(s): feec333

Training in progress, step 500

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "dmis-lab/biobert-base-cased-v1.2",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
 
1
  {
2
+ "_name_or_path": "sciarrilli/biobert-base-cased-v1.2-finetuned-ner",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4e53b0601f7a55a1acda8eec68d305415a43f0adfeb4ccac7de4d9e62f83c2ed
3
  size 430995953
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4361bcc86ddc8271482bb1e38b1d283225220dc2f7824da300dca4661df3a07c
3
  size 430995953
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "special_tokens_map_file": null, "name_or_path": "dmis-lab/biobert-base-cased-v1.2", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "BertTokenizer"}
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "special_tokens_map_file": null, "name_or_path": "sciarrilli/biobert-base-cased-v1.2-finetuned-ner", "do_basic_tokenize": true, "never_split": null, "tokenizer_class": "BertTokenizer"}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:078d580dee8d7731d387a774a6a3e69cd5d98ac3c68282e13af6f4bbe274d648
3
  size 2927
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84bee8f1efbc708d8ad4875b78282507eef3874cec2dc042b51d39a231a030a0
3
  size 2927