quim-motger commited on
Commit
49140de
1 Parent(s): 548f836

Upload 12 files

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "roberta-base",
3
  "architectures": [
4
  "RobertaForTokenClassification"
5
  ],
@@ -30,7 +30,7 @@
30
  "pad_token_id": 1,
31
  "position_embedding_type": "absolute",
32
  "torch_dtype": "float32",
33
- "transformers_version": "4.30.2",
34
  "type_vocab_size": 1,
35
  "use_cache": true,
36
  "vocab_size": 50265
 
1
  {
2
+ "_name_or_path": "data/further_pretraining/roberta-base/checkpoint-16373",
3
  "architectures": [
4
  "RobertaForTokenClassification"
5
  ],
 
30
  "pad_token_id": 1,
31
  "position_embedding_type": "absolute",
32
  "torch_dtype": "float32",
33
+ "transformers_version": "4.39.1",
34
  "type_vocab_size": 1,
35
  "use_cache": true,
36
  "vocab_size": 50265
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bafbae8fe8dea92a388e93e105869d450b3c2c7e79cc85510d9ec67f5e8f5d4f
3
+ size 496253316
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:733e497bb47cdce3c722d944ec16a0ca562b7bd2249630b31f86f7a86e258887
3
- size 992575493
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cc39f5c534760a502bbd0c81c5ff95dc7860bf85f793d033ce71d75f666f5f3
3
+ size 992625210
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:20dd2fbd7f85a1ea702fb954d9d6f4ee89ef264b0c90fe34c1a7f559768e1b20
3
- size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4508b02e23b5fc2b51eaed019d385acf0c7d24d11b60eeb81389ce5cf020b7f4
3
+ size 14244
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:30269608c9e0540a595d8484ff39d209fca7698cf0f2639d73b5d81109885fed
3
- size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1f42bafe915d489bc148c745aee5cb40266ca728ccbcb98c9e87b2cedeab3b6
3
+ size 1064
tokenizer.json CHANGED
@@ -14,7 +14,7 @@
14
  "single_word": false,
15
  "lstrip": false,
16
  "rstrip": false,
17
- "normalized": false,
18
  "special": true
19
  },
20
  {
@@ -23,7 +23,7 @@
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
- "normalized": false,
27
  "special": true
28
  },
29
  {
@@ -32,7 +32,7 @@
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
- "normalized": false,
36
  "special": true
37
  },
38
  {
@@ -41,7 +41,7 @@
41
  "single_word": false,
42
  "lstrip": false,
43
  "rstrip": false,
44
- "normalized": false,
45
  "special": true
46
  },
47
  {
 
14
  "single_word": false,
15
  "lstrip": false,
16
  "rstrip": false,
17
+ "normalized": true,
18
  "special": true
19
  },
20
  {
 
23
  "single_word": false,
24
  "lstrip": false,
25
  "rstrip": false,
26
+ "normalized": true,
27
  "special": true
28
  },
29
  {
 
32
  "single_word": false,
33
  "lstrip": false,
34
  "rstrip": false,
35
+ "normalized": true,
36
  "special": true
37
  },
38
  {
 
41
  "single_word": false,
42
  "lstrip": false,
43
  "rstrip": false,
44
+ "normalized": true,
45
  "special": true
46
  },
47
  {
tokenizer_config.json CHANGED
@@ -1,5 +1,47 @@
1
  {
2
  "add_prefix_space": true,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "bos_token": "<s>",
4
  "clean_up_tokenization_spaces": true,
5
  "cls_token": "<s>",
 
1
  {
2
  "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
  "bos_token": "<s>",
46
  "clean_up_tokenization_spaces": true,
47
  "cls_token": "<s>",
trainer_state.json CHANGED
@@ -1,46 +1,66 @@
1
  {
2
- "best_metric": 0.005157737527042627,
3
- "best_model_checkpoint": "data/train-test///model/checkpoint-1676",
4
- "epoch": 1.0,
 
5
  "global_step": 1676,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
- "epoch": 0.3,
12
- "learning_rate": 1.7016706443914083e-05,
13
- "loss": 0.0422,
 
14
  "step": 500
15
  },
16
  {
17
- "epoch": 0.6,
18
- "learning_rate": 1.4033412887828164e-05,
19
- "loss": 0.0062,
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  "step": 1000
21
  },
22
  {
23
- "epoch": 0.89,
24
- "learning_rate": 1.1050119331742244e-05,
25
- "loss": 0.0052,
 
26
  "step": 1500
27
  },
28
  {
29
- "epoch": 1.0,
30
  "eval_accuracy": null,
31
- "eval_f1": 0.9925373134328359,
32
- "eval_loss": 0.005157737527042627,
33
- "eval_precision": 0.9906890130353817,
34
- "eval_recall": 0.994392523364486,
35
- "eval_runtime": 0.382,
36
- "eval_samples_per_second": 2539.591,
37
- "eval_steps_per_second": 159.706,
38
  "step": 1676
39
  }
40
  ],
41
- "max_steps": 3352,
 
 
42
  "num_train_epochs": 2,
43
- "total_flos": 787274342216514.0,
 
 
44
  "trial_name": null,
45
  "trial_params": null
46
  }
 
1
  {
2
+ "best_metric": 0.9963201471941122,
3
+ "best_model_checkpoint": "data/train-test/roberta-base-output//model/checkpoint-1676",
4
+ "epoch": 2.0,
5
+ "eval_steps": 500,
6
  "global_step": 1676,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.6,
13
+ "grad_norm": 1.825121283531189,
14
+ "learning_rate": 1.4033412887828164e-05,
15
+ "loss": 0.0322,
16
  "step": 500
17
  },
18
  {
19
+ "epoch": 1.0,
20
+ "eval_accuracy": null,
21
+ "eval_f1": 0.9926538108356291,
22
+ "eval_loss": 0.00715277437120676,
23
+ "eval_precision": 0.9908340971585701,
24
+ "eval_recall": 0.9944802207911684,
25
+ "eval_runtime": 1.6458,
26
+ "eval_samples_per_second": 589.386,
27
+ "eval_steps_per_second": 18.836,
28
+ "step": 838
29
+ },
30
+ {
31
+ "epoch": 1.19,
32
+ "grad_norm": 0.7001124620437622,
33
+ "learning_rate": 8.066825775656326e-06,
34
+ "loss": 0.0046,
35
  "step": 1000
36
  },
37
  {
38
+ "epoch": 1.79,
39
+ "grad_norm": 0.1312304437160492,
40
+ "learning_rate": 2.100238663484487e-06,
41
+ "loss": 0.0027,
42
  "step": 1500
43
  },
44
  {
45
+ "epoch": 2.0,
46
  "eval_accuracy": null,
47
+ "eval_f1": 0.9963201471941122,
48
+ "eval_loss": 0.004037069622427225,
49
+ "eval_precision": 0.9963201471941122,
50
+ "eval_recall": 0.9963201471941122,
51
+ "eval_runtime": 1.6277,
52
+ "eval_samples_per_second": 595.928,
53
+ "eval_steps_per_second": 19.045,
54
  "step": 1676
55
  }
56
  ],
57
+ "logging_steps": 500,
58
+ "max_steps": 1676,
59
+ "num_input_tokens_seen": 0,
60
  "num_train_epochs": 2,
61
+ "save_steps": 500,
62
+ "total_flos": 2013781275950328.0,
63
+ "train_batch_size": 32,
64
  "trial_name": null,
65
  "trial_params": null
66
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bb50de76f6ddda59be0ab4580535665e5b1e6934bc7aa1e68ecd5d14330a6190
3
- size 3963
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57817dee2c4bf33835c981bcbee2a13e690f7c763ce659b0aa87a28b502670d3
3
+ size 4984