raulgdp commited on
Commit
80b43b0
·
verified ·
1 Parent(s): d85688f

End of training

Browse files
README.md ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: mit
4
+ base_model: FacebookAI/xlm-roberta-large
5
+ tags:
6
+ - generated_from_trainer
7
+ datasets:
8
+ - conll2002
9
+ metrics:
10
+ - precision
11
+ - recall
12
+ - f1
13
+ - accuracy
14
+ model-index:
15
+ - name: xlm-roberta-large-finetuned-ner
16
+ results:
17
+ - task:
18
+ name: Token Classification
19
+ type: token-classification
20
+ dataset:
21
+ name: conll2002
22
+ type: conll2002
23
+ config: es
24
+ split: validation
25
+ args: es
26
+ metrics:
27
+ - name: Precision
28
+ type: precision
29
+ value: 0.86443345323741
30
+ - name: Recall
31
+ type: recall
32
+ value: 0.8835018382352942
33
+ - name: F1
34
+ type: f1
35
+ value: 0.8738636363636364
36
+ - name: Accuracy
37
+ type: accuracy
38
+ value: 0.9787686065955755
39
+ ---
40
+
41
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
42
+ should probably proofread and complete it, then remove this comment. -->
43
+
44
+ # xlm-roberta-large-finetuned-ner
45
+
46
+ This model is a fine-tuned version of [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) on the conll2002 dataset.
47
+ It achieves the following results on the evaluation set:
48
+ - Loss: 0.0973
49
+ - Precision: 0.8644
50
+ - Recall: 0.8835
51
+ - F1: 0.8739
52
+ - Accuracy: 0.9788
53
+
54
+ ## Model description
55
+
56
+ More information needed
57
+
58
+ ## Intended uses & limitations
59
+
60
+ More information needed
61
+
62
+ ## Training and evaluation data
63
+
64
+ More information needed
65
+
66
+ ## Training procedure
67
+
68
+ ### Training hyperparameters
69
+
70
+ The following hyperparameters were used during training:
71
+ - learning_rate: 2e-05
72
+ - train_batch_size: 16
73
+ - eval_batch_size: 16
74
+ - seed: 42
75
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
76
+ - lr_scheduler_type: linear
77
+ - num_epochs: 3
78
+
79
+ ### Training results
80
+
81
+ | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
82
+ |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
83
+ | 0.1382 | 1.0 | 521 | 0.0906 | 0.8502 | 0.8830 | 0.8663 | 0.9782 |
84
+ | 0.048 | 2.0 | 1042 | 0.0861 | 0.8472 | 0.8729 | 0.8599 | 0.9780 |
85
+ | 0.0294 | 3.0 | 1563 | 0.0973 | 0.8644 | 0.8835 | 0.8739 | 0.9788 |
86
+
87
+
88
+ ### Framework versions
89
+
90
+ - Transformers 4.46.3
91
+ - Pytorch 2.5.1
92
+ - Datasets 3.1.0
93
+ - Tokenizers 0.20.3
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:640d80e5405badea55d79ed10a3eb348c68903819dae336669ff21c606e0050a
3
  size 2235448756
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41c8bb24cb591fe707a6062d89ff61f7bfa5ba57c32d84d4cef72d052cfc9801
3
  size 2235448756
runs/Nov22_13-59-56_DESKTOP-P79TL96/events.out.tfevents.1732301998.DESKTOP-P79TL96 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:99986646955e9a9a4625cefb5b5e2fcc2d4529a39de6b8de4973949cfbb01131
3
- size 6803
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a89b81a98d0341d0b649a779bf1502c036b338d6befad9be76232735f552e71
3
+ size 7840