Update README.md
Browse files
README.md
CHANGED
@@ -9,6 +9,7 @@ This model is a fine-tuned version of XLM-RoBERTa (xlm-roberta-base) for Named E
|
|
9 |
|
10 |
PER: Person names
|
11 |
ORG: Organization names
|
|
|
12 |
|
13 |
|
14 |
<!-- Provide a quick summary of what the model is/does. -->
|
@@ -40,7 +41,12 @@ This modelcard aims to be a base template for new models. It has been generated
|
|
40 |
- **Demo [optional]:** [More Information Needed]
|
41 |
|
42 |
## Uses
|
|
|
43 |
|
|
|
|
|
|
|
|
|
44 |
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
45 |
|
46 |
### Direct Use
|
@@ -80,6 +86,10 @@ Use the code below to get started with the model.
|
|
80 |
[More Information Needed]
|
81 |
|
82 |
## Training Details
|
|
|
|
|
|
|
|
|
83 |
|
84 |
### Training Data
|
85 |
|
@@ -97,6 +107,7 @@ Use the code below to get started with the model.
|
|
97 |
|
98 |
|
99 |
#### Training Hyperparameters
|
|
|
100 |
|
101 |
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
102 |
|
|
|
9 |
|
10 |
PER: Person names
|
11 |
ORG: Organization names
|
12 |
+
LOC: Location names
|
13 |
|
14 |
|
15 |
<!-- Provide a quick summary of what the model is/does. -->
|
|
|
41 |
- **Demo [optional]:** [More Information Needed]
|
42 |
|
43 |
## Uses
|
44 |
+
This model is suitable for multilingual NER tasks, especially in scenarios where extracting and classifying person, organization, and location names in text across different languages is required.
|
45 |
|
46 |
+
Applications:
|
47 |
+
Information extraction
|
48 |
+
Multilingual NER tasks
|
49 |
+
Automated text analysis for businesses
|
50 |
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
51 |
|
52 |
### Direct Use
|
|
|
86 |
[More Information Needed]
|
87 |
|
88 |
## Training Details
|
89 |
+
Base Model: xlm-roberta-base
|
90 |
+
Training Dataset: The model is trained on the PAN-X subset of the XTREME dataset, which includes labeled NER data for multiple languages.
|
91 |
+
Training Framework: Hugging Face transformers library with PyTorch backend.
|
92 |
+
Data Preprocessing: Tokenization was performed using XLM-RoBERTa tokenizer, with attention paid to aligning token labels to subword tokens.
|
93 |
|
94 |
### Training Data
|
95 |
|
|
|
107 |
|
108 |
|
109 |
#### Training Hyperparameters
|
110 |
+
The model's performance is evaluated using the F1 score for NER. The predictions are aligned with gold-standard labels, ignoring sub-token predictions where appropriate.
|
111 |
|
112 |
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
113 |
|