MassMin commited on
Commit
af96897
·
verified ·
1 Parent(s): f3f513f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -0
README.md CHANGED
@@ -9,6 +9,7 @@ This model is a fine-tuned version of XLM-RoBERTa (xlm-roberta-base) for Named E
9
 
10
  PER: Person names
11
  ORG: Organization names
 
12
 
13
 
14
  <!-- Provide a quick summary of what the model is/does. -->
@@ -40,7 +41,12 @@ This modelcard aims to be a base template for new models. It has been generated
40
  - **Demo [optional]:** [More Information Needed]
41
 
42
  ## Uses
 
43
 
 
 
 
 
44
  <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
45
 
46
  ### Direct Use
@@ -80,6 +86,10 @@ Use the code below to get started with the model.
80
  [More Information Needed]
81
 
82
  ## Training Details
 
 
 
 
83
 
84
  ### Training Data
85
 
@@ -97,6 +107,7 @@ Use the code below to get started with the model.
97
 
98
 
99
  #### Training Hyperparameters
 
100
 
101
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
102
 
 
9
 
10
  PER: Person names
11
  ORG: Organization names
12
+ LOC: Location names
13
 
14
 
15
  <!-- Provide a quick summary of what the model is/does. -->
 
41
  - **Demo [optional]:** [More Information Needed]
42
 
43
  ## Uses
44
+ This model is suitable for multilingual NER tasks, especially in scenarios where extracting and classifying person, organization, and location names in text across different languages is required.
45
 
46
+ Applications:
47
+ Information extraction
48
+ Multilingual NER tasks
49
+ Automated text analysis for businesses
50
  <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
51
 
52
  ### Direct Use
 
86
  [More Information Needed]
87
 
88
  ## Training Details
89
+ Base Model: xlm-roberta-base
90
+ Training Dataset: The model is trained on the PAN-X subset of the XTREME dataset, which includes labeled NER data for multiple languages.
91
+ Training Framework: Hugging Face transformers library with PyTorch backend.
92
+ Data Preprocessing: Tokenization was performed using XLM-RoBERTa tokenizer, with attention paid to aligning token labels to subword tokens.
93
 
94
  ### Training Data
95
 
 
107
 
108
 
109
  #### Training Hyperparameters
110
+ The model's performance is evaluated using the F1 score for NER. The predictions are aligned with gold-standard labels, ignoring sub-token predictions where appropriate.
111
 
112
  - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
113