TajaKuzman
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -216,7 +216,7 @@ and enriched with information which specific subtopics belong to the top-level t
|
|
216 |
## Training data
|
217 |
|
218 |
The model was fine-tuned on a training dataset consisting of 15,000 news in four languages (Croatian, Slovenian, Catalan and Greek).
|
219 |
-
The news texts were extracted from the [MaCoCu web corpora](
|
220 |
The training dataset was automatically annotated with the IPTC Media Topic labels by
|
221 |
the [GPT-4o](https://platform.openai.com/docs/models/gpt-4o) model (yielding 0.72 micro-F1 and 0.73 macro-F1 on the test dataset).
|
222 |
|
|
|
216 |
## Training data
|
217 |
|
218 |
The model was fine-tuned on a training dataset consisting of 15,000 news in four languages (Croatian, Slovenian, Catalan and Greek).
|
219 |
+
The news texts were extracted from the [MaCoCu-Genre web corpora](http://hdl.handle.net/11356/1969) based on the "News" genre label, predicted with the [X-GENRE classifier](https://huggingface.co/classla/xlm-roberta-base-multilingual-text-genre-classifier).
|
220 |
The training dataset was automatically annotated with the IPTC Media Topic labels by
|
221 |
the [GPT-4o](https://platform.openai.com/docs/models/gpt-4o) model (yielding 0.72 micro-F1 and 0.73 macro-F1 on the test dataset).
|
222 |
|