Commit
·
84a991a
1
Parent(s):
fd3794c
Update README.md
Browse files
README.md
CHANGED
@@ -76,7 +76,7 @@ The models were evaluated on downstream tasks organized into two groups.
|
|
76 |
In one group, we have the two data sets from the [ASSIN 2 benchmark](https://huggingface.co/datasets/assin2), namely STS and RTE, that were used to evaluate the previous state-of-the-art model [BERTimbau Large](https://huggingface.co/neuralmind/bert-large-portuguese-cased).
|
77 |
In the other group of data sets, we have the translations into PT-BR and PT-PT of the English data sets used for a few of the tasks in the widely-used [GLUE benchmark](https://huggingface.co/datasets/glue), which allowed us to test both Albertina-PT-* variants on a wider variety of downstream tasks.
|
78 |
|
79 |
-
|
80 |
|
81 |
[ASSIN 2](https://huggingface.co/datasets/assin2) is a **PT-BR data** set of approximately 10,000 sentence pairs, split into 6,500 for training, 500 for validation, and 2,448 for testing, annotated with semantic relatedness scores (range 1 to 5) and with binary entailment judgments.
|
82 |
This data set supports the task of semantic text similarity (STS), which consists of assigning a score of how semantically related two sentences are; and the task of recognizing textual entailment (RTE), which given a pair of sentences, consists of determining whether the first entails the second.
|
@@ -87,7 +87,7 @@ This data set supports the task of semantic text similarity (STS), which consist
|
|
87 |
| BERTimbau-large | 0.8913 | 0.8531 |
|
88 |
|
89 |
|
90 |
-
|
91 |
|
92 |
We resort to [PLUE](https://huggingface.co/datasets/dlb/plue) (Portuguese Language Understanding Evaluation), a data set that was obtained by automatically translating GLUE into **PT-BR**.
|
93 |
We address four tasks from those in PLUE, namely:
|
@@ -113,7 +113,7 @@ We automatically translated the same four tasks from GLUE using [DeepL Translate
|
|
113 |
| **Albertina-PT-BR** | 0.7942 | 0.4085 | 0.9048 | **0.8847** |
|
114 |
|
115 |
|
116 |
-
|
117 |
|
118 |
You can use this model directly with a pipeline for masked language modeling:
|
119 |
|
|
|
76 |
In one group, we have the two data sets from the [ASSIN 2 benchmark](https://huggingface.co/datasets/assin2), namely STS and RTE, that were used to evaluate the previous state-of-the-art model [BERTimbau Large](https://huggingface.co/neuralmind/bert-large-portuguese-cased).
|
77 |
In the other group of data sets, we have the translations into PT-BR and PT-PT of the English data sets used for a few of the tasks in the widely-used [GLUE benchmark](https://huggingface.co/datasets/glue), which allowed us to test both Albertina-PT-* variants on a wider variety of downstream tasks.
|
78 |
|
79 |
+
### ASSIN 2
|
80 |
|
81 |
[ASSIN 2](https://huggingface.co/datasets/assin2) is a **PT-BR data** set of approximately 10,000 sentence pairs, split into 6,500 for training, 500 for validation, and 2,448 for testing, annotated with semantic relatedness scores (range 1 to 5) and with binary entailment judgments.
|
82 |
This data set supports the task of semantic text similarity (STS), which consists of assigning a score of how semantically related two sentences are; and the task of recognizing textual entailment (RTE), which given a pair of sentences, consists of determining whether the first entails the second.
|
|
|
87 |
| BERTimbau-large | 0.8913 | 0.8531 |
|
88 |
|
89 |
|
90 |
+
### GLUE tasks translated
|
91 |
|
92 |
We resort to [PLUE](https://huggingface.co/datasets/dlb/plue) (Portuguese Language Understanding Evaluation), a data set that was obtained by automatically translating GLUE into **PT-BR**.
|
93 |
We address four tasks from those in PLUE, namely:
|
|
|
113 |
| **Albertina-PT-BR** | 0.7942 | 0.4085 | 0.9048 | **0.8847** |
|
114 |
|
115 |
|
116 |
+
# How to use
|
117 |
|
118 |
You can use this model directly with a pipeline for masked language modeling:
|
119 |
|