jarodrigues commited on
Commit
84a991a
·
1 Parent(s): fd3794c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -76,7 +76,7 @@ The models were evaluated on downstream tasks organized into two groups.
76
  In one group, we have the two data sets from the [ASSIN 2 benchmark](https://huggingface.co/datasets/assin2), namely STS and RTE, that were used to evaluate the previous state-of-the-art model [BERTimbau Large](https://huggingface.co/neuralmind/bert-large-portuguese-cased).
77
  In the other group of data sets, we have the translations into PT-BR and PT-PT of the English data sets used for a few of the tasks in the widely-used [GLUE benchmark](https://huggingface.co/datasets/glue), which allowed us to test both Albertina-PT-* variants on a wider variety of downstream tasks.
78
 
79
- ## ASSIN 2
80
 
81
  [ASSIN 2](https://huggingface.co/datasets/assin2) is a **PT-BR data** set of approximately 10,000 sentence pairs, split into 6,500 for training, 500 for validation, and 2,448 for testing, annotated with semantic relatedness scores (range 1 to 5) and with binary entailment judgments.
82
  This data set supports the task of semantic text similarity (STS), which consists of assigning a score of how semantically related two sentences are; and the task of recognizing textual entailment (RTE), which given a pair of sentences, consists of determining whether the first entails the second.
@@ -87,7 +87,7 @@ This data set supports the task of semantic text similarity (STS), which consist
87
  | BERTimbau-large | 0.8913 | 0.8531 |
88
 
89
 
90
- ## GLUE tasks translated
91
 
92
  We resort to [PLUE](https://huggingface.co/datasets/dlb/plue) (Portuguese Language Understanding Evaluation), a data set that was obtained by automatically translating GLUE into **PT-BR**.
93
  We address four tasks from those in PLUE, namely:
@@ -113,7 +113,7 @@ We automatically translated the same four tasks from GLUE using [DeepL Translate
113
  | **Albertina-PT-BR** | 0.7942 | 0.4085 | 0.9048 | **0.8847** |
114
 
115
 
116
- ## How to use
117
 
118
  You can use this model directly with a pipeline for masked language modeling:
119
 
 
76
  In one group, we have the two data sets from the [ASSIN 2 benchmark](https://huggingface.co/datasets/assin2), namely STS and RTE, that were used to evaluate the previous state-of-the-art model [BERTimbau Large](https://huggingface.co/neuralmind/bert-large-portuguese-cased).
77
  In the other group of data sets, we have the translations into PT-BR and PT-PT of the English data sets used for a few of the tasks in the widely-used [GLUE benchmark](https://huggingface.co/datasets/glue), which allowed us to test both Albertina-PT-* variants on a wider variety of downstream tasks.
78
 
79
+ ### ASSIN 2
80
 
81
  [ASSIN 2](https://huggingface.co/datasets/assin2) is a **PT-BR data** set of approximately 10,000 sentence pairs, split into 6,500 for training, 500 for validation, and 2,448 for testing, annotated with semantic relatedness scores (range 1 to 5) and with binary entailment judgments.
82
  This data set supports the task of semantic text similarity (STS), which consists of assigning a score of how semantically related two sentences are; and the task of recognizing textual entailment (RTE), which given a pair of sentences, consists of determining whether the first entails the second.
 
87
  | BERTimbau-large | 0.8913 | 0.8531 |
88
 
89
 
90
+ ### GLUE tasks translated
91
 
92
  We resort to [PLUE](https://huggingface.co/datasets/dlb/plue) (Portuguese Language Understanding Evaluation), a data set that was obtained by automatically translating GLUE into **PT-BR**.
93
  We address four tasks from those in PLUE, namely:
 
113
  | **Albertina-PT-BR** | 0.7942 | 0.4085 | 0.9048 | **0.8847** |
114
 
115
 
116
+ # How to use
117
 
118
  You can use this model directly with a pipeline for masked language modeling:
119