ahmetustun
commited on
Commit
•
fc44676
1
Parent(s):
d53e4f9
Update README.md
Browse files
README.md
CHANGED
@@ -127,7 +127,7 @@ metrics:
|
|
127 |
|
128 |
- **Developed by:** Cohere For AI
|
129 |
- **Model type:** a Transformer style autoregressive massively multilingual language model.
|
130 |
-
- **Paper**: [Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model](arxiv.
|
131 |
- **Point of Contact**: Cohere For AI: [cohere.for.ai](cohere.for.ai)
|
132 |
- **Languages**: Refer to the list of languages in the `language` section of this model card.
|
133 |
- **License**: Apache-2.0
|
@@ -180,16 +180,16 @@ The Aya model is trained on the following datasets:
|
|
180 |
- [DataProvenance collection](https://huggingface.co/datasets/DataProvenanceInitiative/Commercially-Verified-Licenses)
|
181 |
- ShareGPT-Command
|
182 |
|
183 |
-
All datasets are subset to the 101 languages supported by [mT5]. See the [paper](arxiv.
|
184 |
|
185 |
## Evaluation
|
186 |
|
187 |
-
We refer to Section 5 from our paper for multilingual eval across 99 languages – including discriminative
|
188 |
|
189 |
## Bias, Risks, and Limitations
|
190 |
|
191 |
|
192 |
-
For a detailed overview of our effort at safety mitigation and benchmarking toxicity and bias across multiple languages, we refer Sections 6 and 7 of our paper: [Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model](arxiv.
|
193 |
|
194 |
We hope that the release of the Aya model will make community-based redteaming efforts possible, by exposing an open-source massively-multilingual model for community research.
|
195 |
|
|
|
127 |
|
128 |
- **Developed by:** Cohere For AI
|
129 |
- **Model type:** a Transformer style autoregressive massively multilingual language model.
|
130 |
+
- **Paper**: [Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model](https://arxiv.org/abs/2402.07827)
|
131 |
- **Point of Contact**: Cohere For AI: [cohere.for.ai](cohere.for.ai)
|
132 |
- **Languages**: Refer to the list of languages in the `language` section of this model card.
|
133 |
- **License**: Apache-2.0
|
|
|
180 |
- [DataProvenance collection](https://huggingface.co/datasets/DataProvenanceInitiative/Commercially-Verified-Licenses)
|
181 |
- ShareGPT-Command
|
182 |
|
183 |
+
All datasets are subset to the 101 languages supported by [mT5]. See the [paper](https://arxiv.org/abs/2402.07827) for details about filtering and pruning.
|
184 |
|
185 |
## Evaluation
|
186 |
|
187 |
+
We refer to Section 5 from our paper for multilingual eval across 99 languages – including discriminative and generative tasks, human evaluation, and simulated win rates that cover both held-out tasks and in-distribution performance.
|
188 |
|
189 |
## Bias, Risks, and Limitations
|
190 |
|
191 |
|
192 |
+
For a detailed overview of our effort at safety mitigation and benchmarking toxicity and bias across multiple languages, we refer to Sections 6 and 7 of our paper: [Aya Model: An Instruction Finetuned Open-Access Multilingual Language Model](https://arxiv.org/abs/2402.07827).
|
193 |
|
194 |
We hope that the release of the Aya model will make community-based redteaming efforts possible, by exposing an open-source massively-multilingual model for community research.
|
195 |
|