Update README.md
Browse files
README.md
CHANGED
@@ -181,7 +181,8 @@ Note: A small amount of English data was kept to avoid catastrophic forgetting.
|
|
181 |
|
182 |
The training corpus has been tokenized using a byte version of [Byte-Pair Encoding (BPE)](https://github.com/openai/gpt-2) with a vocabulary size of 50,257 tokens.
|
183 |
After training a new tokenizer and adapting [falcon-7b](https://huggingface.co/tiiuae/falcon-7b)'s embedding layer, the model was
|
184 |
-
further pre-trained in three target languages: Catalan, Spanish and English.
|
|
|
185 |
The training lasted a total of 320 hours on 8 NVIDIA H100 GPUs with 80GB RAM.
|
186 |
|
187 |
|
@@ -233,11 +234,11 @@ This work was partially funded by:
|
|
233 |
<details>
|
234 |
<summary>Click to expand</summary>
|
235 |
|
236 |
-
The model published in this repository is intended for a generalist purpose and is available to third parties.
|
237 |
|
238 |
-
|
239 |
|
240 |
-
When third parties deploy or provide systems and/or services to other parties using this model (or any
|
241 |
or become users of the model, they should note that it is their responsibility to mitigate the risks arising from its use and,
|
242 |
in any event, to comply with applicable regulations, including regulations regarding the use of Artificial Intelligence.
|
243 |
|
|
|
181 |
|
182 |
The training corpus has been tokenized using a byte version of [Byte-Pair Encoding (BPE)](https://github.com/openai/gpt-2) with a vocabulary size of 50,257 tokens.
|
183 |
After training a new tokenizer and adapting [falcon-7b](https://huggingface.co/tiiuae/falcon-7b)'s embedding layer, the model was
|
184 |
+
further pre-trained in three target languages: Catalan, Spanish and English.
|
185 |
+
|
186 |
The training lasted a total of 320 hours on 8 NVIDIA H100 GPUs with 80GB RAM.
|
187 |
|
188 |
|
|
|
234 |
<details>
|
235 |
<summary>Click to expand</summary>
|
236 |
|
237 |
+
The model published in this repository is intended for a generalist purpose and is available to third parties under a permissive Apache License, Version 2.0.
|
238 |
|
239 |
+
Be aware that the model may have biases and/or any other undesirable distortions.
|
240 |
|
241 |
+
When third parties deploy or provide systems and/or services to other parties using this model (or any system based on it)
|
242 |
or become users of the model, they should note that it is their responsibility to mitigate the risks arising from its use and,
|
243 |
in any event, to comply with applicable regulations, including regulations regarding the use of Artificial Intelligence.
|
244 |
|