Update README.md
Browse files
README.md
CHANGED
@@ -17,7 +17,7 @@ language:
|
|
17 |
Similarly to its base model, Pleias-1b, Pleias-1b-RAG 0.1 aims to be a fully open model (weights, code, data), only trained on content with a permissible license and fully compliant with the upcoming European AI Act.
|
18 |
|
19 |
## Description
|
20 |
-
PleIAs-1b-RAG is continuous pretraining of Pleias-
|
21 |
|
22 |
Pleias-1b-RAG includes the main features of the original base model:
|
23 |
* Only trained on open data under a permissible license and in compliance with the European AI Act. By design, all Pleias model are unable to output copyrighted content.
|
@@ -30,12 +30,12 @@ Pleias-1b-RAG supports retrieval-augmented generation with enhanced verifiabilit
|
|
30 |
* Source analysis/criticism which also acts as an integrated reranker step.
|
31 |
* Generation of ground answers with references and excerpts linked to the original sources.
|
32 |
|
33 |
-
While the base model Pleias-1b-RAG has been made available as an experimental preview, we release Pleias-
|
34 |
|
35 |
## Training
|
36 |
-
PleIAs-1b-RAG was trained
|
37 |
|
38 |
-
PleIAs-1b-RAG derives from the last checkpoint of PleIAs-
|
39 |
|
40 |
Training covers the entire RAG dataset we have been designing out of Common Corpus for 3 epochs.
|
41 |
|
|
|
17 |
Similarly to its base model, Pleias-1b, Pleias-1b-RAG 0.1 aims to be a fully open model (weights, code, data), only trained on content with a permissible license and fully compliant with the upcoming European AI Act.
|
18 |
|
19 |
## Description
|
20 |
+
PleIAs-1b-RAG is continuous pretraining of Pleias-1b on a new dataset of 45,088,768,000 tokens modeling common retrieval tasks. All the content of the dataset is ultimately coming from Common Corpus.
|
21 |
|
22 |
Pleias-1b-RAG includes the main features of the original base model:
|
23 |
* Only trained on open data under a permissible license and in compliance with the European AI Act. By design, all Pleias model are unable to output copyrighted content.
|
|
|
30 |
* Source analysis/criticism which also acts as an integrated reranker step.
|
31 |
* Generation of ground answers with references and excerpts linked to the original sources.
|
32 |
|
33 |
+
While the base model Pleias-1b-RAG has been made available as an experimental preview, we release Pleias-1b-RAG 0.1 as an early version. Pleias-3b-RAG 0.1 has been already tested and integrated into multiple applied RAG projects, including Pleias flagship application Scholastikai.
|
34 |
|
35 |
## Training
|
36 |
+
PleIAs-1b-RAG was trained pretrained on TractoAI on ISEG GPU cluster by Nebius AI on the fork Nanotron developed by TractoAI. We provide the complete settings as a yaml file as part of our release.
|
37 |
|
38 |
+
PleIAs-1b-RAG derives from the last checkpoint of PleIAs-1b (369,000). The training schedule reused the last learning rate value (5e-6) without decay for 43,000 steps. Each step is about 10 time smaller than the original steps from the base model training (roughly 1M tokens per step vs. 12M tokens)
|
39 |
|
40 |
Training covers the entire RAG dataset we have been designing out of Common Corpus for 3 epochs.
|
41 |
|