Update README.md
Browse files
README.md
CHANGED
@@ -81,8 +81,8 @@ widget:
|
|
81 |
> With a new decentralized training algorithm, we fine-tuned GPT-J (6B) on 3.53 billion tokens, resulting in GPT-JT (6B), a model that outperforms many 100B+ parameter models on classification benchmarks.
|
82 |
|
83 |
We incorporated a collection of open techniques and datasets to build GPT-JT:
|
84 |
-
- GPT-JT
|
85 |
-
- We used [UL2](https://github.com/google-research/google-research/tree/master/ul2)'s training objective,
|
86 |
- The model was trained on a large collection of diverse data, including [Chain-of-Thought (CoT)](https://ai.googleblog.com/2022/05/language-models-perform-reasoning-via.html), [Public Pool of Prompts (P3) dataset](https://huggingface.co/datasets/bigscience/P3), [Natural-Instructions (NI) dataset](https://github.com/allenai/natural-instructions).
|
87 |
|
88 |
With the help of techniques mentioned above, GPT-JT significantly improves the performance of classification tasks over the original GPT-J, and even outperforms most 100B+ parameter models!
|
|
|
81 |
> With a new decentralized training algorithm, we fine-tuned GPT-J (6B) on 3.53 billion tokens, resulting in GPT-JT (6B), a model that outperforms many 100B+ parameter models on classification benchmarks.
|
82 |
|
83 |
We incorporated a collection of open techniques and datasets to build GPT-JT:
|
84 |
+
- GPT-JT is a folk of [GPT-J (6B)](https://huggingface.co/EleutherAI/gpt-j-6B), which is created by [EleutherAI](https://www.eleuther.ai);
|
85 |
+
- We used [UL2](https://github.com/google-research/google-research/tree/master/ul2)'s training objective, allowing the model to see bidirectional context of the prompt;
|
86 |
- The model was trained on a large collection of diverse data, including [Chain-of-Thought (CoT)](https://ai.googleblog.com/2022/05/language-models-perform-reasoning-via.html), [Public Pool of Prompts (P3) dataset](https://huggingface.co/datasets/bigscience/P3), [Natural-Instructions (NI) dataset](https://github.com/allenai/natural-instructions).
|
87 |
|
88 |
With the help of techniques mentioned above, GPT-JT significantly improves the performance of classification tasks over the original GPT-J, and even outperforms most 100B+ parameter models!
|