Spaces:
No application file
No application file
Update README.md
Browse files
README.md
CHANGED
@@ -7,4 +7,106 @@ sdk: streamlit
|
|
7 |
pinned: false
|
8 |
---
|
9 |
|
10 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
pinned: false
|
8 |
---
|
9 |
|
10 |
+
# Version 1
|
11 |
+
|
12 |
+
Perform a deep dive synopsis in markdown code describing the datasets and input datasetts used by two models in comparison - delving much deeper into real time papers and information on these datasets and ideally find the URL where the dataset or dataset paper can be viewed. Fix the article I have written below with a start on the datasets that were used to train the two models:
|
13 |
+
|
14 |
+
# Language Models π£οΈ
|
15 |
+
π Bloom sets new record for most performant and efficient AI model in science! πΈ
|
16 |
+
|
17 |
+
### Comparison of Large Language Models
|
18 |
+
| Model Name | Model Size (in Parameters) |
|
19 |
+
| ----------------- | -------------------------- |
|
20 |
+
| BigScience-tr11-176B | 176 billion |
|
21 |
+
| GPT-3 | 175 billion |
|
22 |
+
|
23 |
+
## GPT-3 Datasets π
|
24 |
+
- WebText
|
25 |
+
- Common Crawl
|
26 |
+
- BooksCorpus
|
27 |
+
- English Wikipedia
|
28 |
+
- Toronto Books Corpus
|
29 |
+
- OpenWebText
|
30 |
+
-
|
31 |
+
## ChatGPT Datasets - Details π
|
32 |
+
- **WebText:** A dataset of web pages crawled from domains on the Alexa top 5,000 list. This dataset was used to pretrain GPT-2.
|
33 |
+
- [WebText: A Large-Scale Unsupervised Text Corpus by Radford et al.](https://paperswithcode.com/dataset/webtext)
|
34 |
+
- **Common Crawl:** A dataset of web pages from a variety of domains, which is updated regularly. This dataset was used to pretrain GPT-3.
|
35 |
+
- [Language Models are Few-Shot Learners](https://paperswithcode.com/dataset/common-crawl) by Brown et al.
|
36 |
+
- **BooksCorpus:** A dataset of over 11,000 books from a variety of genres.
|
37 |
+
- [Scalable Methods for 8 Billion Token Language Modeling](https://paperswithcode.com/dataset/bookcorpus) by Zhu et al.
|
38 |
+
- **English Wikipedia:** A dump of the English-language Wikipedia as of 2018, with articles from 2001-2017.
|
39 |
+
- [Improving Language Understanding by Generative Pre-Training](https://huggingface.co/spaces/awacke1/WikipediaUltimateAISearch?logs=build) Space for Wikipedia Search
|
40 |
+
- **Toronto Books Corpus:** A dataset of over 7,000 books from a variety of genres, collected by the University of Toronto.
|
41 |
+
- [Massively Multilingual Sentence Embeddings for Zero-Shot Cross-Lingual Transfer and Beyond](https://paperswithcode.com/dataset/bookcorpus) by Schwenk and Douze.
|
42 |
+
- **OpenWebText:** A dataset of web pages that were filtered to remove content that was likely to be low-quality or spammy. This dataset was used to pretrain GPT-3.
|
43 |
+
- [Language Models are Few-Shot Learners](https://paperswithcode.com/dataset/openwebtext) by Brown et al.
|
44 |
+
|
45 |
+
## Big Science Model π
|
46 |
+
- π Papers:
|
47 |
+
1. BLOOM: A 176B-Parameter Open-Access Multilingual Language Model [Paper](https://arxiv.org/abs/2211.05100)
|
48 |
+
2. Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism [Paper](https://arxiv.org/abs/1909.08053)
|
49 |
+
3. 8-bit Optimizers via Block-wise Quantization [Paper](https://arxiv.org/abs/2110.02861)
|
50 |
+
4. Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation [Paper](https://arxiv.org/abs/2108.12409)
|
51 |
+
5. [Other papers related to Big Science](https://huggingface.co/models?other=doi:10.57967/hf/0003)
|
52 |
+
6. [217 other models optimized for use with Bloom](https://huggingface.co/models?other=bloom)
|
53 |
+
|
54 |
+
- π Datasets:
|
55 |
+
|
56 |
+
**Datasets:**
|
57 |
+
1. - **Universal Dependencies:** A collection of annotated corpora for natural language processing in a range of languages, with a focus on dependency parsing.
|
58 |
+
- [Universal Dependencies official website.](https://universaldependencies.org/)
|
59 |
+
2. - **WMT 2014:** The fourth edition of the Workshop on Statistical Machine Translation, featuring shared tasks on translating between English and various other languages.
|
60 |
+
- [WMT14 website.](http://www.statmt.org/wmt14/)
|
61 |
+
3. - **The Pile:** An English language corpus of diverse text, sourced from various places on the internet.
|
62 |
+
- [The Pile official website.](https://pile.eleuther.ai/)
|
63 |
+
4. - **HumanEval:** A dataset of English sentences, annotated with human judgments on a range of linguistic qualities.
|
64 |
+
- [HumanEval: An Evaluation Benchmark for Language Understanding](https://github.com/google-research-datasets/humaneval) by Gabriel Ilharco, Daniel Loureiro, Pedro Rodriguez, and Afonso Mendes.
|
65 |
+
5. - **FLORES-101:** A dataset of parallel sentences in 101 languages, designed for multilingual machine translation.
|
66 |
+
- [FLORES-101: A Massively Multilingual Parallel Corpus for Language Understanding](https://flores101.opennmt.net/) by Aman Madaan, Shruti Rijhwani, Raghav Gupta, and Mitesh M. Khapra.
|
67 |
+
6. - **CrowS-Pairs:** A dataset of sentence pairs, designed for evaluating the plausibility of generated text.
|
68 |
+
- [CrowS-Pairs: A Challenge Dataset for Plausible Plausibility Judgments](https://github.com/stanford-cogsci/crows-pairs) by Andrea Madotto, Zhaojiang Lin, Chien-Sheng Wu, Pascale Fung, and Caiming Xiong.
|
69 |
+
7. - **WikiLingua:** A dataset of parallel sentences in 75 languages, sourced from Wikipedia.
|
70 |
+
- [WikiLingua: A New Benchmark Dataset for Cross-Lingual Wikification](https://arxiv.org/abs/2105.08031) by Jiarui Yao, Yanqiao Zhu, Ruihan Bao, Guosheng Lin, Lidong Bing, and Bei Shi.
|
71 |
+
8. - **MTEB:** A dataset of English sentences, annotated with their entailment relationships with respect to other sentences.
|
72 |
+
- [Multi-Task Evaluation Benchmark for Natural Language Inference](https://github.com/google-research-datasets/mteb) by MichaΕ Lukasik, Marcin Junczys-Dowmunt, and Houda Bouamor.
|
73 |
+
9. - **xP3:** A dataset of English sentences, annotated with their paraphrase relationships with respect to other sentences.
|
74 |
+
- [xP3: A Large-Scale Evaluation Benchmark for Paraphrase Identification in Context](https://github.com/nyu-dl/xp3) by Aniket Didolkar, James Mayfield, Markus Saers, and Jason Baldridge.
|
75 |
+
10. - **DiaBLa:** A dataset of English dialogue, annotated with dialogue acts.
|
76 |
+
- [A Large-Scale Corpus for Conversation Disentanglement](https://github.com/HLTCHKUST/DiaBLA) by Samuel Broscheit, AntΓ³nio Branco, and AndrΓ© F. T. Martins.
|
77 |
+
|
78 |
+
- π Dataset Papers with Code
|
79 |
+
1. [Universal Dependencies](https://paperswithcode.com/dataset/universal-dependencies)
|
80 |
+
2. [WMT 2014](https://paperswithcode.com/dataset/wmt-2014)
|
81 |
+
3. [The Pile](https://paperswithcode.com/dataset/the-pile)
|
82 |
+
4. [HumanEval](https://paperswithcode.com/dataset/humaneval)
|
83 |
+
5. [FLORES-101](https://paperswithcode.com/dataset/flores-101)
|
84 |
+
6. [CrowS-Pairs](https://paperswithcode.com/dataset/crows-pairs)
|
85 |
+
7. [WikiLingua](https://paperswithcode.com/dataset/wikilingua)
|
86 |
+
8. [MTEB](https://paperswithcode.com/dataset/mteb)
|
87 |
+
9. [xP3](https://paperswithcode.com/dataset/xp3)
|
88 |
+
10. [DiaBLa](https://paperswithcode.com/dataset/diabla)
|
89 |
+
|
90 |
+
# Deep RL ML Strategy π§
|
91 |
+
The AI strategies are:
|
92 |
+
- Language Model Preparation using Human Augmented with Supervised Fine Tuning π€
|
93 |
+
- Reward Model Training with Prompts Dataset Multi-Model Generate Data to Rank π
|
94 |
+
- Fine Tuning with Reinforcement Reward and Distance Distribution Regret Score π―
|
95 |
+
- Proximal Policy Optimization Fine Tuning π€
|
96 |
+
- Variations - Preference Model Pretraining π€
|
97 |
+
- Use Ranking Datasets Sentiment - Thumbs Up/Down, Distribution π
|
98 |
+
- Online Version Getting Feedback π¬
|
99 |
+
- OpenAI - InstructGPT - Humans generate LM Training Text π
|
100 |
+
- DeepMind - Advantage Actor Critic Sparrow, GopherCite π¦
|
101 |
+
- Reward Model Human Prefence Feedback π
|
102 |
+
|
103 |
+
|
104 |
+
For more information on specific techniques and implementations, check out the following resources:
|
105 |
+
- OpenAI's paper on [GPT-3](https://arxiv.org/abs/2005.14165) which details their Language Model Preparation approach
|
106 |
+
- DeepMind's paper on [SAC](https://arxiv.org/abs/1801.01290) which describes the Advantage Actor Critic algorithm
|
107 |
+
- OpenAI's paper on [Reward Learning](https://arxiv.org/abs/1810.06580) which explains their approach to training Reward Models
|
108 |
+
- OpenAI's blog post on [GPT-3's fine-tuning process](https://openai.com/blog/fine-tuning-gpt-3/)
|
109 |
+
|
110 |
+
# Version 2:
|
111 |
+
|
112 |
+
|