hop
Browse files- dist/index.html +1 -1
- src/index.html +1 -1
dist/index.html
CHANGED
@@ -73,6 +73,7 @@
|
|
73 |
|
74 |
<p>The performance of a large language model (LLM) depends heavily on the quality and size of its pretraining dataset.
|
75 |
However, the pretraining datasets for state-of-the-art open LLMs like Llama 3<d-cite bibtex-key="llama3modelcard"></d-cite> and Mixtral<d-cite bibtex-key="jiang2024mixtral"></d-cite> are not publicly available and very little is known about how they were created.</p>
|
|
|
76 |
|
77 |
<p>Recently, we released <a href="https://huggingface.co/datasets/HuggingFaceFW/fineweb"><strong>π· FineWeb</strong></a>, a new, large-scale
|
78 |
(<strong>15-trillion tokens, 44TB disk space</strong>) dataset for LLM pretraining. FineWeb is derived from 96 <a href="https://commoncrawl.org/">CommonCrawl</a> snapshots and produces <strong>better-performing LLMs than other open pretraining datasets</strong>. To bring more clarity in machine learning and advance the open understanding of how to train good quality large language models, we decided to carefully document and ablate all of the design choices used in FineWeb, including in-depth investigations of deduplication and filtering strategies. The present long form report is a deep dive in how to create a large and high-quality web-scale dataset for LLM pretraining. The dataset it-self, π· FineWeb, is available <a href="https://huggingface.co/datasets/HuggingFaceFW/fineweb">here</a>.
|
@@ -86,7 +87,6 @@
|
|
86 |
|
87 |
<p><strong>TLDR:</strong> This blog covers a discussion on processing and evaluating data quality at scale, the π· FineWeb
|
88 |
recipe (listing and explaining all of our design choices), and the process followed to create its π FineWeb-Edu subset.</p>
|
89 |
-
<aside>For the best possible reading experience, we recommend not using a mobile phone.</aside>
|
90 |
|
91 |
<h2>What's web data</h2>
|
92 |
<h3>Finding the data</h3>
|
|
|
73 |
|
74 |
<p>The performance of a large language model (LLM) depends heavily on the quality and size of its pretraining dataset.
|
75 |
However, the pretraining datasets for state-of-the-art open LLMs like Llama 3<d-cite bibtex-key="llama3modelcard"></d-cite> and Mixtral<d-cite bibtex-key="jiang2024mixtral"></d-cite> are not publicly available and very little is known about how they were created.</p>
|
76 |
+
<aside>Reading time: 45 min. For the best reading experience, we recommend not using a mobile phone.</aside>
|
77 |
|
78 |
<p>Recently, we released <a href="https://huggingface.co/datasets/HuggingFaceFW/fineweb"><strong>π· FineWeb</strong></a>, a new, large-scale
|
79 |
(<strong>15-trillion tokens, 44TB disk space</strong>) dataset for LLM pretraining. FineWeb is derived from 96 <a href="https://commoncrawl.org/">CommonCrawl</a> snapshots and produces <strong>better-performing LLMs than other open pretraining datasets</strong>. To bring more clarity in machine learning and advance the open understanding of how to train good quality large language models, we decided to carefully document and ablate all of the design choices used in FineWeb, including in-depth investigations of deduplication and filtering strategies. The present long form report is a deep dive in how to create a large and high-quality web-scale dataset for LLM pretraining. The dataset it-self, π· FineWeb, is available <a href="https://huggingface.co/datasets/HuggingFaceFW/fineweb">here</a>.
|
|
|
87 |
|
88 |
<p><strong>TLDR:</strong> This blog covers a discussion on processing and evaluating data quality at scale, the π· FineWeb
|
89 |
recipe (listing and explaining all of our design choices), and the process followed to create its π FineWeb-Edu subset.</p>
|
|
|
90 |
|
91 |
<h2>What's web data</h2>
|
92 |
<h3>Finding the data</h3>
|
src/index.html
CHANGED
@@ -73,6 +73,7 @@
|
|
73 |
|
74 |
<p>The performance of a large language model (LLM) depends heavily on the quality and size of its pretraining dataset.
|
75 |
However, the pretraining datasets for state-of-the-art open LLMs like Llama 3<d-cite bibtex-key="llama3modelcard"></d-cite> and Mixtral<d-cite bibtex-key="jiang2024mixtral"></d-cite> are not publicly available and very little is known about how they were created.</p>
|
|
|
76 |
|
77 |
<p>Recently, we released <a href="https://huggingface.co/datasets/HuggingFaceFW/fineweb"><strong>π· FineWeb</strong></a>, a new, large-scale
|
78 |
(<strong>15-trillion tokens, 44TB disk space</strong>) dataset for LLM pretraining. FineWeb is derived from 96 <a href="https://commoncrawl.org/">CommonCrawl</a> snapshots and produces <strong>better-performing LLMs than other open pretraining datasets</strong>. To bring more clarity in machine learning and advance the open understanding of how to train good quality large language models, we decided to carefully document and ablate all of the design choices used in FineWeb, including in-depth investigations of deduplication and filtering strategies. The present long form report is a deep dive in how to create a large and high-quality web-scale dataset for LLM pretraining. The dataset it-self, π· FineWeb, is available <a href="https://huggingface.co/datasets/HuggingFaceFW/fineweb">here</a>.
|
@@ -86,7 +87,6 @@
|
|
86 |
|
87 |
<p><strong>TLDR:</strong> This blog covers a discussion on processing and evaluating data quality at scale, the π· FineWeb
|
88 |
recipe (listing and explaining all of our design choices), and the process followed to create its π FineWeb-Edu subset.</p>
|
89 |
-
<aside>For the best possible reading experience, we recommend not using a mobile phone.</aside>
|
90 |
|
91 |
<h2>What's web data</h2>
|
92 |
<h3>Finding the data</h3>
|
|
|
73 |
|
74 |
<p>The performance of a large language model (LLM) depends heavily on the quality and size of its pretraining dataset.
|
75 |
However, the pretraining datasets for state-of-the-art open LLMs like Llama 3<d-cite bibtex-key="llama3modelcard"></d-cite> and Mixtral<d-cite bibtex-key="jiang2024mixtral"></d-cite> are not publicly available and very little is known about how they were created.</p>
|
76 |
+
<aside>Reading time: 45 min. For the best reading experience, we recommend not using a mobile phone.</aside>
|
77 |
|
78 |
<p>Recently, we released <a href="https://huggingface.co/datasets/HuggingFaceFW/fineweb"><strong>π· FineWeb</strong></a>, a new, large-scale
|
79 |
(<strong>15-trillion tokens, 44TB disk space</strong>) dataset for LLM pretraining. FineWeb is derived from 96 <a href="https://commoncrawl.org/">CommonCrawl</a> snapshots and produces <strong>better-performing LLMs than other open pretraining datasets</strong>. To bring more clarity in machine learning and advance the open understanding of how to train good quality large language models, we decided to carefully document and ablate all of the design choices used in FineWeb, including in-depth investigations of deduplication and filtering strategies. The present long form report is a deep dive in how to create a large and high-quality web-scale dataset for LLM pretraining. The dataset it-self, π· FineWeb, is available <a href="https://huggingface.co/datasets/HuggingFaceFW/fineweb">here</a>.
|
|
|
87 |
|
88 |
<p><strong>TLDR:</strong> This blog covers a discussion on processing and evaluating data quality at scale, the π· FineWeb
|
89 |
recipe (listing and explaining all of our design choices), and the process followed to create its π FineWeb-Edu subset.</p>
|
|
|
90 |
|
91 |
<h2>What's web data</h2>
|
92 |
<h3>Finding the data</h3>
|