Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
Dask
License:
koalazf99 commited on
Commit
cf32520
·
verified ·
1 Parent(s): 5bbc364

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -2
README.md CHANGED
@@ -32,7 +32,6 @@ MegaMath is the largest open math pre-training dataset to date, surpassing DeepS
32
  <div style="display: flex; justify-content: center; gap: 20px;">
33
  <img src="https://cdn-uploads.huggingface.co/production/uploads/628f6e5ab90dde28ef57d293/_yAulJQ1lcL56Dvqnxr4W.png" width="75%" />
34
  </div>
35
- ![image/png]()
36
 
37
  ## MegaMath Delivers with High Quality
38
  During development, we use extensive experiments to find optimal practice for text extraction, deduplication, fasttext training, etc. Training MegaMath data shows better performance than existing open datasets.
@@ -47,7 +46,7 @@ During development, we use extensive experiments to find optimal practice for te
47
  We also release two proof-of-concept models which is based on [Llama-3.2-1B](https://huggingface.co/meta-llama/Llama-3.2-1B) and [LLama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B).
48
  Training MegaMath on Llama-3.2-1B and LLama-3.2-3B brings about 15% ~ 20% performance boost on 10 downstream benchmarks, demonstrateing its high data quality.
49
  <div style="display: flex; justify-content: center; gap: 20px;">
50
- <img src="https://cdn-uploads.huggingface.co/production/uploads/628f6e5ab90dde28ef57d293/EIReQ8TIbyn1V3JfsEKiL.png" width="30%" />
51
  </div>
52
 
53
  ## Detailed Statistics
 
32
  <div style="display: flex; justify-content: center; gap: 20px;">
33
  <img src="https://cdn-uploads.huggingface.co/production/uploads/628f6e5ab90dde28ef57d293/_yAulJQ1lcL56Dvqnxr4W.png" width="75%" />
34
  </div>
 
35
 
36
  ## MegaMath Delivers with High Quality
37
  During development, we use extensive experiments to find optimal practice for text extraction, deduplication, fasttext training, etc. Training MegaMath data shows better performance than existing open datasets.
 
46
  We also release two proof-of-concept models which is based on [Llama-3.2-1B](https://huggingface.co/meta-llama/Llama-3.2-1B) and [LLama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B).
47
  Training MegaMath on Llama-3.2-1B and LLama-3.2-3B brings about 15% ~ 20% performance boost on 10 downstream benchmarks, demonstrateing its high data quality.
48
  <div style="display: flex; justify-content: center; gap: 20px;">
49
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/628f6e5ab90dde28ef57d293/EIReQ8TIbyn1V3JfsEKiL.png" width="50%" />
50
  </div>
51
 
52
  ## Detailed Statistics