Update README.md
Browse files
README.md
CHANGED
@@ -18,9 +18,9 @@ dataset_info:
|
|
18 |
- name: sha256
|
19 |
dtype: string
|
20 |
- name: original_height
|
21 |
-
dtype:
|
22 |
- name: original_width
|
23 |
-
dtype:
|
24 |
---
|
25 |
|
26 |
# Dataset Card for DataComp-12M
|
@@ -32,7 +32,7 @@ We distribute the image url-text samples and metadata under a standard Creative
|
|
32 |
Image-text models trained on DataComp-12M are significantly better than on CC-12M/YFCC-15M as well as DataComp-Small/Medium.
|
33 |
|
34 |
DataComp-12M was introduced in [MobileCLIP paper](https://arxiv.org/abs/2311.17049) and along with the reinforced dataset [DataCompDR-12M](https://huggingface.co/datasets/apple/DataCompDR-12M).
|
35 |
-
The UIDs per shards match between [mlfoundations/DataComp-12M](https://huggingface.co/datasets/mlfoundations/DataComp-12M) and [apple/DataCompDR-12M](https://huggingface.co/datasets/apple/DataCompDR-12M)
|
36 |
|
37 |
## Terms and Conditions
|
38 |
|
@@ -41,7 +41,7 @@ We have terms of service that are similar to those adopted by HuggingFace (https
|
|
41 |
|
42 |
## Citation
|
43 |
|
44 |
-
**[
|
45 |
Gadre, Samir Yitzhak, et al.
|
46 |
```
|
47 |
@article{gadre2024datacomp,
|
|
|
18 |
- name: sha256
|
19 |
dtype: string
|
20 |
- name: original_height
|
21 |
+
dtype: int32
|
22 |
- name: original_width
|
23 |
+
dtype: int32
|
24 |
---
|
25 |
|
26 |
# Dataset Card for DataComp-12M
|
|
|
32 |
Image-text models trained on DataComp-12M are significantly better than on CC-12M/YFCC-15M as well as DataComp-Small/Medium.
|
33 |
|
34 |
DataComp-12M was introduced in [MobileCLIP paper](https://arxiv.org/abs/2311.17049) and along with the reinforced dataset [DataCompDR-12M](https://huggingface.co/datasets/apple/DataCompDR-12M).
|
35 |
+
The UIDs per shards match between [mlfoundations/DataComp-12M](https://huggingface.co/datasets/mlfoundations/DataComp-12M) and [apple/DataCompDR-12M](https://huggingface.co/datasets/apple/DataCompDR-12M).
|
36 |
|
37 |
## Terms and Conditions
|
38 |
|
|
|
41 |
|
42 |
## Citation
|
43 |
|
44 |
+
**[DataComp: In search of the next generation of multimodal datasets](https://arxiv.org/abs/2304.14108). (NeurIPS 2024)**
|
45 |
Gadre, Samir Yitzhak, et al.
|
46 |
```
|
47 |
@article{gadre2024datacomp,
|