Datasets:
Tasks:
Token Classification
Modalities:
Text
Formats:
parquet
Languages:
Thai
Size:
100K - 1M
Tags:
word-tokenization
License:
Update files from the datasets library (from 1.3.0)
Browse filesRelease notes: https://github.com/huggingface/datasets/releases/tag/1.3.0
README.md
CHANGED
@@ -43,6 +43,7 @@ task_ids:
|
|
43 |
- [Dataset Curators](#dataset-curators)
|
44 |
- [Licensing Information](#licensing-information)
|
45 |
- [Citation Information](#citation-information)
|
|
|
46 |
|
47 |
## Dataset Description
|
48 |
|
@@ -181,3 +182,7 @@ Character type features:
|
|
181 |
year={2009}
|
182 |
}
|
183 |
```
|
|
|
|
|
|
|
|
|
|
43 |
- [Dataset Curators](#dataset-curators)
|
44 |
- [Licensing Information](#licensing-information)
|
45 |
- [Citation Information](#citation-information)
|
46 |
+
- [Contributions](#contributions)
|
47 |
|
48 |
## Dataset Description
|
49 |
|
|
|
182 |
year={2009}
|
183 |
}
|
184 |
```
|
185 |
+
|
186 |
+
### Contributions
|
187 |
+
|
188 |
+
Thanks to [@cstorm125](https://github.com/cstorm125) for adding this dataset.
|