Datasets:
text
stringlengths 1
56
|
---|
to
|
in
|
no
|
toward
|
actually
|
on
|
is
|
allele
|
early
|
straight
|
want
|
n
|
a
|
com
|
half
|
Love
|
thunderous
|
your
|
mostly
|
precisely
|
made
|
vegan
|
Doctors
|
more
|
have
|
answers
|
region
|
through
|
exposure
|
files
|
as
|
of
|
first
|
came
|
ago
|
draft
|
guys
|
he
|
same
|
bridge
|
Shrimp
|
end
|
best
|
commander
|
come
|
urging
|
over
|
sparked
|
party
|
Kurdish
|
last
|
worked
|
C
|
for
|
oil
|
patriotic
|
But
|
figures
|
Russian
|
fan
|
here
|
buds
|
humped
|
like
|
B
|
layout
|
hearing
|
match
|
opened
|
body
|
gaming
|
Service
|
grown
|
parent
|
expected
|
after
|
al
|
rapid
|
dissertations
|
would
|
compiled
|
whether
|
hands
|
view
|
happening
|
Windows
|
some
|
levels
|
an
|
proposal
|
entered
|
parents
|
did
|
Armstrong
|
men
|
still
|
that
|
Cambodia
|
More
|
final
|
WikiSpell
Description
This dataset is a custom implementation of the WikiSpell dataset introduced in Character-Aware Models Improve Visual Text Rendering by Liu et al. (2022).
Similarly to the original WikiSpell dataset, the training set is composed of 5000 words taken uniformly from the 50% least common Wiktionary words (taken from this Wiktionary extraction), and 5000 words sampled according to their frequencies taken from the 50% most common Wiktionary words.
The validation and test are splitted in 5 sets, sampled depending on their frequency in the corpus:
- 1% most common words
- 1 - 10% most common words
- 10 - 20% most common words
- 20 - 30% most common words
- 50% least common words
Contrary to the original WikiSpell dataset, we compute the frequency of the words using the first 100k sentences from OpenWebText (Skylion007/openwebtext) instead of mC4.
Usage
This dataset is used for testing spelling in Large Language Models. To do so, the labels should be computed like in the following snippet:
sample = ds["train"][0]
label = " ".join(sample["text"])
The labels are not included in the dataset files directly.
Citation
Please cite the original paper introducing WikiSpell if you're using this dataset:
@inproceedings{liu-etal-2023-character,
title = "Character-Aware Models Improve Visual Text Rendering",
author = "Liu, Rosanne and
Garrette, Dan and
Saharia, Chitwan and
Chan, William and
Roberts, Adam and
Narang, Sharan and
Blok, Irina and
Mical, Rj and
Norouzi, Mohammad and
Constant, Noah",
booktitle = "Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = jul,
year = "2023",
address = "Toronto, Canada",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.acl-long.900",
pages = "16270--16297",
}
- Downloads last month
- 91