|
{ |
|
"paper_id": "S01-1003", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T15:35:29.629624Z" |
|
}, |
|
"title": "Dutch Word Sense Disambiguation: Data and Preliminary Results", |
|
"authors": [ |
|
{ |
|
"first": "Iris", |
|
"middle": [], |
|
"last": "Hendrickx", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Tilburg University", |
|
"location": { |
|
"postCode": "NL-5000 LE", |
|
"settlement": "Tilburg", |
|
"country": "The Netherlands" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Antal", |
|
"middle": [], |
|
"last": "Van Den Bosch", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Tilburg University", |
|
"location": { |
|
"postCode": "NL-5000 LE", |
|
"settlement": "Tilburg", |
|
"country": "The Netherlands" |
|
} |
|
}, |
|
"email": "" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "We describe the Dutch word sense disambiguation data submitted to SENSEVAL-2, and give preliminary results on the data using a WSD system based on memory-based learning and statistical keyword selection.", |
|
"pdf_parse": { |
|
"paper_id": "S01-1003", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "We describe the Dutch word sense disambiguation data submitted to SENSEVAL-2, and give preliminary results on the data using a WSD system based on memory-based learning and statistical keyword selection.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Solving lexical ambiguity, or word sense disambiguation (WSD), is an important task in Natural Language Processing systems. Much like syntactic word-class disambiguation, it is not a end in itself, but rather a subtask of other natural language processing tasks (Kilgarriff and Rozenzweig, 2000) . The problem is far from solved, and research and competition in the development of WSD systems in isolation is merited, preferrably on many different languages and genres.", |
|
"cite_spans": [ |
|
{ |
|
"start": 262, |
|
"end": 295, |
|
"text": "(Kilgarriff and Rozenzweig, 2000)", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "Here we introduce the first electronic Dutch word-sense annotated corpus, that was collected under a sociolinguistic research project (Schrooten and Vermeer, 1994) , and was kindly donated by the team coordinators to the WSD systems community. In this paper we describe the original data and the preprocessing steps that were applied to it before submission to the SENSEVAL-2, in Section 2. We also present the first, preliminary, results obtained with MBWSD-D, the Memory-Based Word-Sense Disambiguation system for Dutch, that uses statistical keyword selection, in Section 3.", |
|
"cite_spans": [ |
|
{ |
|
"start": 134, |
|
"end": 163, |
|
"text": "(Schrooten and Vermeer, 1994)", |
|
"ref_id": "BIBREF4" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "The Dutch WSD corpus was built as a part of a sociolinguistic project, led by Walter Schrooten and Anne Vermeer (1994) , on the active vocabulary of children in the age of 4 to 12 in the Netherlands. The aim of developing the corpus was to have a realistic wordlist of the most common words used at elementary schools. This wordlist was further used in the study to make literacy tests, including tests how many senses of ambiguous words were known by children of different ages. The corpus consists of texts of 102 illustrated children books in the age range of 4 to 12. Each word in these texts is manually annotated with its appropriate sense. The data was annotated by six persons who all processed a different part of the data.", |
|
"cite_spans": [ |
|
{ |
|
"start": 85, |
|
"end": 118, |
|
"text": "Schrooten and Anne Vermeer (1994)", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data: The Dutch child book corpus", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Each word in the dataset has a nonhierarchical, symbolic sense tag, realised as a mnemonic description of the specific meaning the word has in the sentence, often using a related term. As there was no gold standard sense set of Dutch available, Schrooten and Vermeer have made their own set of senses.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data: The Dutch child book corpus", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Sense tags consist of the word's lemma and a sense description of one or two words ( dro-gen_nat) or a reference of the grammatical category (fiets_N, fietsen_ V). Verbs have as their tag their lemma and often a reference to their function in the sentence (is/ zijn_kww). When a word has only one sense, this is represented with a simple\"=\". Names and sound imitations also have \"=\" as their sense tag.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data: The Dutch child book corpus", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The dataset also contains senses that span over multiple words. These multi-word expressions cover idiomatic expressions, sayings, proverbs, and strong collocations. Each word in the corpus that is part of such multi-word expression has as its meaning the atomic meaning of the expression.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data: The Dutch child book corpus", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "These are two example sentences in the corpus:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data: The Dutch child book corpus", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "\"/= het/heLiidwoord raadsel/= vanjvan_prepositie de/=. verdwenenjverdwijnen regenboog/= kan /ku n nen JTiogelijkheid aileen j a lleen_adv ,\" /= zeiden/zeggen_praten de/= koningenjkoning .J= toenjtoen_adv verklaar-denjverklaren_oorlog zej= elkaar/=de/= oorlog/= .J=", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data: The Dutch child book corpus", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The dataset needed some adaptations to make it fully usable for computational purposes. First, spelling and consistency errors have been corrected for most part, but in the data submitted to SENSEVAL-2, a certain amount of errors is still present. Second, in Dutch, prepositions are often combined with verbs as particles and these combinations have other meanings than the two separate words. Unfortunately the annotations of these cases were rather inconsistent and for that reason it was decided to give all prepositions the same sense tag \"/prepositie\" after their lemma.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data: The Dutch child book corpus", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The dataset consists of approximately 150,000 tokens (words and punctuation tokens) and about 10,000 different word forms. Nine thousand of these words have only one sense, leaving a thousand word types to disambiguate. These ambiguous types account for 54 % of the tokens in the corpus. The basic numbers can be found in Table 1 .", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 322, |
|
"end": 329, |
|
"text": "Table 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Data: The Dutch child book corpus", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "For the SENSEVAL-2 competition, the dataset was divided in two parts. The training set consisted of 76 books and approximately 115.000 words. The test set consisted of the remaining 26 books and had about 38.000 words.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data: The Dutch child book corpus", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "We first describe the representation of the corpus data in examples presented to a memory-14 based learner in Subsection 3.1. We then describe the architecture of the system in Subsection 3.2, and we then present its preliminary results in Subsection 4.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The MBWSD-D system and preliminary results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "3.1 Representation: Local and keyword features As a general idea, disambiguation information is assumed to be present in the not-too-distant context of ambiguous words; the present instantiation of MBWSD-D limits this to the sentence the ambiguous word occurs in. Sentences are not represented as is, but rather as limited sets of features expected to give salient information about which sense of the word applies.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The MBWSD-D system and preliminary results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "The first source of useful disambiguation information can be found immediately adjacent to the ambiguous word.It has been found that a four-word window, two words before the target word and two words after gives good results; cf. (Veenstra et al., 2000) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 230, |
|
"end": 253, |
|
"text": "(Veenstra et al., 2000)", |
|
"ref_id": "BIBREF5" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The MBWSD-D system and preliminary results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Second, information about the grammatical category of the target word and its direct context words can also be valuable. Consequently, each sentence of the Dutch corpus was tagged and the part-of-speech (POS) tags of the word and its direct context (two left, two right) are included in the representation of the sentence. Part-of-speech tagging was done with the Memory Based Tagger (Daelemans et al., 1996) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 384, |
|
"end": 408, |
|
"text": "(Daelemans et al., 1996)", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The MBWSD-D system and preliminary results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Third, informative words in the context ('keywords') are detected based on the statistical chi-squared test. Chi-square estimates the significance, or degree of surprise, of the number of keyword occurrences with respect to the expected number of occurrences (apriori probability):", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The MBWSD-D system and preliminary results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "x2 = t (ik -ek) 2 k=l ek (1)", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The MBWSD-D system and preliminary results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "where fi is the keyword frequency and ei is the expected frequency. fi is the word frequency and ei is the expected word frequency. The expected frequency of the keyword is given in equation 3.1. It must be noted that the Chi-Square method cannot be considered reliable when the expected frequency has a value below 5: ei = Uwi/ fw) * ]k, where fi is the frequency the ambiguous word w of sensei, fw is the frequency of word wand fk is the frequency of the keyword.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The MBWSD-D system and preliminary results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "The number of occurrences of a very good keyword will have a strong deviation of its expected number of occurrences divided over the senses. The expected probability with respect to all senses can be seen as a distribution of the keyword. A good keyword is a word that differs from the expected distribution and always cooccurs with a certain sense, or never co-occurs with a certain sense.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The MBWSD-D system and preliminary results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "In sum, a representation of an instance of an. ambiguous word consists of the two words before the target word, two words after the word, the POS tags of these words and of the target word itself, a number of selected keywords, and of course the annotated sense of the word as the class label.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The MBWSD-D system and preliminary results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Following the example of ILK's previous wordsense disambiguation system for English (Veenstra et al., 2000) , it was decided to use word experts. Berleant (Berleant, 1995) defines a word expert as follows: \"A word expert is a small expert system-like module for processing a particular word based on other words in its vicinity\" (1995, p.1). Word experts are common in the field of word sense disambiguation, because words are very different from each other. Words all have different numbers of senses, different frequencies and. need different information sources for di~ambiguation. With word experts, each word can be treated with its own optimal method.", |
|
"cite_spans": [ |
|
{ |
|
"start": 84, |
|
"end": 107, |
|
"text": "(Veenstra et al., 2000)", |
|
"ref_id": "BIBREF5" |
|
}, |
|
{ |
|
"start": 155, |
|
"end": 171, |
|
"text": "(Berleant, 1995)", |
|
"ref_id": "BIBREF0" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "System architecture", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "Making word experts for every ambiguous word may not be useful because many words occur only a few times in the corpus. It was decided to create word experts for wordforms with a threshold of minimal 10 occurrences in the training set. There are 524 of such words in the training set. 10 is a rather low threshold, but many words can be easily disambiguated by knowledge a single feature value, such as of their part-of-speech tag.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "System architecture", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "The software for emulating memory-based learning used in this research is TiMBL (Tilburg Memory-Based Learner). TiMBL (Daelemans et al., 2001 ) is a software package developed by the ILK research group at Tilburg University. TiMBL implements several memory-based classifiers. In essence, memory-based classifiers use stored classified examples to disambiguate new examples.", |
|
"cite_spans": [ |
|
{ |
|
"start": 118, |
|
"end": 141, |
|
"text": "(Daelemans et al., 2001", |
|
"ref_id": "BIBREF2" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "System architecture", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "For each word a TiMBL word expert was trained on that portion of the training corpus that consisted of sentence representations containing that word. TiMBL was trained 300 times, each time with another combination of parameters. Each of these training sessions was evaluated with leave-one-out cross validation (Weiss and Kulikowski, 1991) to select the optimal TiMBL setting for a particular word, to be used eventually for classifying the test material.", |
|
"cite_spans": [ |
|
{ |
|
"start": 311, |
|
"end": 339, |
|
"text": "(Weiss and Kulikowski, 1991)", |
|
"ref_id": "BIBREF6" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "System architecture", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "For each word expert a total of 300 experiments were performed, each with another combination of parameter settings. In this study the following options were used ( cf. (Daelemans et al.,. 2001) for first pointers to descriptions of these metrics and functions): distance-weighted voting : (1) all neighbors have equal weight;", |
|
"cite_spans": [ |
|
{ |
|
"start": 169, |
|
"end": 194, |
|
"text": "(Daelemans et al.,. 2001)", |
|
"ref_id": "BIBREF2" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "System architecture", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "(2) Inverse Distance weighting;", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "System architecture", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "(3) Inverse Linear weighting feature weighting : (1) no weighting; 2Gain Ratio;", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "System architecture", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "(3) Information Gain; ( 4) Chi Square; (5) Shared Variance similarity metric : (1) Overlap metric; (2) MVDM number of nearest neighbours : 1, 3, 5, 7, 9, 11, 15, 25, 45, and 75 The last step for each word expert was to test the optimal settings on the test set. To evaluate the results, described in the next Section, theresults were compared with a baseline score. The baseline was to select for each word the most frequent sense.", |
|
"cite_spans": [ |
|
{ |
|
"start": 142, |
|
"end": 176, |
|
"text": "3, 5, 7, 9, 11, 15, 25, 45, and 75", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "System architecture", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "The top line of Table 2 shows the mean score of all the word experts together on the test set. The score of the word experts on the test set, 84.1%, is generously higher than the baseline score of 7 4.1%. These are the results of the word experts only; the second row also includes the best-guess outputs for the lowerfrequency words, lowering the system's performance slightly.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 16, |
|
"end": 23, |
|
"text": "Table 2", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "The same results, now split on the frequency of the words in the training set, can be seen in We can also calculate the score on all the words in the text, including the unambiguous words, to give an impression of the overall performance. The unambiguous words are given a score of 100%, because the task was to disambiguate the ambiguous words. It might be useful for a disambiguation system to tag unambiguous words with their lemma, but the kind of tagging this is not of interest in our task. The third row of Table 2 shows the results on all words in which the system was applied with a threshold of 10: The system scores 4 % higher than the baseline.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 514, |
|
"end": 521, |
|
"text": "Table 2", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "This paper introduced a Dutch child book corpus, generously donated to the WSD community by the team leaders of the sociolinguistic project that produced the corpus. The data is annotated with a non-hierarchical mnemonic sense inventory. The data has been cleaned up and split for the SENSEVAL-2 competition.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Discussion", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "The data provides an arguably interesting case of a \"fiat\" semantic tagging, where there is obviously no gain from a governing wordnet, but alternatively it is not negatively biased by an inappropriate or badly-structured wordnet either. Learnability results are therefore an interesting baseline to beat when the data would be annotated with a Dutch wordnet.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Discussion", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "The system applied to the data as a first indication of its complexity and learnability, consisted of an ensemble of word experts trained to disambiguate particular ambiguous word forms. The score of the system on the 16686 ambiguous words in the test set was 83.8% compared to a baseline score of 7 4.6%. On free held out text the system achieved a result of 92.9%; 4% over the baseline of 88.8%, or in other words yielding an error reduction of about 37%. These absolute and relative figures are roughly comparable to performances of other systems on other data, indicating at least that the data represents learnability properties typical for the WSD area.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Discussion", |
|
"sec_num": "5" |
|
} |
|
], |
|
"back_matter": [], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "Engineering word-experts for word disambiguation", |
|
"authors": [ |
|
{ |
|
"first": "D", |
|
"middle": [], |
|
"last": "Berleant", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1995, |
|
"venue": "Natural Language Engineering", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "339--362", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "D. Berleant. 1995. Engineering word-experts for word disambiguation. Natural Language Engineering, pages 339-362.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Part-of-speech tagging of dutch with mbt, a memory-based tagger generator", |
|
"authors": [ |
|
{ |
|
"first": "W", |
|
"middle": [], |
|
"last": "Daelemans", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Zavrel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Berek", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1996, |
|
"venue": "Congresboek van de Interdisciplinaire Onderzoeksconferentie Informatiewetenchap", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "W. Daelemans, J. Zavrel, and P. Berek. 1996. Part-of-speech tagging of dutch with mbt, a memory-based tagger generator. In Congres- boek van de Interdisciplinaire Onderzoekscon- ferentie Informatiewetenchap.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Timbl: Tilburg memory based learner", |
|
"authors": [ |
|
{ |
|
"first": "W", |
|
"middle": [], |
|
"last": "Daelemans", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Zavrel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "K", |
|
"middle": [], |
|
"last": "Van Der Sloot", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Van Den", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Bosch", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2001, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "W. Daelemans, J. Zavrel, K. van der Sloot, and A. van den Bosch. 2001. Timbl: Tilburg memory based learner, version 4.0, reference guide. Technical report, Tilburg University.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Framework and results for english senseval. Computers and the Humanities", |
|
"authors": [ |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Kilgarriff", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Rozenzweig", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "", |
|
"volume": "34", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "A. Kilgarriff and J. Rozenzweig. 2000. Frame- work and results for english senseval. Com- puters and the Humanities, 34.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Woorden in het basisonderwijs. 15.000 woorden aangeboden aan leerlingen. TUP(Studies in meertaligheid 6)", |
|
"authors": [ |
|
{ |
|
"first": "W", |
|
"middle": [], |
|
"last": "Schrooten", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Vermeer", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1994, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "W. Schrooten and A. Vermeer. 1994. Woorden in het basisonderwijs. 15.000 woorden aange- boden aan leerlingen. TUP(Studies in meer- taligheid 6).", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Memory-based word sense disambiguation", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Veenstra", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Van Den", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Bosch", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "W", |
|
"middle": [], |
|
"last": "Buchholz", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Daelemans", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Zavrel", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "Computers and the Humanities", |
|
"volume": "34", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "J. Veenstra, A. van den Bosch, S. Buch- holz, W. Daelemans, and J. Zavrel. 2000. Memory-based word sense disambiguation. Computers and the Humanities, 34.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "computer systems that learn", |
|
"authors": [ |
|
{ |
|
"first": "S", |
|
"middle": [], |
|
"last": "Weiss", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "C", |
|
"middle": [], |
|
"last": "Kulikowski", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1991, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "S. Weiss and C. Kulikowski. 1991. computer systems that learn. Morgan Kaufmann.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"TABREF1": { |
|
"html": null, |
|
"type_str": "table", |
|
"num": null, |
|
"text": "The first column shows the frequency groups, based on the word frequencies in the training set, the second the number of words in", |
|
"content": "<table><tr><td>test selection</td><td/><td>#words</td><td colspan=\"2\">baseline system</td></tr><tr><td colspan=\"2\">word-expert words</td><td>15365</td><td>74.1</td><td>84.1</td></tr><tr><td colspan=\"2\">all ambiguous words</td><td>16686</td><td>74.6</td><td>83.8</td></tr><tr><td>all words</td><td/><td>37770</td><td>88.8</td><td>92.9</td></tr><tr><td colspan=\"5\">Tabl~ 2: Summary of results on test material</td></tr><tr><td colspan=\"5\">the test set, and the third column shows the</td></tr><tr><td colspan=\"5\">mean score of the WSD system. The scores</td></tr><tr><td colspan=\"5\">tend to get better as the frequency goes up, ex-</td></tr><tr><td colspan=\"5\">cept for the group of 40-49, which has the lowest</td></tr><tr><td colspan=\"5\">score of all. Note that the baseline score of the</td></tr><tr><td colspan=\"5\">group of words with a frequency below 10 is rel-</td></tr><tr><td colspan=\"2\">atively high: 80.5%.</td><td/><td/></tr><tr><td colspan=\"5\">frequency #words baseline system</td></tr><tr><td><10</td><td colspan=\"2\">1321 -</td><td/><td>80.5</td></tr><tr><td>10-19</td><td colspan=\"2\">868 63.0</td><td/><td>76.8</td></tr><tr><td>20-29</td><td colspan=\"2\">644 70.3</td><td/><td>79.5</td></tr><tr><td>30-39</td><td colspan=\"2\">503 75.9</td><td/><td>83.3</td></tr><tr><td>40-49</td><td colspan=\"2\">390 66.7</td><td/><td>75.9</td></tr><tr><td>50-99</td><td colspan=\"2\">1873 73.7</td><td/><td>85.4</td></tr><tr><td>100-199</td><td colspan=\"2\">2289 77.7</td><td/><td>83.1</td></tr><tr><td>> 200</td><td colspan=\"2\">8798 74.6</td><td/><td>85.6</td></tr><tr><td>> 100</td><td colspan=\"2\">10995 75.3</td><td/><td>85.1</td></tr></table>" |
|
}, |
|
"TABREF2": { |
|
"html": null, |
|
"type_str": "table", |
|
"num": null, |
|
"text": "Results divided into frequency groups", |
|
"content": "<table/>" |
|
} |
|
} |
|
} |
|
} |