|
{ |
|
"paper_id": "S07-1016", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T15:22:43.103282Z" |
|
}, |
|
"title": "SemEval-2007 Task 17: English Lexical Sample, SRL and All Words", |
|
"authors": [ |
|
{ |
|
"first": "Sameer", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Pradhan", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "BBN Technologies", |
|
"location": { |
|
"postCode": "02138", |
|
"settlement": "Cambridge", |
|
"region": "MA" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Edward", |
|
"middle": [], |
|
"last": "Loper", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "University of Pennsylvania", |
|
"location": { |
|
"postCode": "19104", |
|
"settlement": "Philadelphia", |
|
"region": "PA" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Dmitriy", |
|
"middle": [], |
|
"last": "Dligach", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "University of Colorado", |
|
"location": { |
|
"postCode": "80303", |
|
"settlement": "Boulder", |
|
"region": "CO" |
|
} |
|
}, |
|
"email": "" |
|
}, |
|
{ |
|
"first": "Martha", |
|
"middle": [], |
|
"last": "Palmer", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "University of Colorado", |
|
"location": { |
|
"postCode": "80303", |
|
"settlement": "Boulder", |
|
"region": "CO" |
|
} |
|
}, |
|
"email": "" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "This paper describes our experience in preparing the data and evaluating the results for three subtasks of SemEval-2007 Task-17-Lexical Sample, Semantic Role Labeling (SRL) and All-Words respectively. We tabulate and analyze the results of participating systems.", |
|
"pdf_parse": { |
|
"paper_id": "S07-1016", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "This paper describes our experience in preparing the data and evaluating the results for three subtasks of SemEval-2007 Task-17-Lexical Sample, Semantic Role Labeling (SRL) and All-Words respectively. We tabulate and analyze the results of participating systems.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Correctly disambiguating words (WSD), and correctly identifying the semantic relationships between those words (SRL), is an important step for building successful natural language processing applications, such as text summarization, question answering, and machine translation. SemEval-2007 Task-17 (English Lexical Sample, SRL and All-Words) focuses on both of these challenges, WSD and SRL, using annotated English text taken from the Wall Street Journal and the Brown Corpus. It includes three subtasks: i) the traditional All-Words task comprising fine-grained word sense disambiguation using a 3,500 word section of the Wall Street Journal, annotated with WordNet 2.1 sense tags, ii) a Lexical Sample task for coarse-grained word sense disambiguation on a selected set of lexemes, and iii) Semantic Role Labeling, using two different types of arguments, on the same subset of lexemes.", |
|
"cite_spans": [ |
|
{ |
|
"start": 278, |
|
"end": 342, |
|
"text": "SemEval-2007 Task-17 (English Lexical Sample, SRL and All-Words)", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "In this task we measure the ability of systems to identify the correct fine-grained WordNet 2.1 word sense for all the verbs and head words of their arguments.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Word Sense Disambiguation 2.1 English fine-grained All-Words", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "We began by selecting three articles wsj 0105.mrg (on homelessness), wsj 0186.mrg (about a book on corruption), and wsj 0239.mrg (about hot-air ballooning) from a section of the WSJ corpus that has been Treebanked and PropBanked. All instances of verbs were identified using the Treebank part-of-speech tags, and also the headwords of their noun arguments (using the PropBank and standard headword rules). The locations of the sentences containing them as well as the locations of the verbs and the nouns within these sentences were recorded for subsequent sense-annotation. A total of 465 lemmas were selected from about 3500 words of text.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data Preparation", |
|
"sec_num": "2.1.1" |
|
}, |
|
{ |
|
"text": "We use a tool called STAMP written by Benjamin Snyder for sense-annotation of these instances. STAMP accepts a list of pointers to the instances that need to be annotated. These pointers consist of the name of the file where the instance is located, the sentence number of the instance, and finally, the word number of the ambiguous word within that sentence. These pointers were obtained as described in the previous paragraph. STAMP also requires a sense inventory, which must be stored in XML format. This sense inventory was obtained by querying WordNet 2.1 and storing the output as a set of XML files (one for each word to be annotated) prior to tagging. STAMP works by displaying to the user the sentence to be annotated with the target word highlighted along with the previous and the following sentences and the senses from the sense inventory. The user can select one of the senses and move on to the next instance.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data Preparation", |
|
"sec_num": "2.1.1" |
|
}, |
|
{ |
|
"text": "Two linguistics students annotated the words with WordNet 2.1 senses. Our annotators examined each instance upon which they disagreed and resolved their disagreements. Finally, we converted the resulting data to the Senseval format. For this dataset, we got an inter-annotator agreement (ITA) of 72% on verbs and 86% for nouns.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data Preparation", |
|
"sec_num": "2.1.1" |
|
}, |
|
{ |
|
"text": "A total of 14 systems were evaluated on the All Words task. These results are shown in Table 1 . We used the standard Senseval scorer -scorer2 1 to score the systems. All the F-scores 2 in this table as well as other tables in this paper are accompanied by a 95% confidence interval calculated using the bootstrap resampling procedure.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 87, |
|
"end": 94, |
|
"text": "Table 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "2.1.2" |
|
}, |
|
{ |
|
"text": "It is quite well accepted at this point that it is difficult to achieve high inter-annotator agreement on the fine-grained WordNet style senses, and without a corpus with high annotator agreement, automatic learning methods cannot perform at a level that would be acceptable for a downstream application. OntoNotes (Hovy et al., 2006 ) is a project that has annotated several layers of semantic information -including word senses, at a high inter-annotator agreement of over 90%. Therefore we decided to use this data for the lexical sample task.", |
|
"cite_spans": [ |
|
{ |
|
"start": 315, |
|
"end": 333, |
|
"text": "(Hovy et al., 2006", |
|
"ref_id": "BIBREF5" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "OntoNotes English Lexical Sample WSD", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "All the data for this task comes from the 1M word WSJ Treebank. For the convenience of the participants who wanted to use syntactic parse information as features using an off-the-shelf syntactic parser, we decided to compose the training data of Sections 02-21. For the test sets, we use data from Sections tem that attempts all the words, both Precision and Recall are the same. Since a few systems had missing answers, they got different Precision and Recall scores. Therefore, for ranking purposes, we consolidated them into an F-score. ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data", |
|
"sec_num": "2.2.1" |
|
}, |
|
{ |
|
"text": "A total of 13 systems were evaluated on the Lexical Sample task. Table 3 shows the Precision/Recall for all these systems. The same scoring software was used to score this task as well.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 65, |
|
"end": 72, |
|
"text": "Table 3", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "For the all words task, the baseline performance using the most frequent WordNet sense for the lemmas is 51.4. The top-performing system was a supervised system that used a Maximum Entropy classifier, and got a Precision/Recall of 59.1% -about 8 points higher than the baseline. Since the coarse and fine-grained disambiguation tasks have been part of the two previous Senseval competitions, and we happen to have access to that data, we can take this opportunity to look at the disambiguation performance trend. Although different test sets were used for every evaluation, we can get a rough indication of the trend. For the fine-grained All Words sense tagging task, which has always used WordNet, the system performance has ranged from our 59% to 65.2 (Sen-seval3, (Decadt et al., 2004) ) to 69% (Seneval2, (Chklovski and Mihalcea, 2002) ). Because of time constraints on the data preparation, this year's task has proportionally more verbs and fewer nouns than previous All-Words English tasks, which may account for the lower scores.", |
|
"cite_spans": [ |
|
{ |
|
"start": 768, |
|
"end": 789, |
|
"text": "(Decadt et al., 2004)", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 810, |
|
"end": 840, |
|
"text": "(Chklovski and Mihalcea, 2002)", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Discussion", |
|
"sec_num": "2.2.3" |
|
}, |
|
{ |
|
"text": "As expected, the Lexical Sample task using coarse Table 3 : System Performance for the OntoNotes Lexical Sample task. Systems marked with an * were post-competition bug-fix submissions.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 50, |
|
"end": 57, |
|
"text": "Table 3", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Discussion", |
|
"sec_num": "2.2.3" |
|
}, |
|
{ |
|
"text": "grained senses provides consistently higher performance than previous more fine-grained Lexical Sample Tasks. The high scores here were foreshadowed in an evaluation involving a subset of the data last summer (Chen et al., 2006) . Note that the best system performance is now closely approaching the ITA for this data of over 90%. Table 4 shows the performance of the top 8 systems on all the individual verbs and nouns in the test set. Owing to space constraints we have removed some lemmas that have perfect or almost perfect accuracies. At the right are mentioned the average, minimum and maximum performances of the teams per lemma, and at the bottom are the average scores per lemma (without considering the lemma frequencies) and broken down by verbs and nouns. A gap of about 10 points between the verb and noun performance seems to indicate that in general the verbs were more difficult than the nouns. However, this might just be owing to this particular test sample having more verbs with higher perplexities, and maybe even ones that are indeed difficult to disambiguate -in spite of high human agreement. The hope is that better knowledge sources can overcome the gap still existing between the system performance and human agreement. Overall, however, this data indicates that the approach suggested by (Palmer, 2000) and that is being adopted in the ongoing OntoNotes project (Hovy et al., 2006) does result in higher system performance. Whether or not the more coarse-grained senses are effective in improving natural language processing applications remains to be seen.", |
|
"cite_spans": [ |
|
{ |
|
"start": 209, |
|
"end": 228, |
|
"text": "(Chen et al., 2006)", |
|
"ref_id": "BIBREF2" |
|
}, |
|
{ |
|
"start": 1316, |
|
"end": 1330, |
|
"text": "(Palmer, 2000)", |
|
"ref_id": "BIBREF10" |
|
}, |
|
{ |
|
"start": 1390, |
|
"end": 1409, |
|
"text": "(Hovy et al., 2006)", |
|
"ref_id": "BIBREF5" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 331, |
|
"end": 338, |
|
"text": "Table 4", |
|
"ref_id": "TABREF4" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Discussion", |
|
"sec_num": "2.2.3" |
|
}, |
|
{ |
|
"text": "Lemma S s T t 1 2 3 4 5 6 7 8 Average Min Max turn.v 13 8 340 62 58 61 40 55 52 53 27 44 49 27 61 go.v 12 6 244 61 64 69 38 66 43 46 31 39 49 31 69 come.v 10 9 186 43 49 46 56 60 37 23 23 49 43 23 60 set.v 9 5 174 42 62 50 52 57 50 57 36 50 52 36 62 hold.v 8 7 129 24 58 46 50 54 54 38 50 67 52 38 67 raise.v 7 6 147 34 50 44 29 26 44 26 24 12 32 12 50 work.v 7 5 230 43 74 65 65 65 72 67 46 65 65 46 74 keep.v 7 6 260 80 56 54 52 64 56 52 48 51 54 48 ", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 6, |
|
"end": 592, |
|
"text": "S s T t 1 2 3 4 5 6 7 8 Average Min Max turn.v 13 8 340 62 58 61 40 55 52 53 27 44 49 27 61 go.v 12 6 244 61 64 69 38 66 43 46 31 39 49 31 69 come.v 10 9 186 43 49 46 56 60 37 23 23 49 43 23 60 set.v 9 5 174 42 62 50 52 57 50 57 36 50 52 36 62 hold.v 8 7 129 24 58 46 50 54 54 38 50 67 52 38 67 raise.v 7 6 147 34 50 44 29 26 44 26 24 12 32 12 50 work.v 7 5 230 43 74 65 65 65 72 67 46 65 65 46 74 keep.v 7 6 260 80 56 54 52 64 56 52 48 51 54 48", |
|
"ref_id": "TABREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Discussion", |
|
"sec_num": "2.2.3" |
|
}, |
|
{ |
|
"text": "Subtask 2 evaluates Semantic Role Labeling (SRL) systems, where the goal is to locate the constituents which are arguments of a given verb, and to assign them appropriate semantic roles that describe how they relate to the verb. SRL systems are an important building block for many larger semantic systems. For example, in order to determine that question (1a) is answered by sentence (1b), but not by sentence (1c), we must determine the relationships between the relevant verbs (eat and feed) and their arguments.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Semantic Role Labeling", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "(1) a. What do lobsters like to eat? b. Recent studies have shown that lobsters primarily feed on live fish, dig for clams, sea urchins, and feed on algae and eel-grass. c. In the early 20th century, Mainers would only eat lobsters because the fish they caught was too valuable to eat themselves.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Semantic Role Labeling", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Traditionally, SRL systems have been trained on either the PropBank corpus (Palmer et al., 2005) -for two years, the CoNLL workshop (Carreras and M\u00e0rquez, 2004; Carreras and M\u00e0rquez, 2005) has made this their shared task, or the FrameNet corpus -Senseval-3 used this for their shared task (Litkowski, 2004) . However, there is still little consensus in the linguistics and NLP communities about what set of role labels are most appropriate. The PropBank corpus avoids this issue by using theoryagnostic labels (ARG0, ARG1, . . . , ARG5), and by defining those labels to have only verb-specific meanings. Under this scheme, PropBank can avoid making any claims about how any one verb's arguments relate to other verbs' arguments, or about general distinctions between verb arguments and adjuncts.", |
|
"cite_spans": [ |
|
{ |
|
"start": 75, |
|
"end": 96, |
|
"text": "(Palmer et al., 2005)", |
|
"ref_id": "BIBREF9" |
|
}, |
|
{ |
|
"start": 132, |
|
"end": 160, |
|
"text": "(Carreras and M\u00e0rquez, 2004;", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 161, |
|
"end": 188, |
|
"text": "Carreras and M\u00e0rquez, 2005)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 289, |
|
"end": 306, |
|
"text": "(Litkowski, 2004)", |
|
"ref_id": "BIBREF7" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Semantic Role Labeling", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "However, there are several limitations to this approach. The first is that it can be difficult to make inferences and generalizations based on role labels that are only meaningful with respect to a single verb. Since each role label is verb-specific, we can not confidently determine when two different verbs' arguments have the same role; and since no encoded meaning is associated with each tag, we can not make generalizations across verb classes. In contrast, the use of a shared set of role labels, such Table 5 : System performance on PropBank arguments.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 509, |
|
"end": 516, |
|
"text": "Table 5", |
|
"ref_id": "TABREF6" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Semantic Role Labeling", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "as VerbNet roles, would facilitate both inferencing and generalization. VerbNet has more traditional labels such as Agent, Patient, Theme, Beneficiary, etc. (Kipper et al., 2006) . Therefore, we chose to annotate the corpus using two different role label sets: the PropBank role set and the VerbNet role set. VerbNet roles were generated using the SemLink mapping (Loper et al., 2007) , which provides a mapping between Prop-Bank and VerbNet role labels. In a small number of cases, no VerbNet role was available (e.g., because VerbNet did not contain the appropriate sense of the verb). In those cases, the PropBank role label was used instead.", |
|
"cite_spans": [ |
|
{ |
|
"start": 157, |
|
"end": 178, |
|
"text": "(Kipper et al., 2006)", |
|
"ref_id": "BIBREF6" |
|
}, |
|
{ |
|
"start": 364, |
|
"end": 384, |
|
"text": "(Loper et al., 2007)", |
|
"ref_id": "BIBREF8" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Semantic Role Labeling", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "We proposed two levels of participation in this task: i) Closed -the systems could use only the annotated data provided and nothing else. ii) Openwhere systems could use PropBank data from Sections 02-21, as well as any other resource for training their labelers.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Semantic Role Labeling", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "We selected 50 verbs from the 65 in the lexical sample task for the SRL task. The partitioning into train and test set was done in the same fashion as for the lexical sample task. Since PropBank does not tag any noun predicates, none of the 35 nouns from the lexical sample task were part of this data.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "For each system, we calculated the precision, recall, and F-measure for both role label sets. Scores were calculated using the srl-eval.pl script from the CoNLL-2005 scoring package (Carreras and M\u00e0rquez, 2005) . Only two teams chose to perform the SRL subtask. The performance of these two teams is shown in Table 6 : System performance on VerbNet roles.", |
|
"cite_spans": [ |
|
{ |
|
"start": 182, |
|
"end": 210, |
|
"text": "(Carreras and M\u00e0rquez, 2005)", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 309, |
|
"end": 316, |
|
"text": "Table 6", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "Given that only two systems participated in the task, it is difficult to form any strong conclusions. It should be noted that since there was no additional VerbNet role data to be used by the Open system, the performance of that on PropBank arguments as well as VerbNet roles is exactly identical. It can be seen that there is almost no difference between the performance of the Open and Closed systems for tagging PropBank arguments. The reason for this is the fact that all the instances of the lemma under consideration was selected from the Propbank corpus, and probably the number of training instances for each lemma as well as the fact that the predicate is such an important feature combine to make the difference negligible. We also realized that more than half of the test instances were contributed by the predicate \"say\" -the performance over whose arguments is in the high 90s. To remove the effect of \"say\" we also computed the performances after excluding examples of \"say\" from the test set. These numbers are shown in the bottom half of the two tables. These results are not directly comparable to the CoNLL-2005 shared task since: i) this test set comprises Sections 01, 22, 23 and 24 as opposed to just Section 23, and ii) this test set comprises data for only 50 predicates as opposed to all the verb predicates in the CoNLL-2005 shared task.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Discussion", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "The results in the previous discussion seem to confirm the hypothesis that there is a predictable correlation between human annotator agreement and system performance. Given high enough ITA rates we can can hope to build sense disambiguation systems that perform at a level that might be of use to a consuming natural language processing application. It is also encouraging that the more informative Verb-Net roles which have better/direct applicability in downstream systems, can also be predicted with almost the same degree of accuracy as the PropBank arguments from which they are mapped.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusions", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "http://www.cse.unt.edu/\u02dcrada/senseval/senseval3/scoring/ 2 scorer2 reports Precision and Recall scores for each system. For a sys-", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "We gratefully acknowledge the support of the Defense Advanced Research Projects Agency (DARPA/IPTO) under the GALE program, DARPA/CMO Contract No. HR0011-06-C-0022; National Science Foundation Grant NSF-0415923, Word Sense Disambiguation; the DTO-AQUAINT NBCHC040036 grant under the University of Illinois subcontract to University of Pennsylvania 2003-07911-01; and NSF-ITR-0325646: Domain-Independent Semantic Interpretation.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Acknowledgments", |
|
"sec_num": "5" |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "Introduction to the CoNLL-2004 shared task: Semantic role labeling", |
|
"authors": [ |
|
{ |
|
"first": "Xavier", |
|
"middle": [], |
|
"last": "Carreras", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Llu\u00eds", |
|
"middle": [], |
|
"last": "M\u00e0rquez", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of CoNLL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Xavier Carreras and Llu\u00eds M\u00e0rquez. 2004. Introduction to the CoNLL-2004 shared task: Semantic role labeling. In Proceedings of CoNLL-2004.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Introduction to the CoNLL-2005 Shared Task: Semantic Role Labeling", |
|
"authors": [ |
|
{ |
|
"first": "Xavier", |
|
"middle": [], |
|
"last": "Carreras", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Llu\u00eds", |
|
"middle": [], |
|
"last": "M\u00e0rquez", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Proceedings of CoNLL-2005", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Xavier Carreras and Llu\u00eds M\u00e0rquez. 2005. Introduction to the CoNLL-2005 Shared Task: Semantic Role Labeling. In Proceedings of CoNLL-2005.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "An empirical study of the behavior of active learning for word sense disambiguation", |
|
"authors": [ |
|
{ |
|
"first": "Jinying", |
|
"middle": [], |
|
"last": "Chen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Andrew", |
|
"middle": [], |
|
"last": "Schein", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Lyle", |
|
"middle": [], |
|
"last": "Ungar", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Martha", |
|
"middle": [], |
|
"last": "Palmer", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Proceedings of HLT/NAACL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jinying Chen, Andrew Schein, Lyle Ungar, and Martha Palmer. 2006. An empirical study of the behavior of active learning for word sense disambiguation. In Proceedings of HLT/NAACL.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Building a sense tagged corpus with open mind word expert", |
|
"authors": [ |
|
{ |
|
"first": "Timothy", |
|
"middle": [], |
|
"last": "Chklovski", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Rada", |
|
"middle": [], |
|
"last": "Mihalcea", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2002, |
|
"venue": "Proceedings of ACL-02 Workshop on WSD", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Timothy Chklovski and Rada Mihalcea. 2002. Building a sense tagged corpus with open mind word expert. In Proceedings of ACL-02 Workshop on WSD.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "GAMBL, genetic algorithm optimization of memory-based wsd", |
|
"authors": [ |
|
{ |
|
"first": "Bart", |
|
"middle": [], |
|
"last": "Decadt", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "V\u00e9ronique", |
|
"middle": [], |
|
"last": "Hoste", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Walter", |
|
"middle": [], |
|
"last": "Daelemans", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Antal", |
|
"middle": [], |
|
"last": "Van Den", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Bosch", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Bart Decadt, V\u00e9ronique Hoste, Walter Daelemans, and Antal Van den Bosch. 2004. GAMBL, genetic algorithm optimization of memory-based wsd. In Senseval-3.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Ontonotes: The 90% solution", |
|
"authors": [ |
|
{ |
|
"first": "Eduard", |
|
"middle": [], |
|
"last": "Hovy", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mitchell", |
|
"middle": [], |
|
"last": "Marcus", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Martha", |
|
"middle": [], |
|
"last": "Palmer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Lance", |
|
"middle": [], |
|
"last": "Ramshaw", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ralph", |
|
"middle": [], |
|
"last": "Weischedel", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Proceedings of HLT/NAACL", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Eduard Hovy, Mitchell Marcus, Martha Palmer, Lance Ramshaw, and Ralph Weischedel. 2006. Ontonotes: The 90% solution. In Proceedings of HLT/NAACL, June.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "Extending VerbNet with novel verb classes", |
|
"authors": [ |
|
{ |
|
"first": "Karin", |
|
"middle": [], |
|
"last": "Kipper", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Anna", |
|
"middle": [], |
|
"last": "Korhonen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Neville", |
|
"middle": [], |
|
"last": "Ryant", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Martha", |
|
"middle": [], |
|
"last": "Palmer", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Karin Kipper, Anna Korhonen, Neville Ryant, and Martha Palmer. 2006. Extending VerbNet with novel verb classes. In LREC-06.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "Senseval-3 task: Automatic labeling of semantic roles", |
|
"authors": [ |
|
{ |
|
"first": "Ken", |
|
"middle": [], |
|
"last": "Litkowski", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of Senseval-3", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Ken Litkowski. 2004. Senseval-3 task: Automatic labeling of semantic roles. In Proceedings of Senseval-3.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "Combining lexical resources: Mapping between propbank and verbnet", |
|
"authors": [ |
|
{ |
|
"first": "Edward", |
|
"middle": [], |
|
"last": "Loper", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yi", |
|
"middle": [], |
|
"last": "Szu Ting", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Martha", |
|
"middle": [], |
|
"last": "Palmer", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2007, |
|
"venue": "Proceedings of the IWCS-7", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Edward Loper, Szu ting Yi, and Martha Palmer. 2007. Combining lexical resources: Mapping between propbank and verbnet. In Proceedings of the IWCS-7.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "The proposition bank: A corpus annotated with semantic roles", |
|
"authors": [ |
|
{ |
|
"first": "Martha", |
|
"middle": [], |
|
"last": "Palmer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Daniel", |
|
"middle": [], |
|
"last": "Gildea", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Paul", |
|
"middle": [], |
|
"last": "Kingsbury", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2005, |
|
"venue": "Computational Linguistics", |
|
"volume": "31", |
|
"issue": "1", |
|
"pages": "71--106", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Martha Palmer, Daniel Gildea, and Paul Kingsbury. 2005. The proposition bank: A corpus annotated with semantic roles. Computational Linguistics, 31(1):71-106.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "Consistent criteria for sense distinctions", |
|
"authors": [ |
|
{ |
|
"first": "Martha", |
|
"middle": [], |
|
"last": "Palmer", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2000, |
|
"venue": "Computers and the Humanities", |
|
"volume": "34", |
|
"issue": "1-1", |
|
"pages": "217--222", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Martha Palmer. 2000. Consistent criteria for sense distinctions. Computers and the Humanities, 34(1-1):217-222.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"TABREF1": { |
|
"type_str": "table", |
|
"num": null, |
|
"text": "The number of instances for Verbs and Nouns in the Train and Test sets for the Lexical Sample WSD task.", |
|
"html": null, |
|
"content": "<table><tr><td>01, 22, 23 and 24. Fortunately, the distribution of</td></tr><tr><td>words was amenable to an acceptable number of in-</td></tr><tr><td>stances for each lemma in the test set. We selected</td></tr><tr><td>a total of 100 lemmas (65 verbs and 35 nouns) con-</td></tr><tr><td>sidering the degree of polysemy and total instances</td></tr><tr><td>that were annotated. The average ITA for these is</td></tr><tr><td>over 90%.</td></tr><tr><td>The training and test set composition is described</td></tr><tr><td>in Table 2. The distribution across all the verbs and</td></tr><tr><td>nouns is displayed in Table 4</td></tr></table>" |
|
}, |
|
"TABREF2": { |
|
"type_str": "table", |
|
"num": null, |
|
"text": "Stephen Tratz <[email protected]> PNNL MaxEnt 59.1\u00b14.5 2 Hwee Tou Ng <[email protected]> NUS-PT SVM 58.7\u00b14.5 3 Rada Mihalcea <[email protected]> UNT-Yahoo Memory-based 58.3\u00b14.5 4 Cai Junfu <[email protected]> NUS-ML naive Bayes 57.6\u00b14.5 5 Oier Lopez de Lacalle <[email protected]> UBC-ALM kNN 54.4\u00b14.5 6 David Martinez <[email protected]>", |
|
"html": null, |
|
"content": "<table><tr><td>Rank Participant</td><td>System ID</td><td>Classifier</td><td>F</td></tr><tr><td colspan=\"2\">1 UBC-UMB-2</td><td>kNN</td><td>54.0\u00b14.5</td></tr><tr><td>7 Jonathan Chang <[email protected]></td><td>PU-BCD</td><td colspan=\"2\">Exponential Model 53.9\u00b14.5</td></tr><tr><td>8 Radu ION <[email protected]></td><td>RACAI</td><td>Unsupervised</td><td>52.7\u00b14.5</td></tr><tr><td>9 Most Frequent WordNet Sense</td><td>Baseline</td><td>N/A</td><td>51.4\u00b14.5</td></tr><tr><td>10 Davide Buscaldi <[email protected]></td><td>UPV-WSD</td><td>Unsupervised</td><td>46.9\u00b14.5</td></tr><tr><td colspan=\"2\">11 Sudip Kumar Naskar <[email protected]> JU-SKNSB</td><td>Unsupervised</td><td>40.2\u00b14.5</td></tr><tr><td>12 David Martinez <[email protected]></td><td>UBC-UMB-1</td><td>Unsupervised</td><td>39.9\u00b14.5</td></tr><tr><td>14 Rafael Berlanga <[email protected]></td><td>tkb-uo</td><td>Unsupervised</td><td>32.5\u00b14.5</td></tr><tr><td>15 Jordan Boyd-Graber <[email protected]></td><td>PUTOP</td><td>Unsupervised</td><td>13.2\u00b14.5</td></tr><tr><td colspan=\"3\">Table 1: System Performance for the All-Words task.</td><td/></tr><tr><td>Rank Participant</td><td>System</td><td>Classifier</td><td>F</td></tr><tr><td>1 Cai Junfu <[email protected]></td><td>NUS-ML</td><td>SVM</td><td>88.7\u00b11.2</td></tr><tr><td colspan=\"2\">2 Oier Lopez de Lacalle <[email protected]> UBC-ALM</td><td>SVD+kNN</td><td>86.9\u00b11.2</td></tr><tr><td>3 Zheng-Yu Niu <niu [email protected]></td><td>I2R</td><td>Supervised</td><td>86.4\u00b11.2</td></tr><tr><td>4 Lucia Specia <[email protected]></td><td>USP-IBM-2</td><td>SVM</td><td>85.7\u00b11.2</td></tr><tr><td>5 Lucia Specia <[email protected]></td><td>USP-IBM-1</td><td>ILP</td><td>85.1\u00b11.2</td></tr><tr><td>5 Deniz Yuret <[email protected]></td><td>KU</td><td>Semi-supervised</td><td>85.1\u00b11.2</td></tr><tr><td colspan=\"2\">6 Saarikoski <[email protected]> OE</td><td colspan=\"2\">naive Bayes, SVM 83.8\u00b11.2</td></tr><tr><td>7 University of Technology Brno</td><td>VUTBR</td><td>naive Bayes</td><td>80.3\u00b11.2</td></tr><tr><td>8 Ana Zelaia <[email protected]></td><td>UBC-ZAS</td><td>SVD+kNN</td><td>79.9\u00b11.2</td></tr><tr><td>9 Carlo Strapparava <[email protected]></td><td>ITC-irst</td><td>SVM</td><td>79.6\u00b11.2</td></tr><tr><td>10 Most frequent sense in training</td><td>Baseline</td><td>N/A</td><td>78.0\u00b11.2</td></tr><tr><td>11 Toby Hawker <[email protected]></td><td>USYD</td><td>SVM</td><td>74.3\u00b11.2</td></tr><tr><td>12 Siddharth Patwardhan <[email protected]></td><td>UMND1</td><td>Unsupervised</td><td>53.8\u00b11.2</td></tr><tr><td>13 Saif Mohammad <[email protected]></td><td>Tor</td><td>Unsupervised</td><td>52.1\u00b11.2</td></tr><tr><td>-Toby Hawker <[email protected]></td><td>USYD *</td><td>SVM</td><td>89.1\u00b11.2</td></tr><tr><td>-Carlo Strapparava <[email protected]></td><td>ITC *</td><td>SVM</td><td>89.1\u00b11.2</td></tr></table>" |
|
}, |
|
"TABREF4": { |
|
"type_str": "table", |
|
"num": null, |
|
"text": "", |
|
"html": null, |
|
"content": "<table/>" |
|
}, |
|
"TABREF6": { |
|
"type_str": "table", |
|
"num": null, |
|
"text": "Table 6.", |
|
"html": null, |
|
"content": "<table><tr><td>System</td><td>Type</td><td colspan=\"2\">Precision Recall</td><td>F</td></tr><tr><td>UBC-UPC</td><td>Open</td><td>85.31</td><td colspan=\"2\">82.08 83.66\u00b10.5</td></tr><tr><td>UBC-UPC</td><td>Closed</td><td>85.31</td><td colspan=\"2\">82.08 83.66\u00b10.5</td></tr><tr><td>RTV</td><td>Closed</td><td>81.58</td><td colspan=\"2\">70.16 75.44\u00b10.6</td></tr><tr><td>Without \"say\"</td><td/><td/><td/></tr><tr><td>UBC-UPC</td><td>Open</td><td>79.23</td><td colspan=\"2\">73.88 76.46\u00b10.8</td></tr><tr><td>UBC-UPC</td><td>Closed</td><td>79.23</td><td colspan=\"2\">73.88 76.46\u00b10.8</td></tr><tr><td>RTV</td><td>Closed</td><td>73.63</td><td colspan=\"2\">57.44 64.53\u00b10.9</td></tr></table>" |
|
} |
|
} |
|
} |
|
} |