|
{ |
|
"paper_id": "S18-1012", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T15:43:39.958621Z" |
|
}, |
|
"title": "Chrono at SemEval-2018 Task 6: A System for Normalizing Temporal Expressions", |
|
"authors": [ |
|
{ |
|
"first": "Amy", |
|
"middle": [ |
|
"L" |
|
], |
|
"last": "Olex", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Virginia Commonwealth University", |
|
"location": { |
|
"region": "Virginia", |
|
"country": "USA" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Luke", |
|
"middle": [ |
|
"G" |
|
], |
|
"last": "Maffey", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Virginia Commonwealth University", |
|
"location": { |
|
"region": "Virginia", |
|
"country": "USA" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Nicholas", |
|
"middle": [], |
|
"last": "Morton", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Virginia Commonwealth University", |
|
"location": { |
|
"region": "Virginia", |
|
"country": "USA" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Bridget", |
|
"middle": [ |
|
"T" |
|
], |
|
"last": "Mcinnes", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "", |
|
"institution": "Virginia Commonwealth University", |
|
"location": { |
|
"region": "Virginia", |
|
"country": "USA" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "Temporal information extraction is a challenging task. Here we describe Chrono, a hybrid rule-based and machine learning system that identifies temporal expressions in text and normalizes them into the SCATE schema. After minor parsing logic adjustments, Chrono has emerged as the top performing system for SemEval 2018 Task 6: Parsing Time Normalizations.", |
|
"pdf_parse": { |
|
"paper_id": "S18-1012", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "Temporal information extraction is a challenging task. Here we describe Chrono, a hybrid rule-based and machine learning system that identifies temporal expressions in text and normalizes them into the SCATE schema. After minor parsing logic adjustments, Chrono has emerged as the top performing system for SemEval 2018 Task 6: Parsing Time Normalizations.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Understanding and processing temporal information is vital for navigating life. The human mind processes subtle temporal expressions instantly and effortlessly; however, it is difficult for computers to do the same. Identifying, processing, and utilizing this information requires knowledge and understanding of syntax, semantics, and context to link temporal information to related events and order them on a time-line. SemEval 2018 Task 6 (Laparra et al., 2018) aims to normalize fine-grained temporal information and relationships into the Semantically Compositional Annotations for Temporal Expressions (SCATE) schema developed by (Bethard and Parker, 2016) . This scheme aims to improve upon the current TIMEX3/TimeML (Pustejovsky et al., 2003) standard by representing a wide variety of temporal expressions, allowing for events to act as anchors, and using mathematical operations over a timeline to define the semantics of each annotation. To address this challenge, we developed Chrono 1 , a hybrid rule-based and machine learning (ML) Python package that normalizes temporal expressions into the SCATE schema.", |
|
"cite_spans": [ |
|
{ |
|
"start": 441, |
|
"end": 463, |
|
"text": "(Laparra et al., 2018)", |
|
"ref_id": "BIBREF3" |
|
}, |
|
{ |
|
"start": 635, |
|
"end": 661, |
|
"text": "(Bethard and Parker, 2016)", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 723, |
|
"end": 749, |
|
"text": "(Pustejovsky et al., 2003)", |
|
"ref_id": "BIBREF5" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "1 https://github.com/AmyOlex/Chrono", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "Our approach to building this hybrid system includes four processing phases: 1) text preprocessing, 2) flagging numeric and temporal tokens, 3) temporal expression identification, and 4) SCATE normalization.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The Chrono System", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "1) Text Pre-processing: Python's Natural Language Toolkit (NLTK) WhitespaceTokenizer and part-of-speech (POS) tagger (Bird and Loper, 2004) process raw text files to identify individual tokens, token spans, and POS tags. Punctuation is not handled at this phase as it is important for identifying correct spans.", |
|
"cite_spans": [ |
|
{ |
|
"start": 117, |
|
"end": 139, |
|
"text": "(Bird and Loper, 2004)", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The Chrono System", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "2) Flagging Numeric and Temporal Tokens: All numeric tokens are flagged regardless of context. Subsequent phases utilize contextual information to determine if a numeric token is part of a temporal expression. Depending on the task, a rule may remove all or some punctuation, and/or convert tokens to lowercase prior to parsing. In the following, RP and LC denote Removing all Punctuation and converting to LowerCase, respectively.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The Chrono System", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Numeric Flagging: Tokens are flagged as numeric if either 1) the token has a POS tag of \"CD\" (Cardinal Number), or 2) the text can be converted to a numeric expression. Textual representations of numeric expressions are converted to numerics with the Word2Number 2 Python module. A custom method recognizes ordinals from \"first\" to \"thirty-first\" and converts them into the associated numerics 1 to 31, respectively. LC normalization is done prior to parsing textual numerics.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The Chrono System", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Temporal Flagging: Temporal tokens are flagged through rule-based parsing using lists of key words and regular expressions. This phase is more liberal in its identification of a temporal token than the SCATE normalization phase, so it identifies a broader range of potential temporal tokens that are refined in future steps. Tokens may be numeric and temporal simultaneously. Numeric tokens with the characters '$', '#', or '%' are NOT marked as temporal. The following types of tokens are flagged as temporal:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The Chrono System", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "\u2022 Formatted date patterns using '/' or '-': mm/dd/yyyy, mm/dd/yy, yyyy/mm/dd, or yy/mm/dd \u2022 Formatted time patterns matching hh:mm:ss \u2022 Sequence of 4 to 8 consecutive digits matching range criteria for 24-hour times or for a year, month, and/or day (e.g. 1998 or 08241998). \u2022 Spelled out month or abbreviation, e.g. \"Mar.\" or \"March\", are flagged after RP except periods as they are required to retrieve correct spans. \u2022 Days of the week, e.g. \"Sat.\" or \"Saturday\", are parsed similar to months. \u2022 Other temporal words, e.g. \"this\", \"now\", \"nearly\", and others, are flagged after RP and LC.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The Chrono System", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "3) Temporal Expression Identification: A temporal expression is represented by a temporal phrase, which we define as two or more consecutive temporal/numeric tokens on the same line, or an isolated temporal token, with some exceptions. If a numeric token contains a '$', '#', or '%', or the text 'million', 'billion', or 'trillion' it is not included in a temporal phrase as these generally refer to non-temporal values. Additionally, isolated numeric tokens are not considered a temporal phrase. 4) SCATE Normalization: Chrono parses each temporal phrase into zero or more SCATE entities, links sub-intervals, and disambiguates the SCATE entities \"Period\" and \"Calendar-Interval\" via a machine learning module. Chrono imple-ments 32 types of entities with 5 parent types that have been described by (Bethard and Parker, 2016) . Parsing strategies differ depending on the composition of a temporal phrase being parsed. Each temporal phrase is interrogated by all of the following parsing strategies.", |
|
"cite_spans": [ |
|
{ |
|
"start": 800, |
|
"end": 826, |
|
"text": "(Bethard and Parker, 2016)", |
|
"ref_id": "BIBREF0" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "The Chrono System", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Formatted dates/times are parsed using regular expressions. To identify which format the date/time is in, Chrono looks for a 2-digit or 4-digit year first, then uses that position for orientation to identify the remaining elements. If a formatted date/time is identified, then the appropriate sub-intervals are linked during element parsing. 4-digit years take precedence over 2-digit years.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Formatted Dates and Times:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Numeric Dates and Times: Header and metadata for Newswire articles frequently have numeric dates listed with no punctuation (e.g. \"19980218\" codes for \"Feb, 18 1998\"), and isolated 4-digit year mentions are frequent. After formatted dates and times are parsed, any phrase containing a numeric token is interrogated for a potential date or year mention. If a numeric token is 4-digits it is tested for a year between 1500 and 2050, 6-digit tokens are parsed for 2-digit year/month/day, and 8-digit strings are parsed for a 4-digit year and 2-digit month/day. All elements must be in the proper range, otherwise the token is skipped. Appropriate sub-intervals are linked during element parsing.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Formatted Dates and Times:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "24-hour Time: 24-hour times are identified by either the format hhmmzzz, where zzz is the time zone, or a 4-digit number that has not been classified as a year. Hour digits must be less than 24 and minutes less than 60. Sub-intervals are linked at this time if existing. Time zones are handled separately and are linked back to the hour entity during the final sub-interval linking step.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Formatted Dates and Times:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Temporal Token Search: The majority of textual temporal entities are identified by looking for specific tokens. Token categories include days of the week, months, parts of a day/week, time zones, and other temporal operators such as \"early\", \"this\", \"before\", etc. Prior to looking for these tokens, text is normalized by RP and LC. Exceptions to RP include searching for day/month abbreviations, such as \"Sat.\" or \"Aug.\". In these cases periods are not removed because they are part of the SCATE span. Another exception to RP and LC is identifying mentions of AM or PM where periods are kept and text is not converted to lowercase in order to capture variations like \"PM\" or \"p.m.\". Non-temporal mentions of the months or seasons of the year \"may\", \"march\", \"spring\", and \"fall\" are disambiguated using POS tags, where tokens that refer to a temporal entity generally have a POS tag of \"NN\" or \"NP\". Sub-intervals are not linked during token searches.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Formatted Dates and Times:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Text Year: Another special case of parsing temporal tokens are textual representations of years such as \"nineteen ninety-seven\". The Word2Number Python module was modified to recognize these phrases. Previously, it would add 19 and 97 together instead of returning 1997.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Formatted Dates and Times:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Periods and Calendar-Intervals: The same temporal token can refer to either a SCATE \"Period\" or \"Calendar-Interval\". For example, in the phrases \"in a week\" vs \"next week\" the token \"week\" is classified differently. Due to language intricacies it is difficult to define a rule-base system to disambiguate these entities as the classification is contingent on the topic being discussed where phrasing around the entity can be different for each instance. Thus, Period/Calendar-Interval tokens are initially identified by a token search using a defined list of terms, then the identified term and its span are passed to a ML algorithm for classification.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Formatted Dates and Times:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Machine Learning Classification: Four ML algorithms are available in Chrono to differentiate between \"Period\" and \"Calendar-Interval\" entities using contextual information. Chrono implements Naive Bayes (NB), Neural Network (NN), Decision Tree (DT), and Support Vector Machine (SVM). Binary feature vectors for all implementations have the following features:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Formatted Dates and Times:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "\u2022 temporal self: If the target is flagged as temporal, this feature is set to \"1\". \u2022 temporal context: If there is at least one temporal word within a 5-word window up-or down-stream of the target this feature is set to \"1\". \u2022 numeric: If there is a numeric expression either directly before or after (a 1-word window) the target, this feature is set to \"1\". \u2022 context: All words within a 5-word window are identified as features and set to \"1\" if that word is present. Prior to identifying these features all words are normalized with RP and LC. The 5-word window includes crossing sentence boundaries before and after the target word.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Formatted Dates and Times:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "We use NLTK with default parameters to implement NB and DT, NN is a simple feed-forward network with three hidden layers implemented using Python's Keras package 3 with epochs set to 5 and batch set to 10, and SVM is implemented using SciKitLearn (Pedregosa et al., 2011) with C set to 0.05 and max iterations set to 3. Ordinals: Ordinals such as \"first\" or \"3rd\" are classified as an \"NthFromStart\" entity in the SCATE schema. These mentions are identified by normalizing with RP and LC before searching for the ordinal tokens representing the numbers 1-31.", |
|
"cite_spans": [ |
|
{ |
|
"start": 247, |
|
"end": 271, |
|
"text": "(Pedregosa et al., 2011)", |
|
"ref_id": "BIBREF4" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Formatted Dates and Times:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Sub-Interval Linking: After all SCATE entities are identified, all temporal phrases are re-parsed to identify sub-intervals within each phrase. For example, entities in the phrase \"August 1998\" are parsed by two different methods leaving the subinterval link vacant. During sub-interval linking, the year \"1998\" has the \"August\" entity added as a sub-interval. Sub-interval linking reviews entities from the smallest to the largest, adding missing sub-intervals as needed. This method assumes each temporal phrase contains zero or one of each type of SCATE entity.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Formatted Dates and Times:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Next/Last Parsing: Determining whether an entity is referring to a date in the future, \"Next\", or past, \"Last\", depends on context and the document time (doc-time). Next/Last parsing is done after all other parsing, and checks two cases: 1) if a temporal phrase contains a year, no additional annotation is made, and 2) if specific modifier words are present (e.g. \"next\" or \"last\") immediately preceding a temporal expression, the modifier is annotated with a sub-interval referencing the following temporal entity. If neither of these cases hold, the year is set as the doc-time year, and the month and day are compared to the full doc-time to determine if it occurs before or after. Note the year assumption is not always valid and more complex, content-based parsing may be required to achieve higher precision. Finally, if a day of the week (e.g. \"Saturday\") is mentioned, Chrono finds the first preceding verb in the sentence, and if it is past tense the temporal entity is annotated as \"Last\", otherwise it is annotated as \"Next\".", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Formatted Dates and Times:", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Training and evaluation of Chrono utilizes the Newswire corpus, consisting of 81 documents, provided by the task organizers. Average preci-sion, recall, and F1-measure of 5-fold cross validation for Track 1 (parsing) are reported in Table 1 (annotations for \"Event\" and \"Modifier\" are ignored). Scores for \"100% Correct Entity\" consider the entity location and all properties (like sub-intervals), and scores for \"Correct Span\" only consider the entity location.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "On average, all ML algorithms perform similarly for the \"100% Correct Entity\". All versions also obtain a higher F1 score when only considering correct spans versus getting all entity properties correct. This indicates that Chrono correctly identifies the majority of temporal entities, but has trouble parsing some of the properties.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "ChronoNN processed the final evaluation dataset, which consisted of 20 previously unseen Newswire articles, and received a F1 of .44. The evaluation dataset contained five articles from BBC that were not represented in the training dataset. Chrono's low performance indicates that it may be over-fit to the the training dataset. This is one downfall of rule-based systems, where new rules need to be developed for each new type of temporal representation. Upon further review we found the submitted version of Chrono had three minor parsing flaws that resulted in unintentional false positives.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "1) Formatted dates such as \"2013-02-22\" were being parsed twice. The first parse specifically looked for a 4-digit year and identified all correct entities, then the second parse looked for a formatted date with a 2-digit year, but didn't check to see if a year had already been found, so returned a 2-digit year with the value \"22\". This was easily fixed by having the 2-digit year parser check for a 4-digit year flag before proceeding (month and day flags were already implemented).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "2) 24-hour time priority was incorrectly placed above 4-digit year. This resulted in any isolated 4digit year being parsed as a 24-hour time expression rather than a year as originally intended. A simple flip of parsing order resolved this issue.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "3) Numeric temporal expressions, such as an isolated 4-digit year, were being parsed as a whole phrase rather than breaking out each token within the phrase. For example, the year in the phrase \"Last 1953\" was not being identified because it was not in a phrase all by itself. To fix this the parsing function was edited to loop through each token in a phrase (a method that was already implemented in most other parsers and was just over- looked here).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "ChronoNN received a Post-Evaluation F1 of .55 for Track 1 after implementing these fixes, which sets ChronoNN as the top performing system for SemEval 2018 Task 6, Track 1.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Results", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Chrono is currently the top performing system for Track 1 of Task 6, but there are still many areas that can be improved. Notably, we plan to implement \"Event\" and \"Between\" parsing, as well as refine current strategies as new temporal expressions are identified. Utilizing sentence tokenization instead of relying on new lines could improve phrase identification; however, this did not appear to be a major source of error in parsing the Newswire dataset. Additionally, usability can be improved by moving all parsing rules to separate, customizable files. We also plan to expand ML use to additional disambiguation tasks, and implement an ensemble system utilizing all four ML methods. We aim to extract the temporal phrase parser into a stand-alone system and compare it's performance directly to existing programs like SUTime (Chang and Manning, 2012) and HeidelTime (Strtgen and Gertz, 2010) as it has done a decent job of identifying temporal entities in this challenge. Finally, we will evaluate Chrono's performance on the THYME dataset (Styler IV et al., 2014) using the post-evaluation submission system.", |
|
"cite_spans": [ |
|
{ |
|
"start": 871, |
|
"end": 896, |
|
"text": "(Strtgen and Gertz, 2010)", |
|
"ref_id": "BIBREF6" |
|
}, |
|
{ |
|
"start": 1045, |
|
"end": 1069, |
|
"text": "(Styler IV et al., 2014)", |
|
"ref_id": "BIBREF7" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusions and Future Work", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "https://github.com/akshaynagpal/w2n", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "https://github.com/keras-team/keras", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "A semantically compositional annotation scheme for time normalization", |
|
"authors": [ |
|
{ |
|
"first": "Steven", |
|
"middle": [], |
|
"last": "Bethard", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jonathan", |
|
"middle": [], |
|
"last": "Parker", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Lrec", |
|
"volume": "2016", |
|
"issue": "", |
|
"pages": "3779--3786", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Steven Bethard and Jonathan Parker. 2016. A semanti- cally compositional annotation scheme for time nor- malization. In Lrec, volume 2016, pages 3779- 3786.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Nltk: the natural language toolkit", |
|
"authors": [ |
|
{ |
|
"first": "Steven", |
|
"middle": [], |
|
"last": "Bird", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Edward", |
|
"middle": [], |
|
"last": "Loper", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of the ACL 2004 on Interactive poster and demonstration sessions, page 31. Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Steven Bird and Edward Loper. 2004. Nltk: the nat- ural language toolkit. In Proceedings of the ACL 2004 on Interactive poster and demonstration ses- sions, page 31. Association for Computational Lin- guistics.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Sutime: A library for recognizing and normalizing time expressions", |
|
"authors": [ |
|
{ |
|
"first": "X", |
|
"middle": [], |
|
"last": "Angel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Chang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "D", |
|
"middle": [], |
|
"last": "Christopher", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Manning", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2012, |
|
"venue": "Lrec", |
|
"volume": "2012", |
|
"issue": "", |
|
"pages": "3735--3740", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Angel X. Chang and Christopher D. Manning. 2012. Sutime: A library for recognizing and normaliz- ing time expressions. In Lrec, volume 2012, pages 3735-3740.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Semeval 2018 task 6: Parsing time normalization", |
|
"authors": [ |
|
{ |
|
"first": "Egoitz", |
|
"middle": [], |
|
"last": "Laparra", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Dongfang", |
|
"middle": [], |
|
"last": "Xu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Steven", |
|
"middle": [], |
|
"last": "Bethard", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ahmed", |
|
"middle": [ |
|
"S" |
|
], |
|
"last": "Elsayed", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Martha", |
|
"middle": [], |
|
"last": "Palmer", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "Proceedings of the 12th International Workshop on Semantic Evaluation, SemEval '18", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Egoitz Laparra, Dongfang Xu, Steven Bethard, Ahmed S. Elsayed, and Martha Palmer. 2018. Sem- eval 2018 task 6: Parsing time normalization. In Proceedings of the 12th International Workshop on Semantic Evaluation, SemEval '18, New Orleans, LA, USA. Association for Computational Linguis- tics.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Scikit-learn: Machine learning in Python", |
|
"authors": [ |
|
{ |
|
"first": "F", |
|
"middle": [], |
|
"last": "Pedregosa", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "G", |
|
"middle": [], |
|
"last": "Varoquaux", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Gramfort", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "V", |
|
"middle": [], |
|
"last": "Michel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "B", |
|
"middle": [], |
|
"last": "Thirion", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "O", |
|
"middle": [], |
|
"last": "Grisel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Blondel", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "P", |
|
"middle": [], |
|
"last": "Prettenhofer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "Weiss", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "V", |
|
"middle": [], |
|
"last": "Dubourg", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Vanderplas", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Passos", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "D", |
|
"middle": [], |
|
"last": "Cournapeau", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Brucher", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Perrot", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "E", |
|
"middle": [], |
|
"last": "Duchesnay", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2011, |
|
"venue": "Journal of Machine Learning Research", |
|
"volume": "12", |
|
"issue": "", |
|
"pages": "2825--2830", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Pretten- hofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Pas- sos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. 2011. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825-2830.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Timeml: Robust specification of event and temporal expressions in text", |
|
"authors": [ |
|
{ |
|
"first": "James", |
|
"middle": [], |
|
"last": "Pustejovsky", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Jos\u00e9", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Robert", |
|
"middle": [], |
|
"last": "Castano", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Roser", |
|
"middle": [], |
|
"last": "Ingria", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Sauri", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Robert", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Andrea", |
|
"middle": [], |
|
"last": "Gaizauskas", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Graham", |
|
"middle": [], |
|
"last": "Setzer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Katz", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Dragomir R Radev", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2003, |
|
"venue": "New directions in question answering", |
|
"volume": "3", |
|
"issue": "", |
|
"pages": "28--34", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "James Pustejovsky, Jos\u00e9 M Castano, Robert Ingria, Roser Sauri, Robert J Gaizauskas, Andrea Set- zer, Graham Katz, and Dragomir R Radev. 2003. Timeml: Robust specification of event and tempo- ral expressions in text. New directions in question answering, 3:28-34.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "Heideltime: High quality rule-based extraction and normalization of temporal expressions", |
|
"authors": [ |
|
{ |
|
"first": "Jannik", |
|
"middle": [], |
|
"last": "Strtgen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Michael", |
|
"middle": [], |
|
"last": "Gertz", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2010, |
|
"venue": "Proceedings of the 5th International Workshop on Semantic Evaluation, SemEval '10", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "321--324", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jannik Strtgen and Michael Gertz. 2010. Heideltime: High quality rule-based extraction and normaliza- tion of temporal expressions. In Proceedings of the 5th International Workshop on Semantic Evalua- tion, SemEval '10, pages 321-324, Stroudsburg, PA, USA. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "Temporal annotation in the clinical domain", |
|
"authors": [ |
|
{ |
|
"first": "F", |
|
"middle": [], |
|
"last": "William", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "I", |
|
"middle": [ |
|
"V" |
|
], |
|
"last": "Styler", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Steven", |
|
"middle": [], |
|
"last": "Bethard", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sean", |
|
"middle": [], |
|
"last": "Finan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Martha", |
|
"middle": [], |
|
"last": "Palmer", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sameer", |
|
"middle": [], |
|
"last": "Pradhan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Piet", |
|
"middle": [ |
|
"C" |
|
], |
|
"last": "De Groen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Brad", |
|
"middle": [], |
|
"last": "Erickson", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Timothy", |
|
"middle": [], |
|
"last": "Miller", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Chen", |
|
"middle": [], |
|
"last": "Lin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Guergana", |
|
"middle": [], |
|
"last": "Savova", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Transactions of the Association for Computational Linguistics", |
|
"volume": "2", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "William F. Styler IV, Steven Bethard, Sean Finan, Martha Palmer, Sameer Pradhan, Piet C. de Groen, Brad Erickson, Timothy Miller, Chen Lin, and Guer- gana Savova. 2014. Temporal annotation in the clin- ical domain. Transactions of the Association for Computational Linguistics, 2:143.", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"text": "Temporal words indicating periods of time, e.g. \"yesterday\" or \"decade\", are flagged after RP and LC.\u2022 Mentions of AM and PM in any format are flagged after RP except periods. \u2022 The parts of a week, e.g. \"weekend\" and \"weekends\", are flagged after RP and LC. \u2022 Seasons of the year are flagged after RP and LC. \u2022 Various parts of a day, e.g. \"noon\" or \"morning\", are flagged after RP and LC. \u2022 Time zones are flagged after RP.", |
|
"uris": null, |
|
"num": null, |
|
"type_str": "figure" |
|
} |
|
} |
|
} |
|
} |