|
{ |
|
"paper_id": "S16-1047", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T15:26:40.197672Z" |
|
}, |
|
"title": "IHS-RD-Belarus at SemEval-2016 Task 5: Detecting Sentiment Polarity Using the Heatmap of Sentence", |
|
"authors": [ |
|
{ |
|
"first": "Maryna", |
|
"middle": [], |
|
"last": "Chernyshevich", |
|
"suffix": "", |
|
"affiliation": {}, |
|
"email": "[email protected]" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "This paper describes the system submitted by IHS-RD-Belarus team for the sentiment detection polarity subtask on Aspect Based Sentiment Analysis task at the SemEval 2016 workshop on semantic evaluation. We developed a system based on artificial neural network to detect the sentiment polarity of opinions. Evaluation on the test data set showed that our system achieved the F-score of 0.83 for restaurants domain (rank 4 out of 28 submissions) and F-score of 0.78 for laptops domain (rank 4 out of 21 submissions).", |
|
"pdf_parse": { |
|
"paper_id": "S16-1047", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "This paper describes the system submitted by IHS-RD-Belarus team for the sentiment detection polarity subtask on Aspect Based Sentiment Analysis task at the SemEval 2016 workshop on semantic evaluation. We developed a system based on artificial neural network to detect the sentiment polarity of opinions. Evaluation on the test data set showed that our system achieved the F-score of 0.83 for restaurants domain (rank 4 out of 28 submissions) and F-score of 0.78 for laptops domain (rank 4 out of 21 submissions).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "Social media texts found in user review services have a great data-mining potential, as they offer real-time data that can be useful to monitor public opinion on brands, products, events, etc.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "Most of recent approaches to sentiment analysis task are based on bag-of-words features, syntactic dependency features, out-of-domain and domainspecific sentiment lexicons, to train an supervised model that predicts the polarity of each given term or aspect category. This approach is very popular but it relies on heavy pre-processing of the data which involves careful choosing of the right features, empirical thresholds and intuitive analysis of the training set (Brun et al., 2014 , Saias, 2015 .", |
|
"cite_spans": [ |
|
{ |
|
"start": 467, |
|
"end": 485, |
|
"text": "(Brun et al., 2014", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 486, |
|
"end": 499, |
|
"text": ", Saias, 2015", |
|
"ref_id": "BIBREF3" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "In this paper, we present an approach to the opinion polarity detection task based on an artificial neural network and sentiment orientation score of words.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "The SemEval-2016 shared task on Aspect based Sentiment Analysis focuses on identifying the Opinion target expressions (OTE), the Aspect categories and the sentiment expressed towards each OTE or Aspect category.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Task description", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The main focus of this paper is polarity subtasks, such as OTE polarity in restaurant subject domain and Aspect category polarity in laptop subject domain.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Task description", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "In the OTE polarity subtask, the input consists of a review sentence and the set of terms or aspect categories. The expected output is a polarity label (positive, negative or neutral) for each of the associated terms or aspect categories.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Task description", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "For example, the system should determine the polarity of fajitas and salads in the following sentence: I hated their fajitas, but their salads were great.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Task description", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "As for the Aspect category polarity, the task is more complicated. In the following sentence, the system has to determine the polarity of display quality (DISPLAY#QUALITY) and display usability (DISPLAY#USABILITY): The display has a great resolution but has difficulty always seeing the small print.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Task description", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The task organizers provided a dataset of customer reviews with manually annotated opinion targets: 2500 sentences for laptop domain and 2000 sentences for restaurant domain.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Task description", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "Evaluation was to be carried out according to Precision, Recall and F1 metrics.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Task description", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The central idea behind our system is the visualization of sentiment orientation in a word sequence using heatmap, that highlights regions with higher or lower \"temperature\". The \"temperature\" of a word is its sentiment polarity, positive or negative, and the intensity is calculated based on sentiment orientation score of word.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "System description", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "Sentiment orientation score indicates the strength of association of words (w) with positive (pr) and negative (nr) reviews. Following Turney and Littman 2003, we calculated SO as the difference using Pointwise Mutual Information (PMI) measures:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Sentiment orientation lexicon", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "( ) = ( , ) \u2212 ( , )", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Sentiment orientation lexicon", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "SO score is positive when the word or phrase tends to occur mostly in positive reviews and negative when the word occurs more often in negative reviews. The magnitude indicates the degree of association.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Sentiment orientation lexicon", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "We followed the lexicon-generation approach proposed by Kiritchenko et al. (Kiritchenko et al., 2014) and when generating the sentiment lexicons we distinguished the terms appearing in negated and affirmative contexts. Sentiment scores were then calculated separately for these two types of contexts.", |
|
"cite_spans": [ |
|
{ |
|
"start": 56, |
|
"end": 101, |
|
"text": "Kiritchenko et al. (Kiritchenko et al., 2014)", |
|
"ref_id": "BIBREF6" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Sentiment orientation lexicon", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "We created uni-, bi-and tri-gram lexicons based on Yelp restaurant reviews and Amazon reviews.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Sentiment orientation lexicon", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "The final SO score of n-gram in a sentence can be affected by some neighbour terms, such as valence shifters and intensifiers (Kennedy and Inkpen, 2006) . We created short wordlists and tuned final SO score of n-gram if words from these lists were found in term context.", |
|
"cite_spans": [ |
|
{ |
|
"start": 126, |
|
"end": 152, |
|
"text": "(Kennedy and Inkpen, 2006)", |
|
"ref_id": "BIBREF2" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Sentiment orientation score", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "Valence shifters are terms that can change the semantic orientation of another term, for example, combining positively valenced words with a nega-tion flips the positive orientation to a negative orientation. The most important shifters are negations such as not, never, none, nobody, etc. (Polanyi and Zaenen, 2004) .", |
|
"cite_spans": [ |
|
{ |
|
"start": 290, |
|
"end": 316, |
|
"text": "(Polanyi and Zaenen, 2004)", |
|
"ref_id": "BIBREF4" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Valence shifters", |
|
"sec_num": "3.2.1" |
|
}, |
|
{ |
|
"text": "As it was mentioned above we included negations, such as not and never, as n-gram postfix into lexicons. If the n-gram in negative context is not found in lexicon or its raw frequency in review corpus is less than 5, the final SO score is calculated based on calculation rules showed in table 1.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Valence shifters", |
|
"sec_num": "3.2.1" |
|
}, |
|
{ |
|
"text": "These negation rules are designed in order to improve the sentiment text analysis and are based on simple assumption: the negation flips positive the valence of a word to a negative with the same strength, but the negation doesn't flip the valence of a negative word, it rather reduces its strength.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Valence shifters", |
|
"sec_num": "3.2.1" |
|
}, |
|
{ |
|
"text": "SO=pos_SO * -1 SO(love) = +2.6 SO(not love) = -2.6 negative SO SO=neg_SO * 0.5 SO(hate) = -2.2 SO(not hate) = -1.1 Table 1 : SO calculation rules in negative context.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 115, |
|
"end": 122, |
|
"text": "Table 1", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "SO of word calculation rule Example positive SO", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "Intensifiers are terms that change the degree of the expressed sentiment. For example, in the sentence \"The waterfly cases are very good.\", the term very good is more positive together than just good alone. So to calculate the final SO score we multiply sentiment score of n-gram by 3.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Intensifiers", |
|
"sec_num": "3.2.2" |
|
}, |
|
{ |
|
"text": "On another side, in the sentence \"The waterfly cases are barely good.\", the term barely makes this statement less positive. We created 4 lists of intensifiers of different intensity of affecting the final SO score of n-gram.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Intensifiers", |
|
"sec_num": "3.2.2" |
|
}, |
|
{ |
|
"text": "A sentence in reviews can express not only a real user experience, but also an unreal opinion, for example a wish, \"I should have bought something better.\" or \"The laptop may be better\". We collected all surface patterns which can express unreality or conditionality. The SO of all positive ngrams in unreal context is flipped to negative with the same strength. The SO of negative n-grams is not changed. ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Unreality and conditionality", |
|
"sec_num": "3.2.3" |
|
}, |
|
{ |
|
"text": "A fully connected multilayer neural network with back-propagation is applied. The network, illustrated in Fig. 1 , contains 3 layers:", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 106, |
|
"end": 112, |
|
"text": "Fig. 1", |
|
"ref_id": "FIGREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Neural network architecture", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "\uf0b7 input with 81 nodes -one for each feature presenting \"temperature\" range; \uf0b7 hidden layer with 80 nodes -each node in the input layer is connected to each node of the hidden layer; \uf0b7 output layer with 3 nodes -one for each of 3 classes. As the activation function the sigmoid function is used. We apply dropout to our hidden layers, as described in Srivastava et al. (Srivastava et al., 2014) , to prevent network overfitting.", |
|
"cite_spans": [ |
|
{ |
|
"start": 350, |
|
"end": 393, |
|
"text": "Srivastava et al. (Srivastava et al., 2014)", |
|
"ref_id": "BIBREF6" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Neural network architecture", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "The network architecture is implemented on Keras 1 , which is an effective deep learning framework implementation in python.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Neural network architecture", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "At the first step, we detect the context of opinion target expressions or category. If the sentence has only one opinion target, the term context will include all words from the beginning to the end of the sentence. If the sentence has many targets, the context will include all words surrounding the term enclosed between two separators. As a separator we consider all punctuation marks, the next opinion target, the end and the beginning of the sentence. In the laptops subtask we had to detect polarity of the category that is not bounded to any word in the sentence. That is why we considered the whole sentence as context for every aspect category.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Neural network architecture", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "For training, each term context was represented as a vector of 81 features that resembles the scale of \"temperature\" from -40 (very negative) to 40 (very positive). Each word according to its SO score was placed on this scale. The value of a feature means the number of words within this range of sentiment orientation. Table 2 illustrates a set of n-grams with their final SO scores and \"temperature\" extracted from the sentence \"Now the speed is disappointing\". The number of training epochs is set to 2. More epochs would lead to overfitting because the training set that is relatively small.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 320, |
|
"end": 327, |
|
"text": "Table 2", |
|
"ref_id": "TABREF1" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Neural network architecture", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "The 3-fold cross-validation procedure was used to test the system performance. It uses 66% of the data to train and the remaining 33% to assess the accuracy. This is repeated 3 times with a different 33% left. On both sets our system did not recognize neutral (mildly positive or mildly negative sentiment) sentiments, although we used 3 classes classification. The reason is the low number of neutral sentiments in the annotated corpus.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Results and further work", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "As a future work, we intend to develop more complex artificial neural network architecture and try to use review-long memory about the polarity of previous targets. Also we are going to investi-gate the influence of words with extremely high or low sentiment orientation score on the SO scores of neighbouring words.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Results and further work", |
|
"sec_num": "4" |
|
}, |
|
{ |
|
"text": "We have presented a simple neural network architecture that predicts polarity of word sequence based the heatmap of sentence that highlights regions with higher or lower sentiment intensity. We submitted runs corresponding to the slot3 subtasks, obtaining competitive results. Our submission was ranked 4 th out of 28 submissions for restaurants domain and 4 th out of 21 submissions for laptops domain.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "http://keras.io/", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "The Theory of Parsing, Translation and Compiling", |
|
"authors": [ |
|
{ |
|
"first": "Alfred", |
|
"middle": [ |
|
"V" |
|
], |
|
"last": "Aho", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jeffrey", |
|
"middle": [ |
|
"D" |
|
], |
|
"last": "Ullman", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 1972, |
|
"venue": "", |
|
"volume": "1", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Alfred. V. Aho and Jeffrey D. Ullman. 1972. The Theory of Parsing, Translation and Compiling, volume 1. Prentice-Hall, Englewood Cliffs, NJ.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "XRCE: Hybrid Classification for Aspect-based Sentiment Analysis", |
|
"authors": [ |
|
{ |
|
"first": "Caroline", |
|
"middle": [], |
|
"last": "Brun", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Diana", |
|
"middle": [ |
|
"Nicoleta" |
|
], |
|
"last": "Popa", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Claude", |
|
"middle": [], |
|
"last": "Roux", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the eight international workshop on Semantic Evaluation Exercises (SemEval-2014)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Caroline Brun, Diana Nicoleta Popa, Claude Roux. 2014. XRCE: Hybrid Classification for Aspect-based Sentiment Analysis. in Proceedings of the eight inter- national workshop on Semantic Evaluation Exercises (SemEval-2014), August 2014, Dublin, Ireland.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Sentiment Classification of Movie Reviews using Contextual Valence Shifters", |
|
"authors": [ |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Kennedy", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "D", |
|
"middle": [], |
|
"last": "Inkpen", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2006, |
|
"venue": "Computational Intelligence", |
|
"volume": "22", |
|
"issue": "2", |
|
"pages": "110--125", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kennedy, A. and D. Inkpen. 2006. Sentiment Classifica- tion of Movie Reviews using Contextual Valence Shifters. Computational Intelligence, vol.22(2), pp.110-125, 2006.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "Sentiue: Target and Aspect based Sentiment Analysis in SemEval-2015 Task 12", |
|
"authors": [ |
|
{ |
|
"first": "Jos\u00e9", |
|
"middle": [], |
|
"last": "Saias", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2015, |
|
"venue": "Proceedings of the 9th International Workshop on Semantic Evaluation", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jos\u00e9 Saias. 2015. Sentiue: Target and Aspect based Sen- timent Analysis in SemEval-2015 Task 12. In Pro- ceedings of the 9th International Workshop on Se- mantic Evaluation (SemEval 2015), Denver, Colora- do.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Contextual valence shifters", |
|
"authors": [ |
|
{ |
|
"first": "Livia", |
|
"middle": [], |
|
"last": "Polanyi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Annie", |
|
"middle": [], |
|
"last": "Zaenen", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2004, |
|
"venue": "Proceedings of the AAAI Symposium on Exploring Attitude and Affect in Text: Theories and Applications (published as AAAI technical report", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Livia Polanyi and Annie Zaenen. Contextual valence shifters. In Proceedings of the AAAI Symposium on Exploring Attitude and Affect in Text: Theories and Applications (published as AAAI technical report SS-04-07), 2004.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "SemEval-2016 Task 5: Aspect Based Sentiment Analysis", |
|
"authors": [ |
|
{ |
|
"first": "Maria", |
|
"middle": [], |
|
"last": "Pontiki", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Dimitrios", |
|
"middle": [], |
|
"last": "Galanis", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Haris", |
|
"middle": [], |
|
"last": "Papageorgiou", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ion", |
|
"middle": [], |
|
"last": "Androutsopoulos", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Suresh", |
|
"middle": [], |
|
"last": "Manandhar", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Al-", |
|
"middle": [], |
|
"last": "Mohammad", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mahmoud", |
|
"middle": [], |
|
"last": "Smadi", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Yanyan", |
|
"middle": [], |
|
"last": "Al-Ayyoub", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Bing", |
|
"middle": [], |
|
"last": "Zhao", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Orph\u00e9e", |
|
"middle": [], |
|
"last": "Qin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "V\u00e9ronique", |
|
"middle": [], |
|
"last": "De Clercq", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Marianna", |
|
"middle": [], |
|
"last": "Hoste", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Xavier", |
|
"middle": [], |
|
"last": "Apidianaki", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Natalia", |
|
"middle": [], |
|
"last": "Tannier", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Evgeny", |
|
"middle": [], |
|
"last": "Loukachevitch", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Kotelnikov", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2016, |
|
"venue": "Proceedings of the 10th International Workshop on Semantic Evaluation, SemEval '16", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Maria Pontiki, Dimitrios Galanis, Haris Papageorgiou, Ion Androutsopoulos, Suresh Manandhar, Moham- mad AL-Smadi, Mahmoud Al-Ayyoub, Yanyan Zhao, Bing Qin, Orph\u00e9e De Clercq, V\u00e9ronique Hoste, Ma- rianna Apidianaki, Xavier Tannier, Natalia Louka- chevitch, Evgeny Kotelnikov, Nuria Bel, Salud Mar\u00eda Jim\u00e9nez Zafra, and G\u00fcl\u015fen Eryi\u011fit. 2016. SemEval- 2016 Task 5: Aspect Based Sentiment Analysis. In Proceedings of the 10th International Workshop on Semantic Evaluation, SemEval '16, San Diego, Cali- fornia, June 2016. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "NRC-Canada-2014: Detecting Aspects and Sentiment in Customer Reviews", |
|
"authors": [ |
|
{ |
|
"first": "Svetlana", |
|
"middle": [], |
|
"last": "Kiritchenko", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Xiaodan", |
|
"middle": [], |
|
"last": "Zhu", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Colin", |
|
"middle": [], |
|
"last": "Cherry", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Saif", |
|
"middle": [ |
|
"M" |
|
], |
|
"last": "Mohammad ; Srivastava", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": ".", |
|
"middle": [ |
|
"N" |
|
], |
|
"last": "Hinton", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "G", |
|
"middle": [], |
|
"last": "Krizhevsky", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "A", |
|
"middle": [], |
|
"last": "Sutskever", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "I", |
|
"middle": [], |
|
"last": "Salakhutdinov", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "R", |
|
"middle": [], |
|
"last": "", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2014, |
|
"venue": "Proceedings of the eight international workshop on Semantic Evaluation Exercises (SemEval-2014)", |
|
"volume": "15", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Svetlana Kiritchenko, Xiaodan Zhu, Colin Cherry, and Saif M. Mohammad. 2014. NRC-Canada-2014: De- tecting Aspects and Sentiment in Customer Reviews. in Proceedings of the eight international workshop on Semantic Evaluation Exercises (SemEval-2014), Au- gust 2014, Dublin, Ireland. Srivastava. N, Hinton, G, Krizhevsky, A, Sutskever, I, Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. Journal of Ma- chine Learning Research 15, 2014", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"type_str": "figure", |
|
"text": "System architecture", |
|
"uris": null, |
|
"num": null |
|
}, |
|
"TABREF1": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table/>", |
|
"text": "SO scores of n-grams" |
|
}, |
|
"TABREF2": { |
|
"num": null, |
|
"type_str": "table", |
|
"html": null, |
|
"content": "<table><tr><td/><td>restaurants</td><td>laptops</td></tr><tr><td>total accuracy</td><td>0.82838934</td><td>0.8030438</td></tr><tr><td>F1 positive</td><td>0.882567</td><td>0.8479</td></tr><tr><td>F1 negative</td><td>0.762167</td><td>0.796867</td></tr><tr><td>F1 neutral</td><td>0.0</td><td>0.0</td></tr><tr><td colspan=\"3\">Table 3: 3-fold cross-validation results</td></tr><tr><td/><td>restaurants</td><td>laptops</td></tr><tr><td>total accuracy</td><td>0.8393481</td><td>0.7790262</td></tr><tr><td>F1 positive</td><td>0.9078</td><td>0.8308</td></tr><tr><td>F1 negative</td><td>0.753</td><td>0.7588</td></tr><tr><td>F1 neutral</td><td>0.0</td><td>0.0</td></tr><tr><td colspan=\"2\">Table 4: Official results</td><td/></tr></table>", |
|
"text": "represents the results.The results on the official test set are given inTable 4. They are similar to our best results attained during the developments stage." |
|
} |
|
} |
|
} |
|
} |