entry_type
stringclasses 4
values | citation_key
stringlengths 10
110
| title
stringlengths 6
276
⌀ | editor
stringclasses 723
values | month
stringclasses 69
values | year
stringdate 1963-01-01 00:00:00
2022-01-01 00:00:00
| address
stringclasses 202
values | publisher
stringclasses 41
values | url
stringlengths 34
62
| author
stringlengths 6
2.07k
⌀ | booktitle
stringclasses 861
values | pages
stringlengths 1
12
⌀ | abstract
stringlengths 302
2.4k
| journal
stringclasses 5
values | volume
stringclasses 24
values | doi
stringlengths 20
39
⌀ | n
stringclasses 3
values | wer
stringclasses 1
value | uas
null | language
stringclasses 3
values | isbn
stringclasses 34
values | recall
null | number
stringclasses 8
values | a
null | b
null | c
null | k
null | f1
stringclasses 4
values | r
stringclasses 2
values | mci
stringclasses 1
value | p
stringclasses 2
values | sd
stringclasses 1
value | female
stringclasses 0
values | m
stringclasses 0
values | food
stringclasses 1
value | f
stringclasses 1
value | note
stringclasses 20
values | __index_level_0__
int64 22k
106k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
inproceedings | hewlett-etal-2017-accurate | Accurate Supervised and Semi-Supervised Machine Reading for Long Documents | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1214/ | Hewlett, Daniel and Jones, Llion and Lacoste, Alexandre and Gur, Izzeddin | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2011--2020 | We introduce a hierarchical architecture for machine reading capable of extracting precise information from long documents. The model divides the document into small, overlapping windows and encodes all windows in parallel with an RNN. It then attends over these window encodings, reducing them to a single encoding, which is decoded into an answer using a sequence decoder. This hierarchical approach allows the model to scale to longer documents without increasing the number of sequential steps. In a supervised setting, our model achieves state of the art accuracy of 76.8 on the WikiReading dataset. We also evaluate the model in a semi-supervised setting by downsampling the WikiReading training set to create increasingly smaller amounts of supervision, while leaving the full unlabeled document corpus to train a sequence autoencoder on document windows. We evaluate models that can reuse autoencoder states and outputs without fine-tuning their weights, allowing for more efficient training and inference. | null | null | 10.18653/v1/D17-1214 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,702 |
inproceedings | jia-liang-2017-adversarial | Adversarial Examples for Evaluating Reading Comprehension Systems | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1215/ | Jia, Robin and Liang, Percy | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2021--2031 | Standard accuracy metrics indicate that reading comprehension systems are making rapid progress, but the extent to which these systems truly understand language remains unclear. To reward systems with real language understanding abilities, we propose an adversarial evaluation scheme for the Stanford Question Answering Dataset (SQuAD). Our method tests whether systems can answer questions about paragraphs that contain adversarially inserted sentences, which are automatically generated to distract computer systems without changing the correct answer or misleading humans. In this adversarial setting, the accuracy of sixteen published models drops from an average of 75{\%} F1 score to 36{\%}; when the adversary is allowed to add ungrammatical sequences of words, average accuracy on four models decreases further to 7{\%}. We hope our insights will motivate the development of new models that understand language more precisely. | null | null | 10.18653/v1/D17-1215 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,703 |
inproceedings | lin-etal-2017-reasoning | Reasoning with Heterogeneous Knowledge for Commonsense Machine Comprehension | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1216/ | Lin, Hongyu and Sun, Le and Han, Xianpei | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2032--2043 | Reasoning with commonsense knowledge is critical for natural language understanding. Traditional methods for commonsense machine comprehension mostly only focus on one specific kind of knowledge, neglecting the fact that commonsense reasoning requires simultaneously considering different kinds of commonsense knowledge. In this paper, we propose a multi-knowledge reasoning method, which can exploit heterogeneous knowledge for commonsense machine comprehension. Specifically, we first mine different kinds of knowledge (including event narrative knowledge, entity semantic knowledge and sentiment coherent knowledge) and encode them as inference rules with costs. Then we propose a multi-knowledge reasoning model, which selects inference rules for a specific reasoning context using attention mechanism, and reasons by summarizing all valid inference rules. Experiments on RocStories show that our method outperforms traditional models significantly. | null | null | 10.18653/v1/D17-1216 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,704 |
inproceedings | yin-etal-2017-document | Document-Level Multi-Aspect Sentiment Classification as Machine Comprehension | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1217/ | Yin, Yichun and Song, Yangqiu and Zhang, Ming | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2044--2054 | Document-level multi-aspect sentiment classification is an important task for customer relation management. In this paper, we model the task as a machine comprehension problem where pseudo question-answer pairs are constructed by a small number of aspect-related keywords and aspect ratings. A hierarchical iterative attention model is introduced to build aspectspecific representations by frequent and repeated interactions between documents and aspect questions. We adopt a hierarchical architecture to represent both word level and sentence level information, and use the attention operations for aspect questions and documents alternatively with the multiple hop mechanism. Experimental results on the TripAdvisor and BeerAdvocate datasets show that our model outperforms classical baselines. We will release our code and data for the method replicability. | null | null | 10.18653/v1/D17-1217 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,705 |
inproceedings | daxenberger-etal-2017-essence | What is the Essence of a Claim? Cross-Domain Claim Identification | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1218/ | Daxenberger, Johannes and Eger, Steffen and Habernal, Ivan and Stab, Christian and Gurevych, Iryna | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2055--2066 | Argument mining has become a popular research area in NLP. It typically includes the identification of argumentative components, e.g. claims, as the central component of an argument. We perform a qualitative analysis across six different datasets and show that these appear to conceptualize claims quite differently. To learn about the consequences of such different conceptualizations of claim for practical applications, we carried out extensive experiments using state-of-the-art feature-rich and deep learning systems, to identify claims in a cross-domain fashion. While the divergent conceptualization of claims in different datasets is indeed harmful to cross-domain classification, we show that there are shared properties on the lexical level as well as system configurations that can help to overcome these gaps. | null | null | 10.18653/v1/D17-1218 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,706 |
inproceedings | du-cardie-2017-identifying | Identifying Where to Focus in Reading Comprehension for Neural Question Generation | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1219/ | Du, Xinya and Cardie, Claire | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2067--2073 | A first step in the task of automatically generating questions for testing reading comprehension is to identify \textit{question-worthy} sentences, i.e. sentences in a text passage that humans find it worthwhile to ask questions about. We propose a hierarchical neural sentence-level sequence tagging model for this task, which existing approaches to question generation have ignored. The approach is fully data-driven {---} with no sophisticated NLP pipelines or any hand-crafted rules/features {---} and compares favorably to a number of baselines when evaluated on the SQuAD data set. When incorporated into an existing neural question generation system, the resulting end-to-end system achieves state-of-the-art performance for paragraph-level question generation for reading comprehension. | null | null | 10.18653/v1/D17-1219 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,707 |
inproceedings | sterckx-etal-2017-break | Break it Down for Me: A Study in Automated Lyric Annotation | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1220/ | Sterckx, Lucas and Naradowsky, Jason and Byrne, Bill and Demeester, Thomas and Develder, Chris | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2074--2080 | Comprehending lyrics, as found in songs and poems, can pose a challenge to human and machine readers alike. This motivates the need for systems that can understand the ambiguity and jargon found in such creative texts, and provide commentary to aid readers in reaching the correct interpretation. We introduce the task of automated lyric annotation (ALA). Like text simplification, a goal of ALA is to rephrase the original text in a more easily understandable manner. However, in ALA the system must often include additional information to clarify niche terminology and abstract concepts. To stimulate research on this task, we release a large collection of crowdsourced annotations for song lyrics. We analyze the performance of translation and retrieval models on this task, measuring performance with both automated and human evaluation. We find that each model captures a unique type of information important to the task. | null | null | 10.18653/v1/D17-1220 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,708 |
inproceedings | li-etal-2017-cascaded | Cascaded Attention based Unsupervised Information Distillation for Compressive Summarization | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1221/ | Li, Piji and Lam, Wai and Bing, Lidong and Guo, Weiwei and Li, Hang | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2081--2090 | When people recall and digest what they have read for writing summaries, the important content is more likely to attract their attention. Inspired by this observation, we propose a cascaded attention based unsupervised model to estimate the salience information from the text for compressive multi-document summarization. The attention weights are learned automatically by an unsupervised data reconstruction framework which can capture the sentence salience. By adding sparsity constraints on the number of output vectors, we can generate condensed information which can be treated as word salience. Fine-grained and coarse-grained sentence compression strategies are incorporated to produce compressive summaries. Experiments on some benchmark data sets show that our framework achieves better results than the state-of-the-art methods. | null | null | 10.18653/v1/D17-1221 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,709 |
inproceedings | li-etal-2017-deep | Deep Recurrent Generative Decoder for Abstractive Text Summarization | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1222/ | Li, Piji and Lam, Wai and Bing, Lidong and Wang, Zihao | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2091--2100 | We propose a new framework for abstractive text summarization based on a sequence-to-sequence oriented encoder-decoder model equipped with a deep recurrent generative decoder (DRGN). Latent structure information implied in the target summaries is learned based on a recurrent latent random model for improving the summarization quality. Neural variational inference is employed to address the intractable posterior inference for the recurrent latent variables. Abstractive summaries are generated based on both the generative latent variables and the discriminative deterministic states. Extensive experiments on some benchmark datasets in different languages show that DRGN achieves improvements over the state-of-the-art methods. | null | null | 10.18653/v1/D17-1222 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,710 |
inproceedings | isonuma-etal-2017-extractive | Extractive Summarization Using Multi-Task Learning with Document Classification | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1223/ | Isonuma, Masaru and Fujino, Toru and Mori, Junichiro and Matsuo, Yutaka and Sakata, Ichiro | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2101--2110 | The need for automatic document summarization that can be used for practical applications is increasing rapidly. In this paper, we propose a general framework for summarization that extracts sentences from a document using externally related information. Our work is aimed at single document summarization using small amounts of reference summaries. In particular, we address document summarization in the framework of multi-task learning using curriculum learning for sentence extraction and document classification. The proposed framework enables us to obtain better feature representations to extract sentences from documents. We evaluate our proposed summarization method on two datasets: financial report and news corpus. Experimental results demonstrate that our summarizers achieve performance that is comparable to state-of-the-art systems. | null | null | 10.18653/v1/D17-1223 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,711 |
inproceedings | zhang-wan-2017-towards | Towards Automatic Construction of News Overview Articles by News Synthesis | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1224/ | Zhang, Jianmin and Wan, Xiaojun | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2111--2116 | In this paper we investigate a new task of automatically constructing an overview article from a given set of news articles about a news event. We propose a news synthesis approach to address this task based on passage segmentation, ranking, selection and merging. Our proposed approach is compared with several typical multi-document summarization methods on the Wikinews dataset, and achieves the best performance on both automatic evaluation and manual evaluation. | null | null | 10.18653/v1/D17-1224 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,712 |
inproceedings | zhao-huang-2017-joint | Joint Syntacto-Discourse Parsing and the Syntacto-Discourse Treebank | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1225/ | Zhao, Kai and Huang, Liang | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2117--2123 | Discourse parsing has long been treated as a stand-alone problem independent from constituency or dependency parsing. Most attempts at this problem rely on annotated text segmentations (Elementary Discourse Units, EDUs) and sophisticated sparse or continuous features to extract syntactic information. In this paper we propose the first end-to-end discourse parser that jointly parses in both syntax and discourse levels, as well as the first syntacto-discourse treebank by integrating the Penn Treebank and the RST Treebank. Built upon our recent span-based constituency parser, this joint syntacto-discourse parser requires no preprocessing efforts such as segmentation or feature extraction, making discourse parsing more convenient. Empirically, our parser achieves the state-of-the-art end-to-end discourse parsing accuracy. | null | null | 10.18653/v1/D17-1225 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,713 |
inproceedings | choubey-huang-2017-event | Event Coreference Resolution by Iteratively Unfolding Inter-dependencies among Events | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1226/ | Choubey, Prafulla Kumar and Huang, Ruihong | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2124--2133 | We introduce a novel iterative approach for event coreference resolution that gradually builds event clusters by exploiting inter-dependencies among event mentions within the same chain as well as across event chains. Among event mentions in the same chain, we distinguish within- and cross-document event coreference links by using two distinct pairwise classifiers, trained separately to capture differences in feature distributions of within- and cross-document event clusters. Our event coreference approach alternates between WD and CD clustering and combines arguments from both event clusters after every merge, continuing till no more merge can be made. And then it performs further merging between event chains that are both closely related to a set of other chains of events. Experiments on the ECB+ corpus show that our model outperforms state-of-the-art methods in joint task of WD and CD event coreference resolution. | null | null | 10.18653/v1/D17-1226 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,714 |
inproceedings | huang-etal-2017-finish | When to Finish? Optimal Beam Search for Neural Text Generation (modulo beam size) | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1227/ | Huang, Liang and Zhao, Kai and Ma, Mingbo | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2134--2139 | In neural text generation such as neural machine translation, summarization, and image captioning, beam search is widely used to improve the output text quality. However, in the neural generation setting, hypotheses can finish in different steps, which makes it difficult to decide when to end beam search to ensure optimality. We propose a provably optimal beam search algorithm that will always return the optimal-score complete hypothesis (modulo beam size), and finish as soon as the optimality is established. To counter neural generation`s tendency for shorter hypotheses, we also introduce a bounded length reward mechanism which allows a modified version of our beam search algorithm to remain optimal. Experiments on neural machine translation demonstrate that our principled beam search algorithm leads to improvement in BLEU score over previously proposed alternatives. | null | null | 10.18653/v1/D17-1227 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,715 |
inproceedings | wang-etal-2017-steering | Steering Output Style and Topic in Neural Response Generation | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1228/ | Wang, Di and Jojic, Nebojsa and Brockett, Chris and Nyberg, Eric | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2140--2150 | We propose simple and flexible training and decoding methods for influencing output style and topic in neural encoder-decoder based language generation. This capability is desirable in a variety of applications, including conversational systems, where successful agents need to produce language in a specific style and generate responses steered by a human puppeteer or external knowledge. We decompose the neural generation process into empirically easier sub-problems: a faithfulness model and a decoding method based on selective-sampling. We also describe training and sampling algorithms that bias the generation process with a specific language style restriction, or a topic restriction. Human evaluation results show that our proposed methods are able to to restrict style and topic without degrading output quality in conversational tasks. | null | null | 10.18653/v1/D17-1228 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,716 |
inproceedings | tran-etal-2017-preserving | Preserving Distributional Information in Dialogue Act Classification | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1229/ | Tran, Quan Hung and Zukerman, Ingrid and Haffari, Gholamreza | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2151--2156 | This paper introduces a novel training/decoding strategy for sequence labeling. Instead of greedily choosing a label at each time step, and using it for the next prediction, we retain the probability distribution over the current label, and pass this distribution to the next prediction. This approach allows us to avoid the effect of label bias and error propagation in sequence learning/decoding. Our experiments on dialogue act classification demonstrate the effectiveness of this approach. Even though our underlying neural network model is relatively simple, it outperforms more complex neural models, achieving state-of-the-art results on the MapTask and Switchboard corpora. | null | null | 10.18653/v1/D17-1229 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,717 |
inproceedings | li-etal-2017-adversarial | Adversarial Learning for Neural Dialogue Generation | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1230/ | Li, Jiwei and Monroe, Will and Shi, Tianlin and Jean, S{\'e}bastien and Ritter, Alan and Jurafsky, Dan | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2157--2169 | We apply adversarial training to open-domain dialogue generation, training a system to produce sequences that are indistinguishable from human-generated dialogue utterances. We cast the task as a reinforcement learning problem where we jointly train two systems: a generative model to produce response sequences, and a discriminator{---}analagous to the human evaluator in the Turing test{---} to distinguish between the human-generated dialogues and the machine-generated ones. In this generative adversarial network approach, the outputs from the discriminator are used to encourage the system towards more human-like dialogue. Further, we investigate models for adversarial evaluation that uses success in fooling an adversary as a dialogue evaluation metric, while avoiding a number of potential pitfalls. Experimental results on several metrics, including adversarial evaluation, demonstrate that the adversarially-trained system generates higher-quality responses than previous baselines | null | null | 10.18653/v1/D17-1230 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,718 |
inproceedings | liu-etal-2017-using-context | Using Context Information for Dialog Act Classification in {DNN} Framework | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1231/ | Liu, Yang and Han, Kun and Tan, Zhao and Lei, Yun | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2170--2178 | Previous work on dialog act (DA) classification has investigated different methods, such as hidden Markov models, maximum entropy, conditional random fields, graphical models, and support vector machines. A few recent studies explored using deep learning neural networks for DA classification, however, it is not clear yet what is the best method for using dialog context or DA sequential information, and how much gain it brings. This paper proposes several ways of using context information for DA classification, all in the deep learning framework. The baseline system classifies each utterance using the convolutional neural networks (CNN). Our proposed methods include using hierarchical models (recurrent neural networks (RNN) or CNN) for DA sequence tagging where the bottom layer takes the sentence CNN representation as input, concatenating predictions from the previous utterances with the CNN vector for classification, and performing sequence decoding based on the predictions from the sentence CNN model. We conduct thorough experiments and comparisons on the Switchboard corpus, demonstrate that incorporating context information significantly improves DA classification, and show that we achieve new state-of-the-art performance for this task. | null | null | 10.18653/v1/D17-1231 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,719 |
inproceedings | jo-etal-2017-modeling | Modeling Dialogue Acts with Content Word Filtering and Speaker Preferences | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1232/ | Jo, Yohan and Yoder, Michael and Jang, Hyeju and Ros{\'e}, Carolyn | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2179--2189 | We present an unsupervised model of dialogue act sequences in conversation. By modeling topical themes as transitioning more slowly than dialogue acts in conversation, our model de-emphasizes content-related words in order to focus on conversational function words that signal dialogue acts. We also incorporate speaker tendencies to use some acts more than others as an additional predictor of dialogue act prevalence beyond temporal dependencies. According to the evaluation presented on two dissimilar corpora, the CNET forum and NPS Chat corpus, the effectiveness of each modeling assumption is found to vary depending on characteristics of the data. De-emphasizing content-related words yields improvement on the CNET corpus, while utilizing speaker tendencies is advantageous on the NPS corpus. The components of our model complement one another to achieve robust performance on both corpora and outperform state-of-the-art baseline models. | null | null | 10.18653/v1/D17-1232 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,720 |
inproceedings | yao-etal-2017-towards | Towards Implicit Content-Introducing for Generative Short-Text Conversation Systems | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1233/ | Yao, Lili and Zhang, Yaoyuan and Feng, Yansong and Zhao, Dongyan and Yan, Rui | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2190--2199 | The study on human-computer conversation systems is a hot research topic nowadays. One of the prevailing methods to build the system is using the generative Sequence-to-Sequence (Seq2Seq) model through neural networks. However, the standard Seq2Seq model is prone to generate trivial responses. In this paper, we aim to generate a more meaningful and informative reply when answering a given question. We propose an implicit content-introducing method which incorporates additional information into the Seq2Seq model in a flexible way. Specifically, we fuse the general decoding and the auxiliary cue word information through our proposed hierarchical gated fusion unit. Experiments on real-life data demonstrate that our model consistently outperforms a set of competitive baselines in terms of BLEU scores and human evaluation. | null | null | 10.18653/v1/D17-1233 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,721 |
inproceedings | chang-etal-2017-affordable | Affordable On-line Dialogue Policy Learning | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1234/ | Chang, Cheng and Yang, Runzhe and Chen, Lu and Zhou, Xiang and Yu, Kai | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2200--2209 | The key to building an evolvable dialogue system in real-world scenarios is to ensure an affordable on-line dialogue policy learning, which requires the on-line learning process to be safe, efficient and economical. But in reality, due to the scarcity of real interaction data, the dialogue system usually grows slowly. Besides, the poor initial dialogue policy easily leads to bad user experience and incurs a failure of attracting users to contribute training data, so that the learning process is unsustainable. To accurately depict this, two quantitative metrics are proposed to assess safety and efficiency issues. For solving the unsustainable learning problem, we proposed a complete companion teaching framework incorporating the guidance from the human teacher. Since the human teaching is expensive, we compared various teaching schemes answering the question how and when to teach, to economically utilize teaching budget, so that make the online learning process affordable. | null | null | 10.18653/v1/D17-1234 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,722 |
inproceedings | shao-etal-2017-generating | Generating High-Quality and Informative Conversation Responses with Sequence-to-Sequence Models | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1235/ | Shao, Yuanlong and Gouws, Stephan and Britz, Denny and Goldie, Anna and Strope, Brian and Kurzweil, Ray | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2210--2219 | Sequence-to-sequence models have been applied to the conversation response generation problem where the source sequence is the conversation history and the target sequence is the response. Unlike translation, conversation responding is inherently creative. The generation of long, informative, coherent, and diverse responses remains a hard task. In this work, we focus on the single turn setting. We add self-attention to the decoder to maintain coherence in longer responses, and we propose a practical approach, called the glimpse-model, for scaling to large datasets. We introduce a stochastic beam-search algorithm with segment-by-segment reranking which lets us inject diversity earlier in the generation process. We trained on a combined data set of over 2.3B conversation messages mined from the web. In human evaluation studies, our method produces longer responses overall, with a higher proportion rated as acceptable and excellent as length increases, compared to baseline sequence-to-sequence models with explicit length-promotion. A back-off strategy produces better responses overall, in the full spectrum of lengths. | null | null | 10.18653/v1/D17-1235 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,723 |
inproceedings | eshghi-etal-2017-bootstrapping | Bootstrapping incremental dialogue systems from minimal data: the generalisation power of dialogue grammars | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1236/ | Eshghi, Arash and Shalyminov, Igor and Lemon, Oliver | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2220--2230 | We investigate an end-to-end method for automatically inducing task-based dialogue systems from small amounts of unannotated dialogue data. It combines an incremental semantic grammar - Dynamic Syntax and Type Theory with Records (DS-TTR) - with Reinforcement Learning (RL), where language generation and dialogue management are a joint decision problem. The systems thus produced are incremental: dialogues are processed word-by-word, shown previously to be essential in supporting natural, spontaneous dialogue. We hypothesised that the rich linguistic knowledge within the grammar should enable a combinatorially large number of dialogue variations to be processed, even when trained on very few dialogues. Our experiments show that our model can process 74{\%} of the Facebook AI bAbI dataset even when trained on only 0.13{\%} of the data (5 dialogues). It can in addition process 65{\%} of bAbI+, a corpus we created by systematically adding incremental dialogue phenomena such as restarts and self-corrections to bAbI. We compare our model with a state-of-the-art retrieval model, MEMN2N. We find that, in terms of semantic accuracy, the MEMN2N model shows very poor robustness to the bAbI+ transformations even when trained on the full bAbI dataset. | null | null | 10.18653/v1/D17-1236 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,724 |
inproceedings | peng-etal-2017-composite | Composite Task-Completion Dialogue Policy Learning via Hierarchical Deep Reinforcement Learning | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1237/ | Peng, Baolin and Li, Xiujun and Li, Lihong and Gao, Jianfeng and Celikyilmaz, Asli and Lee, Sungjin and Wong, Kam-Fai | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2231--2240 | Building a dialogue agent to fulfill complex tasks, such as travel planning, is challenging because the agent has to learn to collectively complete multiple subtasks. For example, the agent needs to reserve a hotel and book a flight so that there leaves enough time for commute between arrival and hotel check-in. This paper addresses this challenge by formulating the task in the mathematical framework of options over Markov Decision Processes (MDPs), and proposing a hierarchical deep reinforcement learning approach to learning a dialogue manager that operates at different temporal scales. The dialogue manager consists of: (1) a top-level dialogue policy that selects among subtasks or options, (2) a low-level dialogue policy that selects primitive actions to complete the subtask given by the top-level policy, and (3) a global state tracker that helps ensure all cross-subtask constraints be satisfied. Experiments on a travel planning task with simulated and real users show that our approach leads to significant improvements over three baselines, two based on handcrafted rules and the other based on flat deep reinforcement learning. | null | null | 10.18653/v1/D17-1237 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,725 |
inproceedings | novikova-etal-2017-need | Why We Need New Evaluation Metrics for {NLG} | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1238/ | Novikova, Jekaterina and Du{\v{s}}ek, Ond{\v{r}}ej and Cercas Curry, Amanda and Rieser, Verena | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2241--2252 | The majority of NLG evaluation relies on automatic metrics, such as BLEU . In this paper, we motivate the need for novel, system- and data-independent automatic evaluation methods: We investigate a wide range of metrics, including state-of-the-art word-based and novel grammar-based ones, and demonstrate that they only weakly reflect human judgements of system outputs as generated by data-driven, end-to-end NLG. We also show that metric performance is data- and system-specific. Nevertheless, our results also suggest that automatic metrics perform reliably at system-level and can support system development by finding cases where a system performs poorly. | null | null | 10.18653/v1/D17-1238 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,726 |
inproceedings | wiseman-etal-2017-challenges | Challenges in Data-to-Document Generation | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1239/ | Wiseman, Sam and Shieber, Stuart and Rush, Alexander | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2253--2263 | Recent neural models have shown significant progress on the problem of generating short descriptive texts conditioned on a small number of database records. In this work, we suggest a slightly more difficult data-to-text generation task, and investigate how effective current approaches are on this task. In particular, we introduce a new, large-scale corpus of data records paired with descriptive documents, propose a series of extractive evaluation methods for analyzing performance, and obtain baseline results using current neural generation methods. Experiments show that these models produce fluent text, but fail to convincingly approximate human-generated documents. Moreover, even templated baselines exceed the performance of these neural models on some metrics, though copy- and reconstruction-based extensions lead to noticeable improvements. | null | null | 10.18653/v1/D17-1239 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,727 |
inproceedings | patro-etal-2017-english | All that is {E}nglish may be {H}indi: Enhancing language identification through automatic ranking of the likeliness of word borrowing in social media | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1240/ | Patro, Jasabanta and Samanta, Bidisha and Singh, Saurabh and Basu, Abhipsa and Mukherjee, Prithwish and Choudhury, Monojit and Mukherjee, Animesh | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2264--2274 | n this paper, we present a set of computational methods to identify the likeliness of a word being borrowed, based on the signals from social media. In terms of Spearman`s correlation values, our methods perform more than two times better ({\ensuremath{\sim}} 0.62) in predicting the borrowing likeliness compared to the best performing baseline ({\ensuremath{\sim}} 0.26) reported in literature. Based on this likeliness estimate we asked annotators to re-annotate the language tags of foreign words in predominantly native contexts. In 88{\%} of cases the annotators felt that the foreign language tag should be replaced by native language tag, thus indicating a huge scope for improvement of automatic language identification systems. | null | null | 10.18653/v1/D17-1240 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,728 |
inproceedings | ding-etal-2017-multi | Multi-View Unsupervised User Feature Embedding for Social Media-based Substance Use Prediction | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1241/ | Ding, Tao and Bickel, Warren K. and Pan, Shimei | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2275--2284 | In this paper, we demonstrate how the state-of-the-art machine learning and text mining techniques can be used to build effective social media-based substance use detection systems. Since a substance use ground truth is difficult to obtain on a large scale, to maximize system performance, we explore different unsupervised feature learning methods to take advantage of a large amount of unsupervised social media data. We also demonstrate the benefit of using multi-view unsupervised feature learning to combine heterogeneous user information such as Facebook {\textquotedblleft}likes{\textquotedblright} and {\textquotedblleft}status updates{\textquotedblright} to enhance system performance. Based on our evaluation, our best models achieved 86{\%} AUC for predicting tobacco use, 81{\%} for alcohol use and 84{\%} for illicit drug use, all of which significantly outperformed existing methods. Our investigation has also uncovered interesting relations between a user`s social media behavior (e.g., word usage) and substance use. | null | null | 10.18653/v1/D17-1241 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,729 |
inproceedings | garimella-etal-2017-demographic | Demographic-aware word associations | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1242/ | Garimella, Aparna and Banea, Carmen and Mihalcea, Rada | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2285--2295 | Variations of word associations across different groups of people can provide insights into people`s psychologies and their world views. To capture these variations, we introduce the task of demographic-aware word associations. We build a new gold standard dataset consisting of word association responses for approximately 300 stimulus words, collected from more than 800 respondents of different gender (male/female) and from different locations (India/United States), and show that there are significant variations in the word associations made by these groups. We also introduce a new demographic-aware word association model based on a neural net skip-gram architecture, and show how computational methods for measuring word associations that specifically account for writer demographics can outperform generic methods that are agnostic to such information. | null | null | 10.18653/v1/D17-1242 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,730 |
inproceedings | cheng-etal-2017-factored | A Factored Neural Network Model for Characterizing Online Discussions in Vector Space | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1243/ | Cheng, Hao and Fang, Hao and Ostendorf, Mari | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2296--2306 | We develop a novel factored neural model that learns comment embeddings in an unsupervised way leveraging the structure of distributional context in online discussion forums. The model links different context with related language factors in the embedding space, providing a way to interpret the factored embeddings. Evaluated on a community endorsement prediction task using a large collection of topic-varying Reddit discussions, the factored embeddings consistently achieve improvement over other text representations. Qualitative analysis shows that the model captures community style and topic, as well as response trigger patterns. | null | null | 10.18653/v1/D17-1243 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,731 |
inproceedings | rashid-blanco-2017-dimensions | Dimensions of Interpersonal Relationships: Corpus and Experiments | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1244/ | Rashid, Farzana and Blanco, Eduardo | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2307--2316 | This paper presents a corpus and experiments to determine dimensions of interpersonal relationships. We define a set of dimensions heavily inspired by work in social science. We create a corpus by retrieving pairs of people, and then annotating dimensions for their relationships. A corpus analysis shows that dimensions can be annotated reliably. Experimental results show that given a pair of people, values to dimensions can be assigned automatically. | null | null | 10.18653/v1/D17-1244 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,732 |
inproceedings | dusmanu-etal-2017-argument | Argument Mining on {T}witter: Arguments, Facts and Sources | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1245/ | Dusmanu, Mihai and Cabrio, Elena and Villata, Serena | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2317--2322 | Social media collect and spread on the Web personal opinions, facts, fake news and all kind of information users may be interested in. Applying argument mining methods to such heterogeneous data sources is a challenging open research issue, in particular considering the peculiarities of the language used to write textual messages on social media. In addition, new issues emerge when dealing with arguments posted on such platforms, such as the need to make a distinction between personal opinions and actual facts, and to detect the source disseminating information about such facts to allow for provenance verification. In this paper, we apply supervised classification to identify arguments on Twitter, and we present two new tasks for argument mining, namely facts recognition and source identification. We study the feasibility of the approaches proposed to address these tasks on a set of tweets related to the Grexit and Brexit news topics. | null | null | 10.18653/v1/D17-1245 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,733 |
inproceedings | aoki-etal-2017-distinguishing | Distinguishing {J}apanese Non-standard Usages from Standard Ones | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1246/ | Aoki, Tatsuya and Sasano, Ryohei and Takamura, Hiroya and Okumura, Manabu | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2323--2328 | We focus on non-standard usages of common words on social media. In the context of social media, words sometimes have other usages that are totally different from their original. In this study, we attempt to distinguish non-standard usages on social media from standard ones in an unsupervised manner. Our basic idea is that non-standardness can be measured by the inconsistency between the expected meaning of the target word and the given context. For this purpose, we use context embeddings derived from word embeddings. Our experimental results show that the model leveraging the context embedding outperforms other methods and provide us with findings, for example, on how to construct context embeddings and which corpus to use. | null | null | 10.18653/v1/D17-1246 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,734 |
inproceedings | sap-etal-2017-connotation | Connotation Frames of Power and Agency in Modern Films | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1247/ | Sap, Maarten and Prasettio, Marcella Cindy and Holtzman, Ari and Rashkin, Hannah and Choi, Yejin | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2329--2334 | The framing of an action influences how we perceive its actor. We introduce connotation frames of power and agency, a pragmatic formalism organized using frame semantic representations, to model how different levels of power and agency are implicitly projected on actors through their actions. We use the new power and agency frames to measure the subtle, but prevalent, gender bias in the portrayal of modern film characters and provide insights that deviate from the well-known Bechdel test. Our contributions include an extended lexicon of connotation frames along with a web interface that provides a comprehensive analysis through the lens of connotation frames. | null | null | 10.18653/v1/D17-1247 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,735 |
inproceedings | preotiuc-pietro-etal-2017-controlling | Controlling Human Perception of Basic User Traits | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1248/ | Preo{\c{t}}iuc-Pietro, Daniel and Chandra Guntuku, Sharath and Ungar, Lyle | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2335--2341 | Much of our online communication is text-mediated and, lately, more common with automated agents. Unlike interacting with humans, these agents currently do not tailor their language to the type of person they are communicating to. In this pilot study, we measure the extent to which human perception of basic user trait information {--} gender and age {--} is controllable through text. Using automatic models of gender and age prediction, we estimate which tweets posted by a user are more likely to mis-characterize his traits. We perform multiple controlled crowdsourcing experiments in which we show that we can reduce the human prediction accuracy of gender to almost random {--} an over 20{\%} drop in accuracy. Our experiments show that it is practically feasible for multiple applications such as text generation, text summarization or machine translation to be tailored to specific traits and perceived as such. | null | null | 10.18653/v1/D17-1248 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,736 |
inproceedings | gautrais-etal-2017-topic | Topic Signatures in Political Campaign Speeches | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1249/ | Gautrais, Cl{\'e}ment and Cellier, Peggy and Quiniou, Ren{\'e} and Termier, Alexandre | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2342--2347 | Highlighting the recurrence of topics usage in candidates speeches is a key feature to identify the main ideas of each candidate during a political campaign. In this paper, we present a method combining standard topic modeling with signature mining for analyzing topic recurrence in speeches of Clinton and Trump during the 2016 American presidential campaign. The results show that the method extracts automatically the main ideas of each candidate and, in addition, provides information about the evolution of these topics during the campaign. | null | null | 10.18653/v1/D17-1249 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,737 |
inproceedings | schwartz-etal-2017-assessing | Assessing Objective Recommendation Quality through Political Forecasting | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1250/ | Schwartz, H. Andrew and Rouhizadeh, Masoud and Bishop, Michael and Tetlock, Philip and Mellers, Barbara and Ungar, Lyle | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2348--2357 | Recommendations are often rated for their subjective quality, but few researchers have studied comment quality in terms of objective utility. We explore recommendation quality assessment with respect to both subjective (i.e. users' ratings) and objective (i.e., did it influence? did it improve decisions?) metrics in a massive online geopolitical forecasting system, ultimately comparing linguistic characteristics of each quality metric. Using a variety of features, we predict all types of quality with better accuracy than the simple yet strong baseline of comment length. Looking at the most predictive content illustrates rater biases; for example, forecasters are subjectively biased in favor of comments mentioning business transactions or dealings as well as material things, even though such comments do not indeed prove any more useful objectively. Additionally, more complex sentence constructions, as evidenced by subordinate conjunctions, are characteristic of comments leading to objective improvements in forecasting. | null | null | 10.18653/v1/D17-1250 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,738 |
inproceedings | shirakawa-etal-2017-never | Never Abandon Minorities: Exhaustive Extraction of Bursty Phrases on Microblogs Using Set Cover Problem | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1251/ | Shirakawa, Masumi and Hara, Takahiro and Maekawa, Takuya | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2358--2367 | We propose a language-independent data-driven method to exhaustively extract bursty phrases of arbitrary forms (e.g., phrases other than simple noun phrases) from microblogs. The burst (i.e., the rapid increase of the occurrence) of a phrase causes the burst of overlapping N-grams including incomplete ones. In other words, bursty incomplete N-grams inevitably overlap bursty phrases. Thus, the proposed method performs the extraction of bursty phrases as the set cover problem in which all bursty N-grams are covered by a minimum set of bursty phrases. Experimental results using Japanese Twitter data showed that the proposed method outperformed word-based, noun phrase-based, and segmentation-based methods both in terms of accuracy and coverage. | null | null | 10.18653/v1/D17-1251 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,739 |
inproceedings | peng-etal-2017-maximum | Maximum Margin Reward Networks for Learning from Explicit and Implicit Supervision | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1252/ | Peng, Haoruo and Chang, Ming-Wei and Yih, Wen-tau | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2368--2378 | Neural networks have achieved state-of-the-art performance on several structured-output prediction tasks, trained in a fully supervised fashion. However, annotated examples in structured domains are often costly to obtain, which thus limits the applications of neural networks. In this work, we propose Maximum Margin Reward Networks, a neural network-based framework that aims to learn from both explicit (full structures) and implicit supervision signals (delayed feedback on the correctness of the predicted structure). On named entity recognition and semantic parsing, our model outperforms previous systems on the benchmark datasets, CoNLL-2003 and WebQuestionsSP. | null | null | 10.18653/v1/D17-1252 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,740 |
inproceedings | wachsmuth-etal-2017-impact | The Impact of Modeling Overall Argumentation with Tree Kernels | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1253/ | Wachsmuth, Henning and Da San Martino, Giovanni and Kiesel, Dora and Stein, Benno | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2379--2389 | Several approaches have been proposed to model either the explicit sequential structure of an argumentative text or its implicit hierarchical structure. So far, the adequacy of these models of overall argumentation remains unclear. This paper asks what type of structure is actually important to tackle downstream tasks in computational argumentation. We analyze patterns in the overall argumentation of texts from three corpora. Then, we adapt the idea of positional tree kernels in order to capture sequential and hierarchical argumentative structure together for the first time. In systematic experiments for three text classification tasks, we find strong evidence for the impact of both types of structure. Our results suggest that either of them is necessary while their combination may be beneficial. | null | null | 10.18653/v1/D17-1253 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,741 |
inproceedings | gan-etal-2017-learning | Learning Generic Sentence Representations Using Convolutional Neural Networks | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1254/ | Gan, Zhe and Pu, Yunchen and Henao, Ricardo and Li, Chunyuan and He, Xiaodong and Carin, Lawrence | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2390--2400 | We propose a new encoder-decoder approach to learn distributed sentence representations that are applicable to multiple purposes. The model is learned by using a convolutional neural network as an encoder to map an input sentence into a continuous vector, and using a long short-term memory recurrent neural network as a decoder. Several tasks are considered, including sentence reconstruction and future sentence prediction. Further, a hierarchical encoder-decoder model is proposed to encode a sentence to predict multiple future sentences. By training our models on a large collection of novels, we obtain a highly generic convolutional sentence encoder that performs well in practice. Experimental results on several benchmark datasets, and across a broad range of applications, demonstrate the superiority of the proposed model over competing methods. | null | null | 10.18653/v1/D17-1254 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,742 |
inproceedings | amiri-etal-2017-repeat | Repeat before Forgetting: Spaced Repetition for Efficient and Effective Training of Neural Networks | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1255/ | Amiri, Hadi and Miller, Timothy and Savova, Guergana | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2401--2410 | We present a novel approach for training artificial neural networks. Our approach is inspired by broad evidence in psychology that shows human learners can learn efficiently and effectively by increasing intervals of time between subsequent reviews of previously learned materials (spaced repetition). We investigate the analogy between training neural models and findings in psychology about human memory model and develop an efficient and effective algorithm to train neural models. The core part of our algorithm is a cognitively-motivated scheduler according to which training instances and their {\textquotedblleft}reviews{\textquotedblright} are spaced over time. Our algorithm uses only 34-50{\%} of data per epoch, is 2.9-4.8 times faster than standard training, and outperforms competing state-of-the-art baselines. Our code is available at \url{scholar.harvard.edu/hadi/RbF/}. | null | null | 10.18653/v1/D17-1255 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,743 |
inproceedings | gui-etal-2017-part | Part-of-Speech Tagging for {T}witter with Adversarial Neural Networks | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1256/ | Gui, Tao and Zhang, Qi and Huang, Haoran and Peng, Minlong and Huang, Xuanjing | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2411--2420 | In this work, we study the problem of part-of-speech tagging for Tweets. In contrast to newswire articles, Tweets are usually informal and contain numerous out-of-vocabulary words. Moreover, there is a lack of large scale labeled datasets for this domain. To tackle these challenges, we propose a novel neural network to make use of out-of-domain labeled data, unlabeled in-domain data, and labeled in-domain data. Inspired by adversarial neural networks, the proposed method tries to learn common features through adversarial discriminator. In addition, we hypothesize that domain-specific features of target domain should be preserved in some degree. Hence, the proposed method adopts a sequence-to-sequence autoencoder to perform this task. Experimental results on three different datasets show that our method achieves better performance than state-of-the-art methods. | null | null | 10.18653/v1/D17-1256 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,744 |
inproceedings | li-etal-2017-investigating | Investigating Different Syntactic Context Types and Context Representations for Learning Word Embeddings | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1257/ | Li, Bofang and Liu, Tao and Zhao, Zhe and Tang, Buzhou and Drozd, Aleksandr and Rogers, Anna and Du, Xiaoyong | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2421--2431 | The number of word embedding models is growing every year. Most of them are based on the co-occurrence information of words and their contexts. However, it is still an open question what is the best definition of context. We provide a systematical investigation of 4 different syntactic context types and context representations for learning word embeddings. Comprehensive experiments are conducted to evaluate their effectiveness on 6 extrinsic and intrinsic tasks. We hope that this paper, along with the published code, would be helpful for choosing the best context type and representation for a given task. | null | null | 10.18653/v1/D17-1257 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,745 |
inproceedings | braud-etal-2017-syntax | Does syntax help discourse segmentation? Not so much | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1258/ | Braud, Chlo{\'e} and Lacroix, Oph{\'e}lie and S{\o}gaard, Anders | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2432--2442 | Discourse segmentation is the first step in building discourse parsers. Most work on discourse segmentation does not scale to real-world discourse parsing across languages, for two reasons: (i) models rely on constituent trees, and (ii) experiments have relied on gold standard identification of sentence and token boundaries. We therefore investigate to what extent constituents can be replaced with universal dependencies, or left out completely, as well as how state-of-the-art segmenters fare in the absence of sentence boundaries. Our results show that dependency information is less useful than expected, but we provide a fully scalable, robust model that only relies on part-of-speech information, and show that it performs well across languages in the absence of any gold-standard annotation. | null | null | 10.18653/v1/D17-1258 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,746 |
inproceedings | lewis-etal-2017-deal | Deal or No Deal? End-to-End Learning of Negotiation Dialogues | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1259/ | Lewis, Mike and Yarats, Denis and Dauphin, Yann and Parikh, Devi and Batra, Dhruv | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2443--2453 | Much of human dialogue occurs in semi-cooperative settings, where agents with different goals attempt to agree on common decisions. Negotiations require complex communication and reasoning skills, but success is easy to measure, making this an interesting task for AI. We gather a large dataset of human-human negotiations on a multi-issue bargaining task, where agents who cannot observe each other`s reward functions must reach an agreement (or a deal) via natural language dialogue. For the first time, we show it is possible to train end-to-end models for negotiation, which must learn both linguistic and reasoning skills with no annotated dialogue states. We also introduce dialogue rollouts, in which the model plans ahead by simulating possible complete continuations of the conversation, and find that this technique dramatically improves performance. Our code and dataset are publicly available. | null | null | 10.18653/v1/D17-1259 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,747 |
inproceedings | chen-etal-2017-agent | Agent-Aware Dropout {DQN} for Safe and Efficient On-line Dialogue Policy Learning | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1260/ | Chen, Lu and Zhou, Xiang and Chang, Cheng and Yang, Runzhe and Yu, Kai | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2454--2464 | Hand-crafted rules and reinforcement learning (RL) are two popular choices to obtain dialogue policy. The rule-based policy is often reliable within predefined scope but not self-adaptable, whereas RL is evolvable with data but often suffers from a bad initial performance. We employ a \textit{companion learning} framework to integrate the two approaches for \textit{on-line} dialogue policy learning, in which a pre-defined rule-based policy acts as a {\textquotedblleft}teacher{\textquotedblright} and guides a data-driven RL system by giving example actions as well as additional rewards. A novel \textit{agent-aware dropout} Deep Q-Network (AAD-DQN) is proposed to address the problem of when to consult the teacher and how to learn from the teacher`s experiences. AAD-DQN, as a data-driven student policy, provides (1) two separate experience memories for student and teacher, (2) an uncertainty estimated by dropout to control the timing of consultation and learning. Simulation experiments showed that the proposed approach can significantly improve both \textit{safety}and \textit{efficiency} of on-line policy optimization compared to other companion learning approaches as well as supervised pre-training using static dialogue corpus. | null | null | 10.18653/v1/D17-1260 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,748 |
inproceedings | potash-rumshisky-2017-towards | Towards Debate Automation: a Recurrent Model for Predicting Debate Winners | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1261/ | Potash, Peter and Rumshisky, Anna | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2465--2475 | In this paper we introduce a practical first step towards the creation of an automated debate agent: a state-of-the-art recurrent predictive model for predicting debate winners. By having an accurate predictive model, we are able to objectively rate the quality of a statement made at a specific turn in a debate. The model is based on a recurrent neural network architecture with attention, which allows the model to effectively account for the entire debate when making its prediction. Our model achieves state-of-the-art accuracy on a dataset of debate transcripts annotated with audience favorability of the debate teams. Finally, we discuss how future work can leverage our proposed model for the creation of an automated debate agent. We accomplish this by determining the model input that will maximize audience favorability toward a given side of a debate at an arbitrary turn. | null | null | 10.18653/v1/D17-1261 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,749 |
inproceedings | ma-etal-2017-investigation | Further Investigation into Reference Bias in Monolingual Evaluation of Machine Translation | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1262/ | Ma, Qingsong and Graham, Yvette and Baldwin, Timothy and Liu, Qun | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2476--2485 | Monolingual evaluation of Machine Translation (MT) aims to simplify human assessment by requiring assessors to compare the meaning of the MT output with a reference translation, opening up the task to a much larger pool of genuinely qualified evaluators. Monolingual evaluation runs the risk, however, of bias in favour of MT systems that happen to produce translations superficially similar to the reference and, consistent with this intuition, previous investigations have concluded monolingual assessment to be strongly biased in this respect. On re-examination of past analyses, we identify a series of potential analytical errors that force some important questions to be raised about the reliability of past conclusions, however. We subsequently carry out further investigation into reference bias via direct human assessment of MT adequacy via quality controlled crowd-sourcing. Contrary to both intuition and past conclusions, results for show no significant evidence of reference bias in monolingual evaluation of MT. | null | null | 10.18653/v1/D17-1262 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,750 |
inproceedings | isabelle-etal-2017-challenge | A Challenge Set Approach to Evaluating Machine Translation | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1263/ | Isabelle, Pierre and Cherry, Colin and Foster, George | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2486--2496 | Neural machine translation represents an exciting leap forward in translation quality. But what longstanding weaknesses does it resolve, and which remain? We address these questions with a challenge set approach to translation evaluation and error analysis. A challenge set consists of a small set of sentences, each hand-designed to probe a system`s capacity to bridge a particular structural divergence between languages. To exemplify this approach, we present an English-French challenge set, and use it to analyze phrase-based and neural systems. The resulting analysis provides not only a more fine-grained picture of the strengths of neural systems, but also insight into which linguistic phenomena remain out of reach. | null | null | 10.18653/v1/D17-1263 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,751 |
inproceedings | nakashole-flauger-2017-knowledge | Knowledge Distillation for Bilingual Dictionary Induction | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1264/ | Nakashole, Ndapandula and Flauger, Raphael | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2497--2506 | Leveraging zero-shot learning to learn mapping functions between vector spaces of different languages is a promising approach to bilingual dictionary induction. However, methods using this approach have not yet achieved high accuracy on the task. In this paper, we propose a bridging approach, where our main contribution is a knowledge distillation training objective. As teachers, rich resource translation paths are exploited in this role. And as learners, translation paths involving low resource languages learn from the teachers. Our training objective allows seamless addition of teacher translation paths for any given low resource pair. Since our approach relies on the quality of monolingual word embeddings, we also propose to enhance vector representations of both the source and target language with linguistic information. Our experiments on various languages show large performance gains from our distillation training objective, obtaining as high as 17{\%} accuracy improvements. | null | null | 10.18653/v1/D17-1264 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,752 |
inproceedings | bawden-2017-machine | Machine Translation, it`s a question of style, innit? The case of {E}nglish tag questions | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1265/ | Bawden, Rachel | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2507--2512 | In this paper, we address the problem of generating English tag questions (TQs) (e.g. it is, isn`t it?) in Machine Translation (MT). We propose a post-edition solution, formulating the problem as a multi-class classification task. We present (i) the automatic annotation of English TQs in a parallel corpus of subtitles and (ii) an approach using a series of classifiers to predict TQ forms, which we use to post-edit state-of-the-art MT outputs. Our method provides significant improvements in English TQ translation when translating from Czech, French and German, in turn improving the fluidity, naturalness, grammatical correctness and pragmatic coherence of MT output. | null | null | 10.18653/v1/D17-1265 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,753 |
inproceedings | pourdamghani-knight-2017-deciphering | Deciphering Related Languages | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1266/ | Pourdamghani, Nima and Knight, Kevin | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2513--2518 | We present a method for translating texts between close language pairs. The method does not require parallel data, and it does not require the languages to be written in the same script. We show results for six language pairs: Afrikaans/Dutch, Bosnian/Serbian, Danish/Swedish, Macedonian/Bulgarian, Malaysian/Indonesian, and Polish/Belorussian. We report BLEU scores showing our method to outperform others that do not use parallel data. | null | null | 10.18653/v1/D17-1266 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,754 |
inproceedings | st-arnaud-etal-2017-identifying | Identifying Cognate Sets Across Dictionaries of Related Languages | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1267/ | St Arnaud, Adam and Beck, David and Kondrak, Grzegorz | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2519--2528 | We present a system for identifying cognate sets across dictionaries of related languages. The likelihood of a cognate relationship is calculated on the basis of a rich set of features that capture both phonetic and semantic similarity, as well as the presence of regular sound correspondences. The similarity scores are used to cluster words from different languages that may originate from a common proto-word. When tested on the Algonquian language family, our system detects 63{\%} of cognate sets while maintaining cluster purity of 70{\%}. | null | null | 10.18653/v1/D17-1267 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,755 |
inproceedings | malaviya-etal-2017-learning | Learning Language Representations for Typology Prediction | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1268/ | Malaviya, Chaitanya and Neubig, Graham and Littell, Patrick | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2529--2535 | One central mystery of neural NLP is what neural models {\textquotedblleft}know{\textquotedblright} about their subject matter. When a neural machine translation system learns to translate from one language to another, does it learn the syntax or semantics of the languages? Can this knowledge be extracted from the system to fill holes in human scientific knowledge? Existing typological databases contain relatively full feature specifications for only a few hundred languages. Exploiting the existence of parallel texts in more than a thousand languages, we build a massive many-to-one NMT system from 1017 languages into English, and use this to predict information missing from typological databases. Experiments show that the proposed method is able to infer not only syntactic, but also phonological and phonetic inventory features, and improves over a baseline that has access to information about the languages geographic and phylogenetic neighbors. | null | null | 10.18653/v1/D17-1268 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,756 |
inproceedings | mayhew-etal-2017-cheap | Cheap Translation for Cross-Lingual Named Entity Recognition | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1269/ | Mayhew, Stephen and Tsai, Chen-Tse and Roth, Dan | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2536--2545 | Recent work in NLP has attempted to deal with low-resource languages but still assumed a resource level that is not present for most languages, e.g., the availability of Wikipedia in the target language. We propose a simple method for cross-lingual named entity recognition (NER) that works well in settings with \textit{very} minimal resources. Our approach makes use of a lexicon to {\textquotedblleft}translate{\textquotedblright} annotated data available in one or several high resource language(s) into the target language, and learns a standard monolingual NER model there. Further, when Wikipedia is available in the target language, our method can enhance Wikipedia based methods to yield state-of-the-art NER results; we evaluate on 7 diverse languages, improving the state-of-the-art by an average of 5.5{\%} F1 points. With the minimal resources required, this is an extremely portable cross-lingual NER approach, as illustrated using a truly low-resource language, Uyghur. | null | null | 10.18653/v1/D17-1269 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,757 |
inproceedings | vulic-etal-2017-cross | Cross-Lingual Induction and Transfer of Verb Classes Based on Word Vector Space Specialisation | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1270/ | Vuli{\'c}, Ivan and Mrk{\v{s}}i{\'c}, Nikola and Korhonen, Anna | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2546--2558 | Existing approaches to automatic VerbNet-style verb classification are heavily dependent on feature engineering and therefore limited to languages with mature NLP pipelines. In this work, we propose a novel cross-lingual transfer method for inducing VerbNets for multiple languages. To the best of our knowledge, this is the first study which demonstrates how the architectures for learning word embeddings can be applied to this challenging syntactic-semantic task. Our method uses cross-lingual translation pairs to tie each of the six target languages into a bilingual vector space with English, jointly specialising the representations to encode the relational information from English VerbNet. A standard clustering algorithm is then run on top of the VerbNet-specialised representations, using vector dimensions as features for learning verb classes. Our results show that the proposed cross-lingual transfer approach sets new state-of-the-art verb classification performance across all six target languages explored in this work. | null | null | 10.18653/v1/D17-1270 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,758 |
inproceedings | friedrich-gateva-2017-classification | Classification of telicity using cross-linguistic annotation projection | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1271/ | Friedrich, Annemarie and Gateva, Damyana | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2559--2565 | This paper addresses the automatic recognition of telicity, an aspectual notion. A telic event includes a natural endpoint ({\textquotedblleft}she walked home{\textquotedblright}), while an atelic event does not ({\textquotedblleft}she walked around{\textquotedblright}). Recognizing this difference is a prerequisite for temporal natural language understanding. In English, this classification task is difficult, as telicity is a covert linguistic category. In contrast, in Slavic languages, aspect is part of a verb`s meaning and even available in machine-readable dictionaries. Our contributions are as follows. We successfully leverage additional silver standard training data in the form of projected annotations from parallel English-Czech data as well as context information, improving automatic telicity classification for English significantly compared to previous work. We also create a new data set of English texts manually annotated with telicity. | null | null | 10.18653/v1/D17-1271 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,759 |
inproceedings | lawrence-etal-2017-counterfactual | Counterfactual Learning from Bandit Feedback under Deterministic Logging : A Case Study in Statistical Machine Translation | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1272/ | Lawrence, Carolin and Sokolov, Artem and Riezler, Stefan | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2566--2576 | The goal of counterfactual learning for statistical machine translation (SMT) is to optimize a target SMT system from logged data that consist of user feedback to translations that were predicted by another, historic SMT system. A challenge arises by the fact that risk-averse commercial SMT systems deterministically log the most probable translation. The lack of sufficient exploration of the SMT output space seemingly contradicts the theoretical requirements for counterfactual learning. We show that counterfactual learning from deterministic bandit logs is possible nevertheless by smoothing out deterministic components in learning. This can be achieved by additive and multiplicative control variates that avoid degenerate behavior in empirical risk minimization. Our simulation experiments show improvements of up to 2 BLEU points by counterfactual learning from deterministic bandit feedback. | null | null | 10.18653/v1/D17-1272 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,760 |
inproceedings | wang-etal-2017-learning-fine | Learning Fine-grained Relations from {C}hinese User Generated Categories | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1273/ | Wang, Chengyu and Fan, Yan and He, Xiaofeng and Zhou, Aoying | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2577--2587 | User generated categories (UGCs) are short texts that reflect how people describe and organize entities, expressing rich semantic relations implicitly. While most methods on UGC relation extraction are based on pattern matching in English circumstances, learning relations from Chinese UGCs poses different challenges due to the flexibility of expressions. In this paper, we present a weakly supervised learning framework to harvest relations from Chinese UGCs. We identify is-a relations via word embedding based projection and inference, extract non-taxonomic relations and their category patterns by graph mining. We conduct experiments on Chinese Wikipedia and achieve high accuracy, outperforming state-of-the-art methods. | null | null | 10.18653/v1/D17-1273 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,761 |
inproceedings | huang-etal-2017-improving | Improving Slot Filling Performance with Attentive Neural Networks on Dependency Structures | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1274/ | Huang, Lifu and Sil, Avirup and Ji, Heng and Florian, Radu | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2588--2597 | Slot Filling (SF) aims to extract the values of certain types of attributes (or slots, such as person:cities{\_}of{\_}residence) for a given entity from a large collection of source documents. In this paper we propose an effective DNN architecture for SF with the following new strategies: (1). Take a regularized dependency graph instead of a raw sentence as input to DNN, to compress the wide contexts between query and candidate filler; (2). Incorporate two attention mechanisms: local attention learned from query and candidate filler, and global attention learned from external knowledge bases, to guide the model to better select indicative contexts to determine slot type. Experiments show that this framework outperforms state-of-the-art on both relation extraction (16{\%} absolute F-score gain) and slot filling validation for each individual system (up to 8.5{\%} absolute F-score gain). | null | null | 10.18653/v1/D17-1274 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,762 |
inproceedings | durrett-etal-2017-identifying | Identifying Products in Online Cybercrime Marketplaces: A Dataset for Fine-grained Domain Adaptation | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1275/ | Durrett, Greg and Kummerfeld, Jonathan K. and Berg-Kirkpatrick, Taylor and Portnoff, Rebecca and Afroz, Sadia and McCoy, Damon and Levchenko, Kirill and Paxson, Vern | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2598--2607 | One weakness of machine-learned NLP models is that they typically perform poorly on out-of-domain data. In this work, we study the task of identifying products being bought and sold in online cybercrime forums, which exhibits particularly challenging cross-domain effects. We formulate a task that represents a hybrid of slot-filling information extraction and named entity recognition and annotate data from four different forums. Each of these forums constitutes its own {\textquotedblleft}fine-grained domain{\textquotedblright} in that the forums cover different market sectors with different properties, even though all forums are in the broad domain of cybercrime. We characterize these domain differences in the context of a learning-based system: supervised models see decreased accuracy when applied to new forums, and standard techniques for semi-supervised learning and domain adaptation have limited effectiveness on this data, which suggests the need to improve these techniques. We release a dataset of 1,938 annotated posts from across the four forums. | null | null | 10.18653/v1/D17-1275 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,763 |
inproceedings | muis-lu-2017-labeling | Labeling Gaps Between Words: Recognizing Overlapping Mentions with Mention Separators | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1276/ | Muis, Aldrian Obaja and Lu, Wei | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2608--2618 | In this paper, we propose a new model that is capable of recognizing overlapping mentions. We introduce a novel notion of mention separators that can be effectively used to capture how mentions overlap with one another. On top of a novel multigraph representation that we introduce, we show that efficient and exact inference can still be performed. We present some theoretical analysis on the differences between our model and a recently proposed model for recognizing overlapping mentions, and discuss the possible implications of the differences. Through extensive empirical analysis on standard datasets, we demonstrate the effectiveness of our approach. | null | null | 10.18653/v1/D17-1276 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,764 |
inproceedings | ganea-hofmann-2017-deep | Deep Joint Entity Disambiguation with Local Neural Attention | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1277/ | Ganea, Octavian-Eugen and Hofmann, Thomas | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2619--2629 | We propose a novel deep learning model for joint document-level entity disambiguation, which leverages learned neural representations. Key components are entity embeddings, a neural attention mechanism over local context windows, and a differentiable joint inference stage for disambiguation. Our approach thereby combines benefits of deep learning with more traditional approaches such as graphical models and probabilistic mention-entity maps. Extensive experiments show that we are able to obtain competitive or state-of-the-art accuracy at moderate computational costs. | null | null | 10.18653/v1/D17-1277 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,765 |
inproceedings | gashteovski-etal-2017-minie | {M}in{IE}: Minimizing Facts in Open Information Extraction | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1278/ | Gashteovski, Kiril and Gemulla, Rainer and del Corro, Luciano | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2630--2640 | The goal of Open Information Extraction (OIE) is to extract surface relations and their arguments from natural-language text in an unsupervised, domain-independent manner. In this paper, we propose MinIE, an OIE system that aims to provide useful, compact extractions with high precision and recall. MinIE approaches these goals by (1) representing information about polarity, modality, attribution, and quantities with semantic annotations instead of in the actual extraction, and (2) identifying and removing parts that are considered overly specific. We conducted an experimental study with several real-world datasets and found that MinIE achieves competitive or higher precision and recall than most prior systems, while at the same time producing shorter, semantically enriched extractions. | null | null | 10.18653/v1/D17-1278 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,766 |
inproceedings | luan-etal-2017-scientific | Scientific Information Extraction with Semi-supervised Neural Tagging | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1279/ | Luan, Yi and Ostendorf, Mari and Hajishirzi, Hannaneh | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2641--2651 | This paper addresses the problem of extracting keyphrases from scientific articles and categorizing them as corresponding to a task, process, or material. We cast the problem as sequence tagging and introduce semi-supervised methods to a neural tagging model, which builds on recent advances in named entity recognition. Since annotated training data is scarce in this domain, we introduce a graph-based semi-supervised algorithm together with a data selection scheme to leverage unannotated articles. Both inductive and transductive semi-supervised learning strategies outperform state-of-the-art information extraction performance on the 2017 SemEval Task 10 ScienceIE task. | null | null | 10.18653/v1/D17-1279 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,767 |
inproceedings | tang-etal-2017-nite | {NITE}: A Neural Inductive Teaching Framework for Domain Specific {NER} | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1280/ | Tang, Siliang and Zhang, Ning and Zhang, Jinjiang and Wu, Fei and Zhuang, Yueting | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2652--2657 | In domain-specific NER, due to insufficient labeled training data, deep models usually fail to behave normally. In this paper, we proposed a novel Neural Inductive TEaching framework (NITE) to transfer knowledge from existing domain-specific NER models into an arbitrary deep neural network in a teacher-student training manner. NITE is a general framework that builds upon transfer learning and multiple instance learning, which collaboratively not only transfers knowledge to a deep student network but also reduces the noise from teachers. NITE can help deep learning methods to effectively utilize existing resources (i.e., models, labeled and unlabeled data) in a small domain. The experiment resulted on Disease NER proved that without using any labeled data, NITE can significantly boost the performance of a CNN-bidirectional LSTM-CRF NER neural network nearly over 30{\%} in terms of F1-score. | null | null | 10.18653/v1/D17-1280 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,768 |
inproceedings | sharma-etal-2017-speeding | Speeding up Reinforcement Learning-based Information Extraction Training using Asynchronous Methods | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1281/ | Sharma, Aditya and Parekh, Zarana and Talukdar, Partha | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2658--2663 | RLIE-DQN is a recently proposed Reinforcement Learning-based Information Extraction (IE) technique which is able to incorporate external evidence during the extraction process. RLIE-DQN trains a single agent sequentially, training on one instance at a time. This results in significant training slowdown which is undesirable. We leverage recent advances in parallel RL training using asynchronous methods and propose RLIE-A3C. RLIE-A3C trains multiple agents in parallel and is able to achieve upto 6x training speedup over RLIE-DQN, while suffering no loss in average accuracy. | null | null | 10.18653/v1/D17-1281 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,769 |
inproceedings | li-etal-2017-leveraging | Leveraging Linguistic Structures for Named Entity Recognition with Bidirectional Recursive Neural Networks | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1282/ | Li, Peng-Hsuan and Dong, Ruo-Ping and Wang, Yu-Siang and Chou, Ju-Chieh and Ma, Wei-Yun | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2664--2669 | In this paper, we utilize the linguistic structures of texts to improve named entity recognition by BRNN-CNN, a special bidirectional recursive network attached with a convolutional network. Motivated by the observation that named entities are highly related to linguistic constituents, we propose a constituent-based BRNN-CNN for named entity recognition. In contrast to classical sequential labeling methods, the system first identifies which text chunks are possible named entities by whether they are linguistic constituents. Then it classifies these chunks with a constituency tree structure by recursively propagating syntactic and semantic information to each constituent node. This method surpasses current state-of-the-art on OntoNotes 5.0 with automatically generated parses. | null | null | 10.18653/v1/D17-1282 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,770 |
inproceedings | strubell-etal-2017-fast | Fast and Accurate Entity Recognition with Iterated Dilated Convolutions | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1283/ | Strubell, Emma and Verga, Patrick and Belanger, David and McCallum, Andrew | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2670--2680 | Today when many practitioners run basic NLP on the entire web and large-volume traffic, faster methods are paramount to saving time and energy costs. Recent advances in GPU hardware have led to the emergence of bi-directional LSTMs as a standard method for obtaining per-token vector representations serving as input to labeling tasks such as NER (often followed by prediction in a linear-chain CRF). Though expressive and accurate, these models fail to fully exploit GPU parallelism, limiting their computational efficiency. This paper proposes a faster alternative to Bi-LSTMs for NER: Iterated Dilated Convolutional Neural Networks (ID-CNNs), which have better capacity than traditional CNNs for large context and structured prediction. Unlike LSTMs whose sequential processing on sentences of length N requires O(N) time even in the face of parallelism, ID-CNNs permit fixed-depth convolutions to run in parallel across entire documents. We describe a distinct combination of network structure, parameter sharing and training procedures that enable dramatic 14-20x test-time speedups while retaining accuracy comparable to the Bi-LSTM-CRF. Moreover, ID-CNNs trained to aggregate context from the entire document are more accurate than Bi-LSTM-CRFs while attaining 8x faster test time speeds. | null | null | 10.18653/v1/D17-1283 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,771 |
inproceedings | gupta-etal-2017-entity | Entity Linking via Joint Encoding of Types, Descriptions, and Context | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1284/ | Gupta, Nitish and Singh, Sameer and Roth, Dan | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2681--2690 | For accurate entity linking, we need to capture various information aspects of an entity, such as its description in a KB, contexts in which it is mentioned, and structured knowledge. Additionally, a linking system should work on texts from different domains without requiring domain-specific training data or hand-engineered features. In this work we present a neural, modular entity linking system that learns a unified dense representation for each entity using multiple sources of information, such as its description, contexts around its mentions, and its fine-grained types. We show that the resulting entity linking system is effective at combining these sources, and performs competitively, sometimes out-performing current state-of-the-art systems across datasets, without requiring any domain-specific training data or hand-engineered features. We also show that our model can effectively {\textquotedblleft}embed{\textquotedblright} entities that are new to the KB, and is able to link its mentions accurately. | null | null | 10.18653/v1/D17-1284 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,772 |
inproceedings | he-etal-2017-insight | An Insight Extraction System on {B}io{M}edical Literature with Deep Neural Networks | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1285/ | He, Hua and Ganjam, Kris and Jain, Navendu and Lundin, Jessica and White, Ryen and Lin, Jimmy | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2691--2701 | Mining biomedical text offers an opportunity to automatically discover important facts and infer associations among them. As new scientific findings appear across a large collection of biomedical publications, our aim is to tap into this literature to automate biomedical knowledge extraction and identify important insights from them. Towards that goal, we develop a system with novel deep neural networks to extract insights on biomedical literature. Evaluation shows our system is able to provide insights with competitive accuracy of human acceptance and its relation extraction component outperforms previous work. | null | null | 10.18653/v1/D17-1285 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,773 |
inproceedings | nastase-strapparava-2017-word | Word Etymology as Native Language Interference | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1286/ | Nastase, Vivi and Strapparava, Carlo | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2702--2707 | We present experiments that show the influence of native language on lexical choice when producing text in another language {--} in this particular case English. We start from the premise that non-native English speakers will choose lexical items that are close to words in their native language. This leads us to an etymology-based representation of documents written by people whose mother tongue is an Indo-European language. Based on this representation we grow a language family tree, that matches closely the Indo-European language tree. | null | null | 10.18653/v1/D17-1286 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,774 |
inproceedings | eisenberg-finlayson-2017-simpler | A Simpler and More Generalizable Story Detector using Verb and Character Features | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1287/ | Eisenberg, Joshua and Finlayson, Mark | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2708--2715 | Story detection is the task of determining whether or not a unit of text contains a story. Prior approaches achieved a maximum performance of 0.66 F1, and did not generalize well across different corpora. We present a new state-of-the-art detector that achieves a maximum performance of 0.75 F1 (a 14{\%} improvement), with significantly greater generalizability than previous work. In particular, our detector achieves performance above 0.70 F1 across a variety of combinations of lexically different corpora for training and testing, as well as dramatic improvements (up to 4,000{\%}) in performance when trained on a small, disfluent data set. The new detector uses two basic types of features{--}ones related to events, and ones related to characters{--}totaling 283 specific features overall; previous detectors used tens of thousands of features, and so this detector represents a significant simplification along with increased performance. | null | null | 10.18653/v1/D17-1287 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,775 |
inproceedings | schulz-kuhn-2017-multi | Multi-modular domain-tailored {OCR} post-correction | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1288/ | Schulz, Sarah and Kuhn, Jonas | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2716--2726 | One of the main obstacles for many Digital Humanities projects is the low data availability. Texts have to be digitized in an expensive and time consuming process whereas Optical Character Recognition (OCR) post-correction is one of the time-critical factors. At the example of OCR post-correction, we show the adaptation of a generic system to solve a specific problem with little data. The system accounts for a diversity of errors encountered in OCRed texts coming from different time periods in the domain of literature. We show that the combination of different approaches, such as e.g. Statistical Machine Translation and spell checking, with the help of a ranking mechanism tremendously improves over single-handed approaches. Since we consider the accessibility of the resulting tool as a crucial part of Digital Humanities collaborations, we describe the workflow we suggest for efficient text recognition and subsequent automatic and manual post-correction | null | null | 10.18653/v1/D17-1288 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,776 |
inproceedings | luo-etal-2017-learning | Learning to Predict Charges for Criminal Cases with Legal Basis | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1289/ | Luo, Bingfeng and Feng, Yansong and Xu, Jianbo and Zhang, Xiang and Zhao, Dongyan | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2727--2736 | The charge prediction task is to determine appropriate charges for a given case, which is helpful for legal assistant systems where the user input is fact description. We argue that relevant law articles play an important role in this task, and therefore propose an attention-based neural network method to jointly model the charge prediction task and the relevant article extraction task in a unified framework. The experimental results show that, besides providing legal basis, the relevant articles can also clearly improve the charge prediction results, and our full model can effectively predict appropriate charges for cases with different expression styles. | null | null | 10.18653/v1/D17-1289 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,777 |
inproceedings | schofield-etal-2017-quantifying | Quantifying the Effects of Text Duplication on Semantic Models | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1290/ | Schofield, Alexandra and Thompson, Laure and Mimno, David | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2737--2747 | Duplicate documents are a pervasive problem in text datasets and can have a strong effect on unsupervised models. Methods to remove duplicate texts are typically heuristic or very expensive, so it is vital to know when and why they are needed. We measure the sensitivity of two latent semantic methods to the presence of different levels of document repetition. By artificially creating different forms of duplicate text we confirm several hypotheses about how repeated text impacts models. While a small amount of duplication is tolerable, substantial over-representation of subsets of the text may overwhelm meaningful topical patterns. | null | null | 10.18653/v1/D17-1290 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,778 |
inproceedings | zhuang-etal-2017-identifying | Identifying Semantically Deviating Outlier Documents | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1291/ | Zhuang, Honglei and Wang, Chi and Tao, Fangbo and Kaplan, Lance and Han, Jiawei | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2748--2757 | A document outlier is a document that substantially deviates in semantics from the majority ones in a corpus. Automatic identification of document outliers can be valuable in many applications, such as screening health records for medical mistakes. In this paper, we study the problem of mining semantically deviating document outliers in a given corpus. We develop a generative model to identify frequent and characteristic semantic regions in the word embedding space to represent the given corpus, and a robust outlierness measure which is resistant to noisy content in documents. Experiments conducted on two real-world textual data sets show that our method can achieve an up to 135{\%} improvement over baselines in terms of recall at top-1{\%} of the outlier ranking. | null | null | 10.18653/v1/D17-1291 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,779 |
inproceedings | kang-etal-2017-detecting | Detecting and Explaining Causes From Text For a Time Series Event | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1292/ | Kang, Dongyeop and Gangal, Varun and Lu, Ang and Chen, Zheng and Hovy, Eduard | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2758--2767 | Explaining underlying causes or effects about events is a challenging but valuable task. We define a novel problem of generating explanations of a time series event by (1) searching cause and effect relationships of the time series with textual data and (2) constructing a connecting chain between them to generate an explanation. To detect causal features from text, we propose a novel method based on the Granger causality of time series between features extracted from text such as N-grams, topics, sentiments, and their composition. The generation of the sequence of causal entities requires a commonsense causative knowledge base with efficient reasoning. To ensure good interpretability and appropriate lexical usage we combine symbolic and neural representations, using a neural reasoning algorithm trained on commonsense causal tuples to predict the next cause step. Our quantitative and human analysis show empirical evidence that our method successfully extracts meaningful causality relationships between time series with textual features and generates appropriate explanation between them. | null | null | 10.18653/v1/D17-1292 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,780 |
inproceedings | jiang-etal-2017-novel | A Novel Cascade Model for Learning Latent Similarity from Heterogeneous Sequential Data of {MOOC} | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1293/ | Jiang, Zhuoxuan and Feng, Shanshan and Cong, Gao and Miao, Chunyan and Li, Xiaoming | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2768--2773 | Recent years have witnessed the proliferation of Massive Open Online Courses (MOOCs). With massive learners being offered MOOCs, there is a demand that the forum contents within MOOCs need to be classified in order to facilitate both learners and instructors. Therefore we investigate a significant application, which is to associate forum threads to subtitles of video clips. This task can be regarded as a document ranking problem, and the key is how to learn a distinguishable text representation from word sequences and learners' behavior sequences. In this paper, we propose a novel cascade model, which can capture both the latent semantics and latent similarity by modeling MOOC data. Experimental results on two real-world datasets demonstrate that our textual representation outperforms state-of-the-art unsupervised counterparts for the application. | null | null | 10.18653/v1/D17-1293 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,781 |
inproceedings | mysore-sathyendra-etal-2017-identifying | Identifying the Provision of Choices in Privacy Policy Text | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1294/ | Mysore Sathyendra, Kanthashree and Wilson, Shomir and Schaub, Florian and Zimmeck, Sebastian and Sadeh, Norman | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2774--2779 | Websites' and mobile apps' privacy policies, written in natural language, tend to be long and difficult to understand. Information privacy revolves around the fundamental principle of Notice and choice, namely the idea that users should be able to make informed decisions about what information about them can be collected and how it can be used. Internet users want control over their privacy, but their choices are often hidden in long and convoluted privacy policy texts. Moreover, little (if any) prior work has been done to detect the provision of choices in text. We address this challenge of enabling user choice by automatically identifying and extracting pertinent choice language in privacy policies. In particular, we present a two-stage architecture of classification models to identify opt-out choices in privacy policy text, labelling common varieties of choices with a mean F1 score of 0.735. Our techniques enable the creation of systems to help Internet users to learn about their choices, thereby effectuating notice and choice and improving Internet privacy. | null | null | 10.18653/v1/D17-1294 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,782 |
inproceedings | goyal-etal-2017-empirical | An Empirical Analysis of Edit Importance between Document Versions | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1295/ | Goyal, Tanya and Kelkar, Sachin and Agarwal, Manas and Grover, Jeenu | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2780--2784 | In this paper, we present a novel approach to infer significance of various textual edits to documents. An author may make several edits to a document; each edit varies in its impact to the content of the document. While some edits are surface changes and introduce negligible change, other edits may change the content/tone of the document significantly. In this paper, we perform an analysis on the human perceptions of edit importance while reviewing documents from one version to the next. We identify linguistic features that influence edit importance and model it in a regression based setting. We show that the predicted importance by our approach is highly correlated with the human perceived importance, established by a Mechanical Turk study. | null | null | 10.18653/v1/D17-1295 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,783 |
inproceedings | wang-etal-2017-transition | Transition-Based Disfluency Detection using {LSTM}s | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1296/ | Wang, Shaolei and Che, Wanxiang and Zhang, Yue and Zhang, Meishan and Liu, Ting | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2785--2794 | In this paper, we model the problem of disfluency detection using a transition-based framework, which incrementally constructs and labels the disfluency chunk of input sentences using a new transition system without syntax information. Compared with sequence labeling methods, it can capture non-local chunk-level features; compared with joint parsing and disfluency detection methods, it is free for noise in syntax. Experiments show that our model achieves state-of-the-art f-score of 87.5{\%} on the commonly used English Switchboard test set, and a set of in-house annotated Chinese data. | null | null | 10.18653/v1/D17-1296 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,784 |
inproceedings | yannakoudakis-etal-2017-neural | Neural Sequence-Labelling Models for Grammatical Error Correction | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1297/ | Yannakoudakis, Helen and Rei, Marek and Andersen, {\O}istein E. and Yuan, Zheng | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2795--2806 | We propose an approach to N-best list reranking using neural sequence-labelling models. We train a compositional model for error detection that calculates the probability of each token in a sentence being correct or incorrect, utilising the full sentence as context. Using the error detection model, we then re-rank the N best hypotheses generated by statistical machine translation systems. Our approach achieves state-of-the-art results on error correction for three different datasets, and it has the additional advantage of only using a small set of easily computed features that require no linguistic input. | null | null | 10.18653/v1/D17-1297 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,785 |
inproceedings | schmaltz-etal-2017-adapting | Adapting Sequence Models for Sentence Correction | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1298/ | Schmaltz, Allen and Kim, Yoon and Rush, Alexander and Shieber, Stuart | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2807--2813 | In a controlled experiment of sequence-to-sequence approaches for the task of sentence correction, we find that character-based models are generally more effective than word-based models and models that encode subword information via convolutions, and that modeling the output data as a series of diffs improves effectiveness over standard approaches. Our strongest sequence-to-sequence model improves over our strongest phrase-based statistical machine translation model, with access to the same data, by 6 M2 (0.5 GLEU) points. Additionally, in the data environment of the standard CoNLL-2014 setup, we demonstrate that modeling (and tuning against) diffs yields similar or better M2 scores with simpler models and/or significantly less data than previous sequence-to-sequence approaches. | null | null | 10.18653/v1/D17-1298 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,786 |
inproceedings | niu-etal-2017-study | A Study of Style in Machine Translation: Controlling the Formality of Machine Translation Output | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1299/ | Niu, Xing and Martindale, Marianna and Carpuat, Marine | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2814--2819 | Stylistic variations of language, such as formality, carry speakers' intention beyond literal meaning and should be conveyed adequately in translation. We propose to use lexical formality models to control the formality level of machine translation output. We demonstrate the effectiveness of our approach in empirical evaluations, as measured by automatic metrics and human assessments. | null | null | 10.18653/v1/D17-1299 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,787 |
inproceedings | devlin-2017-sharp | Sharp Models on Dull Hardware: Fast and Accurate Neural Machine Translation Decoding on the {CPU} | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1300/ | Devlin, Jacob | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2820--2825 | Attentional sequence-to-sequence models have become the new standard for machine translation, but one challenge of such models is a significant increase in training and decoding cost compared to phrase-based systems. In this work we focus on efficient decoding, with a goal of achieving accuracy close the state-of-the-art in neural machine translation (NMT), while achieving CPU decoding speed/throughput close to that of a phrasal decoder. We approach this problem from two angles: First, we describe several techniques for speeding up an NMT beam search decoder, which obtain a 4.4x speedup over a very efficient baseline decoder without changing the decoder output. Second, we propose a simple but powerful network architecture which uses an RNN (GRU/LSTM) layer at bottom, followed by a series of stacked fully-connected layers applied at every timestep. This architecture achieves similar accuracy to a deep recurrent model, at a small fraction of the training and decoding cost. By combining these techniques, our best system achieves a very competitive accuracy of 38.3 BLEU on WMT English-French NewsTest2014, while decoding at 100 words/sec on single-threaded CPU. We believe this is the best published accuracy/speed trade-off of an NMT system. | null | null | 10.18653/v1/D17-1300 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,788 |
inproceedings | wang-etal-2017-exploiting-cross | Exploiting Cross-Sentence Context for Neural Machine Translation | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1301/ | Wang, Longyue and Tu, Zhaopeng and Way, Andy and Liu, Qun | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2826--2831 | In translation, considering the document as a whole can help to resolve ambiguities and inconsistencies. In this paper, we propose a cross-sentence context-aware approach and investigate the influence of historical contextual information on the performance of neural machine translation (NMT). First, this history is summarized in a hierarchical way. We then integrate the historical representation into NMT in two strategies: 1) a warm-start of encoder and decoder states, and 2) an auxiliary context source for updating decoder states. Experimental results on a large Chinese-English translation task show that our approach significantly improves upon a strong attention-based NMT system by up to +2.1 BLEU points. | null | null | 10.18653/v1/D17-1301 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,789 |
inproceedings | kim-etal-2017-cross | Cross-Lingual Transfer Learning for {POS} Tagging without Cross-Lingual Resources | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1302/ | Kim, Joo-Kyung and Kim, Young-Bum and Sarikaya, Ruhi and Fosler-Lussier, Eric | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2832--2838 | Training a POS tagging model with crosslingual transfer learning usually requires linguistic knowledge and resources about the relation between the source language and the target language. In this paper, we introduce a cross-lingual transfer learning model for POS tagging without ancillary resources such as parallel corpora. The proposed cross-lingual model utilizes a common BLSTM that enables knowledge transfer from other languages, and private BLSTMs for language-specific representations. The cross-lingual model is trained with language-adversarial training and bidirectional language modeling as auxiliary objectives to better represent language-general information while not losing the information about a specific target language. Evaluating on POS datasets from 14 languages in the Universal Dependencies corpus, we show that the proposed transfer learning model improves the POS tagging performance of the target languages without exploiting any linguistic knowledge between the source language and the target language. | null | null | 10.18653/v1/D17-1302 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,790 |
inproceedings | gella-etal-2017-image | Image Pivoting for Learning Multilingual Multimodal Representations | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1303/ | Gella, Spandana and Sennrich, Rico and Keller, Frank and Lapata, Mirella | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2839--2845 | In this paper we propose a model to learn multimodal multilingual representations for matching images and sentences in different languages, with the aim of advancing multilingual versions of image search and image understanding. Our model learns a common representation for images and their descriptions in two different languages (which need not be parallel) by considering the image as a pivot between two languages. We introduce a new pairwise ranking loss function which can handle both symmetric and asymmetric similarity between the two modalities. We evaluate our models on image-description ranking for German and English, and on semantic textual similarity of image descriptions in English. In both cases we achieve state-of-the-art performance. | null | null | 10.18653/v1/D17-1303 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,791 |
inproceedings | chen-etal-2017-neural | Neural Machine Translation with Source Dependency Representation | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1304/ | Chen, Kehai and Wang, Rui and Utiyama, Masao and Liu, Lemao and Tamura, Akihiro and Sumita, Eiichiro and Zhao, Tiejun | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2846--2852 | Source dependency information has been successfully introduced into statistical machine translation. However, there are only a few preliminary attempts for Neural Machine Translation (NMT), such as concatenating representations of source word and its dependency label together. In this paper, we propose a novel NMT with source dependency representation to improve translation performance of NMT, especially long sentences. Empirical results on NIST Chinese-to-English translation task show that our method achieves 1.6 BLEU improvements on average over a strong NMT system. | null | null | 10.18653/v1/D17-1304 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,792 |
inproceedings | han-etal-2017-visual | Visual Denotations for Recognizing Textual Entailment | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1305/ | Han, Dan and Mart{\'i}nez-G{\'o}mez, Pascual and Mineshima, Koji | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2853--2859 | In the logic approach to Recognizing Textual Entailment, identifying phrase-to-phrase semantic relations is still an unsolved problem. Resources such as the Paraphrase Database offer limited coverage despite their large size whereas unsupervised distributional models of meaning often fail to recognize phrasal entailments. We propose to map phrases to their visual denotations and compare their meaning in terms of their images. We show that our approach is effective in the task of Recognizing Textual Entailment when combined with specific linguistic and logic features. | null | null | 10.18653/v1/D17-1305 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,793 |
inproceedings | mathur-etal-2017-sequence | Sequence Effects in Crowdsourced Annotations | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1306/ | Mathur, Nitika and Baldwin, Timothy and Cohn, Trevor | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2860--2865 | Manual data annotation is a vital component of NLP research. When designing annotation tasks, properties of the annotation interface can unintentionally lead to artefacts in the resulting dataset, biasing the evaluation. In this paper, we explore sequence effects where annotations of an item are affected by the preceding items. Having assigned one label to an instance, the annotator may be less (or more) likely to assign the same label to the next. During rating tasks, seeing a low quality item may affect the score given to the next item either positively or negatively. We see clear evidence of both types of effects using auto-correlation studies over three different crowdsourced datasets. We then recommend a simple way to minimise sequence effects. | null | null | 10.18653/v1/D17-1306 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,794 |
inproceedings | ture-jojic-2017-need | No Need to Pay Attention: Simple Recurrent Neural Networks Work! | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1307/ | Ture, Ferhan and Jojic, Oliver | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2866--2872 | First-order factoid question answering assumes that the question can be answered by a single fact in a knowledge base (KB). While this does not seem like a challenging task, many recent attempts that apply either complex linguistic reasoning or deep neural networks achieve 65{\%}{--}76{\%} accuracy on benchmark sets. Our approach formulates the task as two machine learning problems: detecting the entities in the question, and classifying the question as one of the relation types in the KB. We train a recurrent neural network to solve each problem. On the SimpleQuestions dataset, our approach yields substantial improvements over previously published results {---} even neural networks based on much more complex architectures. The simplicity of our approach also has practical advantages, such as efficiency and modularity, that are valuable especially in an industry setting. In fact, we present a preliminary analysis of the performance of our model on real queries from Comcast`s X1 entertainment platform with millions of users every day. | null | null | 10.18653/v1/D17-1307 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,795 |
inproceedings | mimno-thompson-2017-strange | The strange geometry of skip-gram with negative sampling | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1308/ | Mimno, David and Thompson, Laure | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2873--2878 | Despite their ubiquity, word embeddings trained with skip-gram negative sampling (SGNS) remain poorly understood. We find that vector positions are not simply determined by semantic similarity, but rather occupy a narrow cone, diametrically opposed to the context vectors. We show that this geometric concentration depends on the ratio of positive to negative examples, and that it is neither theoretically nor empirically inherent in related embedding algorithms. | null | null | 10.18653/v1/D17-1308 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,796 |
inproceedings | botha-etal-2017-natural | Natural Language Processing with Small Feed-Forward Networks | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1309/ | Botha, Jan A. and Pitler, Emily and Ma, Ji and Bakalov, Anton and Salcianu, Alex and Weiss, David and McDonald, Ryan and Petrov, Slav | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2879--2885 | We show that small and shallow feed-forward neural networks can achieve near state-of-the-art results on a range of unstructured and structured language processing tasks while being considerably cheaper in memory and computational requirements than deep recurrent models. Motivated by resource-constrained environments like mobile phones, we showcase simple techniques for obtaining such small neural network models, and investigate different tradeoffs when deciding how to allocate a small memory budget. | null | null | 10.18653/v1/D17-1309 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,797 |
inproceedings | li-lam-2017-deep | Deep Multi-Task Learning for Aspect Term Extraction with Memory Interaction | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1310/ | Li, Xin and Lam, Wai | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2886--2892 | We propose a novel LSTM-based deep multi-task learning framework for aspect term extraction from user review sentences. Two LSTMs equipped with extended memories and neural memory operations are designed for jointly handling the extraction tasks of aspects and opinions via memory interactions. Sentimental sentence constraint is also added for more accurate prediction via another LSTM. Experiment results over two benchmark datasets demonstrate the effectiveness of our framework. | null | null | 10.18653/v1/D17-1310 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,798 |
inproceedings | andreas-klein-2017-analogs | Analogs of Linguistic Structure in Deep Representations | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1311/ | Andreas, Jacob and Klein, Dan | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2893--2897 | We investigate the compositional structure of message vectors computed by a deep network trained on a communication game. By comparing truth-conditional representations of encoder-produced message vectors to human-produced referring expressions, we are able to identify aligned (vector, utterance) pairs with the same meaning. We then search for structured relationships among these aligned pairs to discover simple vector space transformations corresponding to negation, conjunction, and disjunction. Our results suggest that neural representations are capable of spontaneously developing a {\textquotedblleft}syntax{\textquotedblright} with functional analogues to qualitative properties of natural language. | null | null | 10.18653/v1/D17-1311 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,799 |
inproceedings | yang-etal-2017-simple | A Simple Regularization-based Algorithm for Learning Cross-Domain Word Embeddings | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1312/ | Yang, Wei and Lu, Wei and Zheng, Vincent | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2898--2904 | Learning word embeddings has received a significant amount of attention recently. Often, word embeddings are learned in an unsupervised manner from a large collection of text. The genre of the text typically plays an important role in the effectiveness of the resulting embeddings. How to effectively train word embedding models using data from different domains remains a problem that is less explored. In this paper, we present a simple yet effective method for learning word embeddings based on text from different domains. We demonstrate the effectiveness of our approach through extensive experiments on various down-stream NLP tasks. | null | null | 10.18653/v1/D17-1312 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,800 |
inproceedings | noriega-atala-etal-2017-learning | Learning what to read: Focused machine reading | Palmer, Martha and Hwa, Rebecca and Riedel, Sebastian | sep | 2017 | Copenhagen, Denmark | Association for Computational Linguistics | https://aclanthology.org/D17-1313/ | Noriega-Atala, Enrique and Valenzuela-Esc{\'a}rcega, Marco A. and Morrison, Clayton and Surdeanu, Mihai | Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing | 2905--2910 | Recent efforts in bioinformatics have achieved tremendous progress in the machine reading of biomedical literature, and the assembly of the extracted biochemical interactions into large-scale models such as protein signaling pathways. However, batch machine reading of literature at today`s scale (PubMed alone indexes over 1 million papers per year) is unfeasible due to both cost and processing overhead. In this work, we introduce a focused reading approach to guide the machine reading of biomedical literature towards what literature should be read to answer a biomedical query as efficiently as possible. We introduce a family of algorithms for focused reading, including an intuitive, strong baseline, and a second approach which uses a reinforcement learning (RL) framework that learns when to explore (widen the search) or exploit (narrow it). We demonstrate that the RL approach is capable of answering more queries than the baseline, while being more efficient, i.e., reading fewer documents. | null | null | 10.18653/v1/D17-1313 | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | null | 57,801 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.