sentence1
stringlengths 16
446
| sentence2
stringlengths 14
436
|
---|---|
in this paper , we propose a sense-topic model for wsi , which treats sense and topic as two separate latent variables to be inferred jointly .
|
we propose a sense-topic model for wsi , which treats sense and topic as two separate latent variables to be inferred jointly .
|
word space models capture the semantic similarity between two words on the basis of their distribution in a corpus .
|
co-occurrence space models represent the meaning of a word as a vector in high-dimensional space .
|
morante et al and morante and daelemans pioneered the research on negation scope finding by formulating it as a chunking problem , which classifies the words of a sentence as being inside or outside the scope of a negation signal .
|
morante et al and daelemans pioneered the research on scope learning by formulating it as a chunking problem , which classifies the words of a sentence as being inside or outside the scope of a cue .
|
trigram language models are implemented using the srilm toolkit .
|
the trigram language model is implemented in the srilm toolkit .
|
these models can be tuned using minimum error rate training .
|
each translation model is tuned using mert to maximize bleu .
|
language is the primary tool that people use for establishing , maintaining and expressing social relations .
|
language consists of much more than just content .
|
on the other hand , experiments indicate that mental representation and processing of morphologically complex words are not quite language independent .
|
on the other hand , several cross-linguistic experiments have indicated that mental representation and processing of polymorphemic words are not language independent .
|
the task of adapting a system trained on one domain ( called the source domain ) to a new domain ( called the target domain ) is called domain adaptation .
|
if these data sources differ systematically from each other , and/or from the test data , then the problem of combining these disparate data sets to create the best possible translation system is known as domain adaptation .
|
have shown that dual decomposition or lagrangian relaxation is an elegant framework for combining different types of nlp tasks .
|
it has recently been shown that different nlp models can be effectively combined using dual decomposition .
|
to the best of our knowledge , a large-scale quantitative typological analysis of lexical semantics is lacking thus far .
|
to our knowledge , this constitutes the first large-scale quantitative lexical semantic typology that is completely unsupervised , bottom-up , and data-driven .
|
relation detection is significantly more challenging compared to general relation detection tasks .
|
kb relation detection is a key step in kbqa and is significantly different from general relation extraction tasks .
|
jindal and liu base the recognition of comparative predicates on a list of manually compiled keywords .
|
jindal and liu introduce techniques to find and analyze explicit comparison sentences , but this assumes that such sentences exist .
|
one important difference between mstparser and maltparser , on the one hand , and the best performing parsers evaluated in rimell et al , on the other , is that the former were never developed specifically as parsers for english .
|
unlike the best performing grammar-based parsers studied in rimell et al , neither mstparser nor maltparser was developed specifically as a parser for english , and neither has any special mechanism for dealing with unbounded dependencies .
|
we used a phrase-based smt model as implemented in the moses toolkit .
|
we experimented with the phrase-based smt model as implemented in moses .
|
to generate the n-gram language models , we used the kenlm n-gram , language modeling tool .
|
we trained a standard 5-gram language model with modified kneser-ney smoothing using the kenlm toolkit on 4 billion running words .
|
the pre-processed monolingual sentences will be used by srilm or berkeleylm to train a n-gram language model .
|
a 5-gram language model with kneser-ney smoothing was trained with srilm on monolingual english data .
|
cite-p-17-3-2 proposed a recursive neural network designed to model the subtrees , and cnn to capture .
|
cite-p-17-1-2 proposed a simple customizaition of recursive neural networks .
|
sapkota et al showed that classical character n-grams lose some information in merging instances of ngrams like the which could be a prefix , a suffix , or a standalone word .
|
sapkota et al showed that classical character n-grams lose some information in merging together instances of n-grams like the which could be a prefix , a suffix , or a standalone word .
|
transitionbased and graph-based models have attracted the most attention of dependency parsing in recent years .
|
such approaches , for example , transition-based and graph-based models have attracted the most attention in dependency parsing in recent works .
|
this paper describes our investigation into the effectiveness of lexicalization in dependency parsing .
|
this paper discusses our investigation into the effectiveness of lexicalization in dependency parsing .
|
all the weights of those features are tuned by using minimal error rate training .
|
the weights associated to feature functions are optimally combined using the minimum error rate training .
|
we use srilm toolkit to train a trigram language model with modified kneser-ney smoothing on the target side of training corpus .
|
we use srilm to train a 5-gram language model on the target side of our training corpus with modified kneser-ney discounting .
|
for this language model , we built a trigram language model with kneser-ney smoothing using srilm from the same automatically segmented corpus .
|
for the language model we use the corpus of 60,000 simple english wikipedia articles 3 and build a 3-gram language model with kneser-ney smoothing trained with srilm .
|
in order to provide results on additional languages , we present in table 3 a comparison to the work of gillenwater et al , using the conll-x shared task data .
|
following pitler et al , we report in table 1 figures for the training sets of six languages used in the conll-x shared task on dependency parsing .
|
morphologically , arabic is a non-concatenative language .
|
arabic is a highly inflectional language with 85 % of words derived from trilateral roots ( alfedaghi and al-anzi 1989 ) .
|
language models were trained with the kenlm toolkit .
|
the 5-gram language models were built using kenlm .
|
twitter is the medium where people post real time messages to discuss on the different topics , and express their sentiments .
|
twitter is a very popular micro blogging site .
|
in this paper , we propose a feature augmentation approach for dependency parser adaptation .
|
in this paper , we present a parsing adaptation approach focused on the fully supervised case .
|
an english 5-gram language model is trained using kenlm on the gigaword corpus .
|
a 5-gram language model on the english side of the training data was trained with the kenlm toolkit .
|
time normalization is a crucial part of almost any information extraction task that needs to place entities or events along a timeline .
|
time normalization is the task of converting a natural language expression of time into a formal representation of a time on a timeline .
|
lda is a widely used topic model , which views the underlying document distribution as having a dirichlet prior .
|
lda is a probabilistic generative model that can be used to uncover the underlying semantic structure of a document collection .
|
word sense disambiguation ( wsd ) is a difficult natural language processing task which requires that for every content word ( noun , adjective , verb or adverb ) the appropriate meaning is automatically selected from the available sense inventory 1 .
|
word sense disambiguation ( wsd ) is the task to identify the intended sense of a word in a computational manner based on the context in which it appears ( cite-p-13-3-4 ) .
|
turney and littman proposed to compute pair-wised mutual information between a target word and a set of seed positive and negative words to infer the so of the target word .
|
turney and littman determined the polarity of sentiment words by estimating the point-wise mutual information between sentiment words and a set of seed words with strong polarity .
|
i like my x like i like my y , z jokes , where x , y , and z are variables to be filled in .
|
i like my x like i like my y , z , where x , y , and z are slots to be filled in .
|
the srilm toolkit was used to build this language model .
|
we used the sri language modeling toolkit for this purpose .
|
we used the sri language modeling toolkit to train a fivegram model with modified kneser-ney smoothing .
|
for the language model , we used sri language modeling toolkit to train a trigram model with modified kneser-ney smoothing on the 31 , 149 english sentences .
|
for phrase-based smt translation , we used the moses decoder and its support training scripts .
|
we used moses with the default configuration for phrase-based translation .
|
for estimating monolingual word vector models , we use the cbow algorithm as implemented in the word2vec package using a 5-token window .
|
to obtain these features , we use the word2vec implementation available in the gensim toolkit to obtain word vectors with dimension 300 for each word in the responses .
|
we use the moses software package 5 to train a pbmt model .
|
we implement the pbsmt system with the moses toolkit .
|
we used the svm-light-tk 5 to train the reranker with a combination of tree kernels and feature vectors .
|
we use svm-light-tk to train our reranking models , 9 which enables the use of tree kernels in svm-light .
|
silberer and frank use an entity-based coreference resolution model to automatically extended the training set .
|
silberer and frank cast ni resolution as a coreference resolution task , and employ an entity-mention model .
|
furthermore , we also evaluated the system on a similarly drug-focused corpus annotated for anaphora .
|
additionally , we tested our system on drugnerar corpus , which similarly focuses on drug interactions .
|
for this purpose , we use the uplug toolkit which is a collection of tools for processing corpus data , created by j枚rg tiedemann .
|
we use the uplug 5 collection of tools for alignment to extract translations from our specialized parallel corpus .
|
faruqui and dyer introduce canonical correlation analysis to project the embeddings in both languages to a shared vector space .
|
faruqui and dyer introduced canonical correlation analysis to project the embeddings in both languages to a shared vector space .
|
all smt models were developed using the moses phrase-based mt toolkit and the experiment management system .
|
both systems are phrase-based smt models , trained using the moses toolkit .
|
srilm toolkit is used to build these language models .
|
the trigram language model is implemented in the srilm toolkit .
|
relation extraction ( re ) is the task of determining semantic relations between entities mentioned in text .
|
relation extraction ( re ) is the task of extracting instances of semantic relations between entities in unstructured data such as natural language text .
|
we train trigram language models on the training set using the sri language modeling tookit .
|
we use sri language model toolkit to train a 5-gram model with modified kneser-ney smoothing on the target-side training corpus .
|
semantic parsing is the task of mapping a natural language ( nl ) sentence into a complete , formal meaning representation ( mr ) which a computer program can execute to perform some task , like answering database queries or controlling a robot .
|
semantic parsing is the problem of translating human language into computer language , and therefore is at the heart of natural language understanding .
|
system tuning was carried out using minimum error rate training optimised with k-best mira on a held out development set .
|
feature weights were set with minimum error rate training on a tuning set using bleu as the objective function .
|
word sense disambiguation ( wsd ) is a natural language processing ( nlp ) task in which the correct meaning ( sense ) of a word in a given context is to be determined .
|
word sense disambiguation ( wsd ) is a key enabling-technology that automatically chooses the intended sense of a word in context .
|
a hierarchical phrase-based translation grammar was extracted for the nist mt03 chinese-english translation using a suffix array rule extractor .
|
the hierarchical translation grammar was extracted using the joshua toolkit implementation of the suffix array rule extractor algorithm .
|
mutalik et al developed another rule based system called negfinder that recognizes negation patterns in biomedical text .
|
mutalik et al developed negfinder , a rule-based system that recognises negated patterns in medical documents .
|
in this paper , we adopt continuous bag-of-word in word2vec as our context-based embedding model .
|
with english gigaword corpus , we use the skip-gram model as implemented in word2vec 3 to induce embeddings .
|
cardie and wagstaff have proposed an unsupervised approach which also incorporates cluster information into consideration .
|
cardie and wagstaff present an early approach to unsupervised coreference resolution based on a straightforward clustering approach .
|
we use the skipgram model to learn word embeddings .
|
we use the monolingual corpora provided for the wmt translation task .
|
for our chinese-english experiments , we use a simple heuristic that equates anchors with constituents whose corresponding word class belongs to function words-related classes , bearing a close resemblance to .
|
for our chinese-english experiments , we use a simple heuristic that equates as anchors , single-word chunks whose corresponding word class belongs to closed-word classes , bearing a close resemblance to .
|
note that we interpret factuality as event factuality in the sense of saur铆 and pustejovsky .
|
following boye and saur铆 and pustejovsky , we characterize evidential justification in terms of epistemic support .
|
to test this hypothesis , we use a latent dirichlet allocation model .
|
for this feature , we use the latent dirichlet allocation .
|
these methods were normally created based on a large corpus of well-formed native english texts .
|
these methods are normally created based on a large corpus of well-formed native english texts .
|
we used moses as the implementation of the baseline smt systems .
|
we use the moses toolkit to train our phrase-based smt models .
|
system 1 is a new approach using sequence-tosequence models , encoderdecoder , and attention as described in bahdanau et al for machine translation .
|
our system for this shared task 1 is based on an encoder-decoder model proposed by bahdanau et al for neural machine translation .
|
conversational systems must be able to learn new words automatically during human machine conversation .
|
it is desirable that conversational systems can learn new words automatically during human machine conversation .
|
named entity recognition ( ner ) is the task of identifying and typing phrases that contain the names of persons , organizations , locations , and so on .
|
named entity recognition ( ner ) is a key technique for ie and other natural language processing tasks .
|
one way to handle such faulty arguments is to simply disregard them and focus on extracting arguments containing proper support .
|
one way to deal with such implicit arguments is to simply disregard them and focus on extracting arguments containing proper support .
|
ruppenhofer et al argued that semantic role techniques are useful but not completely sufficient for holder and topic identification , and that other linguistic phenomena must be studied as well .
|
ruppenhofer , somasundaran , and wiebe argued that semantic role techniques are useful but not completely sufficient for holder and topic identification , and that other linguistic phenomena must be studied as well .
|
recurrent neural networks are another way to exploit the context of a word by considering the sequence of words preceding it .
|
recurrent neural networks are another natural choice to model text due to their capability of processing arbitrary-length sequences .
|
this paper presents a thorough evaluation of the impact of annotation noise on al .
|
in this paper we present a thorough evaluation of the impact of annotation noise on al .
|
relation classification is the task of finding semantic relations between pairs of nominals , which is useful for many nlp applications , such as information extraction ( cite-p-15-3-3 ) , question answering ( cite-p-15-3-6 ) .
|
relation classification is the task of identifying the semantic relation holding between two nominal entities in text .
|
then , additional alignment points are added according to the growing heuristic algorithm , grow additional alignment points , finally , we select consecutive clusters which are aligned to the same english word as candidates .
|
then , additional alignment points are added according to the growing heuristic algorithm , grow additional alignment points , finally , we select consecutive which are aligned to the same english word as candidates .
|
liu et al and baron et al carried out sentence unit and disfluency prediction as separate tasks .
|
liu et al used conditional random fields for sentence boundary and edit word detection .
|
in cross-lingual settings , the actual translations of a word can be taken as the sense labels .
|
in a multilingual context , word senses can be easily identified using their translations in other languages .
|
we train a linear support vector machine classifier using the efficient liblinear package .
|
we build all the classifiers using the l2-regularized linear logistic regression from the liblinear package .
|
in this study , we presented a novel encoder-decoder model to automatically generate market comments from numerical time-series data of stock prices .
|
in this study , we focus on the task of generating market comments from a time-series of stock prices .
|
semantic similarity has seen major progress in recent times , due largely to the semeval semantic textual similarity ( sts ) task ( cite-p-17-1-0 , cite-p-17-1-1 , cite-p-17-1-2 , cite-p-17-1-3 ) .
|
major progress has been made in this task in recent years , due primarily to the semeval semantic textual similarity ( sts ) task ( cite-p-17-1-0 , cite-p-17-1-1 , cite-p-17-1-2 , cite-p-17-1-3 ) .
|
hamilton et al propose the use of cosine similarities of words in different contexts to detect changes .
|
hamilton et al measured the variation between models by observing semantic change using diachronic corpora .
|
unfortunately , wordnet is a fine-grained resource , which encodes possibly subtle sense distictions .
|
wordnet is a comprehensive lexical resource for word-sense disambiguation ( wsd ) , covering nouns , verbs , adjectives , adverbs , and many multi-word expressions .
|
the dataset we used in the present study is the online edition 2 of the world atlas of language structures .
|
the database of typological features we used is the online edition 8 of the world atlas of language structures .
|
experimental results demonstrate that the distributional similarity based models can significantly outperform their baseline systems .
|
experimental results show that both methods can achieve significant improvements over their baseline settings .
|
a 5-gram language model was created with the sri language modeling toolkit and trained using the gigaword corpus and english sentences from the parallel data .
|
language models were estimated using the sri language modeling toolkit with modified kneser-ney smoothing .
|
translation quality is measured in truecase with bleu and ter .
|
the translation quality is evaluated by case-insensitive bleu and ter metrics using multeval .
|
also , convolutional neural networks have been a popular choice in the image domain .
|
convolutional neural networks have rapidly become the state-of-the-art approach in computer vision .
|
we employ support vector machine as the machine learning approach .
|
in the experiments reported here we use support vector machines through the svm light package .
|
experiments on chinese-english translation show that joint training with generalized agreement achieves significant improvements over two baselines for ( hierarchical ) .
|
experiments on chineseenglish translation show that joint training with generalized agreement achieves significant improvements over two state-of-the-art alignment methods .
|
infrastructural issues are dealt with by the platform , completely transparently to the user : load balancing , efficient data upload and storage , deployment on the virtual machines , security , and fault tolerance .
|
important infrastructural issues are dealt with by the platform , completely transparently for the researcher : load balancing , efficient data upload and storage , deployment on the virtual machines , security , and fault tolerance .
|
hence we use the expectation maximization algorithm for parameter learning .
|
a minimum of this function can be found using the em algorithm .
|
for language model , we used sri language modeling toolkit to train a 4-gram model with modified kneser-ney smoothing .
|
we also use a 4-gram language model trained using srilm with kneser-ney smoothing .
|
we develop translation models using the phrase-based moses smt system .
|
our baseline system is phrase-based moses with feature weights trained using mert .
|
figure 1 : multimodal compact bilinear pooling .
|
figure 2 : multimodal compact bilinear pooling ( mcb )
|
madamira is a tool designed for morphological analysis and disambiguation of modern standard arabic .
|
madamira is a system for morphological analysis and disambiguation of arabic text .
|
goldwasser et al presented a confidence-driven approach to semantic parsing based on self-training .
|
goldwasser et al took an unsupervised approach for semantic parsing based on self-training driven by confidence estimation .
|
this paper proposes the ¡° hierarchical directed acyclic graph ( hdag ) kernel .
|
this paper proposes the hierarchical directed acyclic graph ( hdag ) kernel .
|
also , we initialized all of the word embeddings using the 300 dimensional pre-trained vectors from glove .
|
our word embeddings is initialized with 100-dimensional glove word embeddings .
|
we also use glove vectors to initialize the word embedding matrix in the caption embedding module .
|
we use pre-trained glove vector for initialization of word embeddings .
|
for instance , collobert and weston use a multitask network for different nlp tasks and show that the multi-task setting improves generality among shared tasks .
|
collobert and weston propose a unified deep convolutional neural network for different tasks by using a set of taskindependent word embeddings together with a set of task-specific word embeddings .
|
with dependency treebanks , since ccg lexical categories can be easily extracted from dependency treebanks ( cite-p-19-1-2 , cite-p-19-1-0 ) .
|
our approach can therefore be adapted to languages with dependency treebanks , since ccg lexical categories can be easily extracted from dependency treebanks ( cite-p-19-1-2 , cite-p-19-1-0 ) .
|
in this paper , we present a reranking method of n-best multi-sentence compressions based on keyphrase extraction .
|
this paper presents an n-best reranking method based on keyphrase extraction .
|
the default is the phrase-based variant of cube pruning .
|
the default phrasal search algorithm is cube pruning .
|
axelrod et al , 2011 ) extend this work using the bilingual difference of cross-entropy on in-domain and out-of-domain language models for training sentence selection for smt .
|
axelrod et al improved the perplexitybased approach and proposed bilingual crossentropy difference as a ranking function with inand general-domain language models .
|
semantic role labeling ( srl ) is the task of automatic recognition of individual predicates together with their major roles ( e.g . frame elements ) as they are grammatically realized in input sentences .
|
semantic role labeling ( srl ) is the task of identifying the semantic arguments of a predicate and labeling them with their semantic roles .
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.