id
stringlengths
7
12
sentence1
stringlengths
6
1.27k
sentence2
stringlengths
6
926
label
stringclasses
4 values
train_17200
Data-independent methods employ random projections to construct hash functions without any consideration on data characteristics, like the locality sensitive hashing (LSH) algorithm (Datar et al., 2004).
data dependent hashing seeks to learn a hash function from the given training data in a supervised or an unsupervised way.
contrasting
train_17201
To increase the modeling ability of (1), we may resort to more complex likelihood p θ (D|z), such as using deep neural networks to relate the latent z to the observation x i , instead of the simple softmax function in (2).
as indicated in , employing expressive nonlinear decoders likely destroy the distance-keeping property, which is essential to yield good hashing codes.
contrasting
train_17202
With reductions like the above one, any ANNS methods can be applied for MIPS.
it was shown that there are performance limitations for the reduction MIPS methods (Morozov and Babenko, 2018).
contrasting
train_17203
We found that the edge selection method is vital for the trade-off of effectiveness and efficiency in searching.
the existing edge selection techniques used in HNSW and ip-NSW are actually designed for metric distances, which are inapplicable for the non-metric measure, e.g., inner product.
contrasting
train_17204
On the other hand, the information captured by word embeddings can be seamlessly used in downstream tasks, which makes embedding a potential solution for the aforementioned problem.
conventional word embeddings using one unified embedding for each word are not able to distinguish different relations types (such as various syntactic relations, which is crucial for SP) among words.
contrasting
train_17205
To address this problem, the dependency-based embedding model (Levy and Goldberg, 2014) is proposed to treat a word through separate ones, e.g., 'food@dobj' and 'food@nsubj', under different syntactic relations, with the skip-gram (Mikolov et al., 2013) model being used to train the final embeddings.
this method is limited in two aspects.
contrasting
train_17206
In particular, quantization (Lin et al., 2016;Hubara et al., 2016) and groupwise low rank approximation (Chen et al., 2018a) achieve the state-of-the-art performance on tasks like language modeling and machine translation.
these methods are rather computationally inspired than being motivated by intuitions of the semantic composition and thus suffer from poor interpretability.
contrasting
train_17207
Notice that some previous methods compress model directly during training phase (Khrulkov et al., 2019;Wen et al., 2017).
our problem setup follows (Chen et al., 2018a;Shu and Nakayama, 2017;Chen et al., 2018b) that given a pre-trained model, we want to compress the model with limited fine-tuning.
contrasting
train_17208
This implies that local precision lost for low-rank basis in GroupReduce is more difficult to be recovered.
the collective information of MulCode due to the compositional property is more robust when imprecise local vectors present.
contrasting
train_17209
Recent studies have shown that word embeddings exhibit gender bias inherited from the training corpora.
most studies to date have focused on quantifying and mitigating such bias only in English.
contrasting
train_17210
This is due to morphological agreement and should not be considered as a stereotype.
gender bias in the embeddings of languages with grammatical gender indeed exists.
contrasting
train_17211
In gendered languages, all nouns are assigned a gender class.
inanimate objects (e.g., water and spoon) do not carry the meaning of male or female.
contrasting
train_17212
English) only has the former direction.
when analyzing languages like Spanish and French, considering the second is necessary.
contrasting
train_17213
To simplify the discussion, we focus on noun class systems where two major gender classes are feminine and masculine.
the proposed approach can be generalized to languages with multiple gender classes (e.g., German).
contrasting
train_17214
A few recent studies focus on measuring and reducing gender bias in contextualized word embeddings (Zhao et al., 2019;May et al., 2019;Basta et al., 2019).
they only focus on English embeddings in which the gender is mostly only expressed by pronouns (Stahlberg et al., 2007).
contrasting
train_17215
In English, these paths are extracted from sentences where x and y co-occur.
when x and y are in different languages, a new path definition is required.
contrasting
train_17216
Supervised English system Without crosslingual training samples, we cannot compare weakly supervised and fully supervised training for BILEXNET in a controlled fashion.
the supervised monolingual ENLEXNET model (Section 6) evaluated on the En-En test set offers a reference point: remarkably the F1 scores of BILEXNET are only 1 to 3 points lower than those obtained by the supervised English model (∼44 on the En-En test set).
contrasting
train_17217
This is similar to the the Co- gALex shared task , where the first part of the task is to eliminate completely unrelated pairs, before predicting relations on the remaining pairs.
filtering out unrelated pairs is an easier task than filtering pairs in the Other category.
contrasting
train_17218
For example, given the monolingual example (drop, fall, Forward Entail) the model places the highest weight on the Hindi word , which captures the "moving downward" sense.
for the example (autumn, fall, Equivalence), the model correctly identifies Ú as the right translation.
contrasting
train_17219
Neural WSD systems (Kågebäck and Salomonsson, 2016;Raganato et al., 2017b) feed the continuous word representations into a neural network that captures the whole sentence and the word representation in the sentence.
in both approaches, the word representations are independent of the context.
contrasting
train_17220
(2017b) proposed a self-attention layer on top of the concatenated bidirectional LSTM hidden states for WSD and introduced multi-task learning with part-ofspeech tagging and semantic labeling as auxiliary tasks.
on average across the test sets, their approach did not outperform SVM with word embedding features.
contrasting
train_17221
This is surprising as the embedding directly provides a number's value, thus, the synthetic tasks should be easy to solve.
we had difficulty training models for large ranges, even when using numerous architecture variants (e.g., tiny networks with 10 hidden units and tanh activations) and hyperparameters.
contrasting
train_17222
Denoting the predicted restated query byz, simplifying E (x,y,z)∼D as E, the goal of the RL training is to maximize following objective: where Z is the space of all restated query candidates and r represents the reward defined by comparingz and the annotation z.
the overall candidate space Q×Z is vast, making it impossible to exactly maximize L rl .
contrasting
train_17223
Guided by the golden restated query z, in training, we find outz * by computing the reward of each candidate.
in inference, where there is no golden restate query, we can only obtainz * from F. Specially, for the v-th span in the followup query, we find u * = arg max u F u,v .
contrasting
train_17224
(2016) requires annotations on which phrases should be mapped to unknowns during the training phase.
such supervised knowledge is not required for our method.
contrasting
train_17225
As the interaction proceeds, the user question becomes more complicated as it requires longer SQL query to answer.
more query tokens overlap with the previous query, and thus the number of new tokens remains small at the third turn and beyond.
contrasting
train_17226
A series of pruning methods are then proposed to alleviate the imbalanced distribution, such as the korder pruning .
it does not extend well to other languages, and even hinders the syntax-agnostic SRL model as has experimented with different k values on English.
contrasting
train_17227
Additionally, the use of CNN layers allows us to explicitly control the window size for phrase modeling, which has been shown to be critical for relevance matching (Dai et al., 2018;Rao et al., 2019).
the contextual encoder enables us to obtain long-distance contextual representations for each token.
contrasting
train_17228
Indeed, semantic matching methods alone are ineffective when queries are comprised of only a few keywords, without much semantic information to exploit.
the context-aware representations learned from SM do contribute to RM, leading to the superior results of our complete HCAN model.
contrasting
train_17229
As is well known, the hard parameter sharing approach can provide representations for all the shared tasks and reduce the probability of overfitting on the main task.
this kind of sharing strategy somewhat weakens the representation framework maintains distinct model parameters for each task, due to the neutralization of knowledge introduced by the auxiliary task.
contrasting
train_17230
7 Note that these sentential and phrasal paraphrases are obtained by automatic methods.
dataset creation for downstream tasks generally requires expensive human annotation.
contrasting
train_17231
A recent task by on the CONCODE dataset maps a single utterance to an entire method, conditioned on environment variables and methods.
we tackle the task of general purpose code generation in an interactive setting, using an entire sequence of prior NL and code blocks as context.
contrasting
train_17232
1 Natural language interfaces that allow users to query data and invoke services without programming have been identified as a key application of semantic parsing (Berant et al., 2013;Thomason et al., 2015;Dong and Lapata, 2016;Zhong et al., 2017;Campagna et al., 2017;Su et al., 2017).
existing semantic parsing technologies often fall short when deployed in practice, facing several challenges: (1) user utterances can be inherently ambiguous or vague, making it difficult to get the correct result in one shot, (2) the accuracy of state-of-the-art semantic parsers are still not high enough for real use, and (3) it is hard for users to validate the semantic parsing results, especially with mainstream neural network models that are known for the lack of interpretability.
contrasting
train_17233
Similarly, (Yao et al., 2019) relies on a pre-defined two-level hierarchy among components in an If-Then program and cannot generalize to formal languages with a deeper structure.
mISP aims for a general design principle by explicitly identifying and decoupling important components, such as error detector, question generator and world model.
contrasting
train_17234
For SQLNet, we also compare our system with the reported performance of DialSQL (Gur et al., 2018, Table 4).
since DialSQL is not open-sourced and it is not easy to reproduce it, we are unable to adapt it to SQLova for more comparisons.
contrasting
train_17235
We first observe that, via interactions with simulated users, MISP-SQL improves SyntaxSQL-Net by 10% accuracy with reasonably 3 questions per query.
we also realize that, unlike on WikiSQL, in this setting, the probabilitybased error detector requires more questions than the Bayesian uncertainty-based detector.
contrasting
train_17236
In all settings, MISP-SQL improves the base parser's performance, demonstrating the benefit of involving human interaction.
we also notice that the gain is not as large as in simulation, especially on SQLova.
contrasting
train_17237
Most human evaluation studies for (interactive) semantic parsers so far (Chaurasia and Mooney, 2017;Gur et al., 2018;Yao et al., 2019) use pre-existing test questions (e.g., from datasets like WikiSQL).
this introduces an undesired discrepancy, that is, human evaluators may not necessarily be able to understand the true intent of the given questions in an faithful way, especially when the question is ambiguous, vague, or containing unfamiliar entities.
contrasting
train_17238
In Example (1), though our baseline recovers a propositional phrase for the noun staff and another one for the noun funding, it fails to recognize the anaphora and antecedent relation between the two propositional phrases.
our approach successfully recognizes :prep-for c as a reentrancy node and generates one propositional phrase shared by both nouns staff and funding.
contrasting
train_17239
(2019) utilizes additional in-domain datasets to post-train BERT's weights and then fine-tune it on this task.
such a method requires a large corpus for post-training and the fine-tuning also takes a lot of computation resources and time.
contrasting
train_17240
The main challenge comes from multi-aspect sentences, which express multiple sentiment polarities towards different targets, resulting in overlapped feature representation.
most existing neural models tend to utilize static pooling operation or attention mechanism to identify sentimental words, which therefore insufficient for dealing with overlapped features.
contrasting
train_17241
Traditional approaches have designed rich features about content and syntactic structures to capture the sentiment polarity (Jiang et al., 2011;Pérez-Rosas et al., 2012).
these featurebased methods are labor-intensive and the performance highly depends on the quality of the features.
contrasting
train_17242
Very recently, CNN-based models have shown the strengths in efficiency to tackle the aspect-level sentiment classification (Xue and Li, 2018;Huang and Carley, 2018;.
all the previous methods utilize static pooling operation or attention mechanism to locate the sentimental words, which fails to handle the overlapped features.
contrasting
train_17243
Different from previous researches, we explore the dependence among relevant posts via the authors' backgrounds, since the authors with similar backgrounds, e.g., gender, location, tend to express similar emotions.
such personal attributes are not easy to obtain in most social media websites, and it is hard to capture attributesaware words to connect similar people.
contrasting
train_17244
On one hand, most websites may not contain useful personal information.
people are normally not willing to attach their personal information in social media.
contrasting
train_17245
Recently, with the development of artificial intelligence, neural network models have been successfully applied to various NLP tasks (Collobert et al., 2011;Goldberg, 2016).
few works use neural network models for emotion detection.
contrasting
train_17246
Most of previous studies consider each post individually in emotion detection, one of the most important tasks in sentiment analysis.
since the posts in social media are generated by users, it is natural that these posts can be connected through authors' personal background attributes.
contrasting
train_17247
For instance, all baselines miss proper noun "wendys" in the first example.
our approach increases the specificity of the output, leading to better content preservation.
contrasting
train_17248
With the development of deep learning, neural networks have obtained state-ofthe-art results on many sentiment classification datasets (Kim, 2014;Dong et al., 2014;Tang et al., 2015).
despite the promising results, recent work has shown that these models easily fail in adversarial examples 2 with little perturba- * Equal Contribution.
contrasting
train_17249
Second, we regard the examples that confuse the classifier as good attacking examples and give them high reward to train the generator.
not all confusing examples are useful for robustness improvements.
contrasting
train_17250
We posit that structural and semantic correspondence is both prevalent in opinionated text, especially when associated with attributes, and crucial in accurately revealing its latent aspect and sentiment structure.
it is not recognized by existing approaches.
contrasting
train_17251
Further, Solo and Friend contain reviews of business trips, although the authors did not select Business as the trip type.
this situation does not happen for Couple and Family.
contrasting
train_17252
In practice such models are notoriously hard to train and require the availability of very large datasets.
the injection of finegrained polarity information has been shown to be a key ingredient to build competitive sentiment predictors by Socher et al.
contrasting
train_17253
While these techniques are effective, they are not ideal in domains with limited data.
work such as ELMo (Peters et al., 2018) and BERT (Devlin et al., 2018) propose deeply connected layers to learn sentence embeddings by exploiting bi-directional contexts.
contrasting
train_17254
In their work (K Sarma et al., 2018), the authors demonstrate the effectiveness of KCCA based embeddings via classification experiments.
in order to verify that the word level adaptation performed by the KCCA step can capture relevant domain semantics, we perform the following experiment.
contrasting
train_17255
These words are used extensively when writing reviews about books.
words that shift the least are Nouns (69%) such as 'Higgins', 'Gardner' and 'Schaffer' that correspond to author and character names.
contrasting
train_17256
their inner working mechanisms and suffer from the lack of interpretability.
clearly understanding where and how such a model makes such a decision is rather important for developing real-world applications Marcus, 2018).
contrasting
train_17257
One possible way to alleviate this problem is to also leverage the soft-attention mechanism as proposed in .
this soft-attention mechanism may induce additional noise and lack interpretability because it tends to assign higher weights to some domain-specific words rather than real sentiment-relevant words (Mudinas et al., 2012;Zou et al., 2018).
contrasting
train_17258
(6)) to guide both clause and word selection.
when model training is finished, i.e., both high-level and low-level policy finish all their selections, the goal of sentiment rating predictor is to perform DASC.
contrasting
train_17259
This means that, interestingly, CHIM-based attribute representations have also learned information about the category of the product.
representations learned from the bias-attention method are not able to transfer well on this task, leading to worse results compared to the random and majority baseline.
contrasting
train_17260
The visualization results are shown in Fig 3, we can observe that the model without sentiment regularizer most focus on non-sentiment words such as XiaoMei, she and heard it which are inessential to provoke the emotion.
when we plus the SR. into model, we can see an obvious weights shift on attention distribution.
contrasting
train_17261
(2019) proposed a method based on learning to re-rank candidate emotion cause clauses with extracting a number of emotion-dependent and emotion-independent features.
these methods are heavily dependent on the expensive human-based features and are too difficult in a real-world application.
contrasting
train_17262
In addition, in UKPRank no individual labeling is provided, and individual quality scores are inferred from pairs labeling.
for IBMRank each argument is individually labeled for quality, and we explicitly demonstrate the consistency of these individual labeling with the provided pairwise labeling.
contrasting
train_17263
Table 6 shows that joint learning (BERT-Joint) hurts the performance compared to single-task BERT.
using additional information from the sentence-level for the token-level classification (BERT-Granularity) yields small improvements.
contrasting
train_17264
Thus, the weights in the encoder part will adjust according to the desired task-specific label as well.
in contrast, the decoder part does not have such information.
contrasting
train_17265
These approaches provided strong baselines for text classification (Wang and Manning, 2012).
sequential patterns and semantic structure between words play a crucial role in deciding the category of a text.
contrasting
train_17266
For example, a word like "killer" in the original embedding space is very close to negative sentiment words like, "bad" and "awful", especially if the word embeddings were produced on huge fact based datasets like Wikipedia or news datasets.
in the SST2 dataset, "killer" is often used to describe a movie very positively.
contrasting
train_17267
Incorporating the parent representation only along with the claim representation does not give significant improvement over representing the claim only.
incorporating the flat representation of the larger context along with the claim representation consistently achieves significantly better (p < 0.001) performance than the claim representation alone.
contrasting
train_17268
In such an operation, the GCN only considers the first-order neighborhood of a node when modeling its embeddings.
k successive GCN operations result in the propagation of information across the k-th order neighborhood.
contrasting
train_17269
(2018 CTC Table 1 shows the main BLEU results of different methods on the test set.
we cannot identify the best DA method because their rankings across the four translation tasks vary a bit.
contrasting
train_17270
Some of those measures have complex definition such as linear regions, others are very expensive to compute for models as large as Transformer such as Hessian and stiffness.
we compute weight norm with different forms proposed in Neyshabur et al.
contrasting
train_17271
To generate y given x with the channel model, we wish to compute arg max y log p(x|y) + log p(y).
naïve decoding in this way is computationally expensive because the channel model p(x|y) is conditional on each candidate target prefix.
contrasting
train_17272
For the direct model, it is sufficient to perform a single forward pass over the network parameterizing p(y|x) to obtain output word probabilities for the entire vocabulary.
the channel model requires separate forward passes for each vocabulary word.
contrasting
train_17273
These models enable zero-shot transfer, but achieve lower results than monolingual models.
we focus on making the training of monolingual language models more efficient in a multi-lingual context.
contrasting
train_17274
All of these studies, however, only create small datasets, which are inadequate for pretraining language models.
we are among the first to report the 3 Our method We propose Multi-lingual Fine-tuning (MultiFit).
contrasting
train_17275
Previous works mainly focus on adding different components into the NART model to improve the expressiveness of the network structure to overcome the loss of autoregressive dependency (Gu et al., 2017;Lee et al., 2018;.
the computational overhead of new components will hurt the inference speed, contradicting with the goal of the NART models: to parallelize and speed up neural machine translation models.
contrasting
train_17276
As we explain in §3, our investigation requires a definition of lexical semantics that is independent of grammatical gender.
in many gendered languages, word embeddings effectively encode grammatical gender because this information is trivially recoverable from distributional semantics.
contrasting
train_17277
Although our results provide evidence for the non-arbitrariness of noun-gender assignments, they must be contextualized.
to animate nouns, it is not clear that a single cross-linguistic category explains our results.
contrasting
train_17278
Similarly, we focus on perturbing over person names.
our method is readily extendable to other kinds of models as well as to other entity types.
contrasting
train_17279
Furthermore, we were unable to extract author gender in the professor dataset since the RMP reviews are anonymous.
in future work, we may explore the influence of author gender in the celebrity dataset.
contrasting
train_17280
For example, in Figure 1, it needs at least 4 hops (i.e., "fired"-"evidence"-"blood"-"soldiers") to figure out that the word "fired" means "shot" instead of "dismissed".
the above dependency tree based methods explicitly use only first-order syntactic relations, although they may also implicitly capture high-order syntactic relations by stacking more GCN layers.
contrasting
train_17281
Therefore, the trigger words and their related entities are unable to interact in multi-order graph attention network, which leads to a lower recall.
mOGANED still achieves the best performance in terms of precision, recall, F 1 -measure among all dependency based methods, which suggests the effective of multi-order representations.
contrasting
train_17282
Depending on the task at hand, IE often achieves high correctness (sometimes above 90%).
evaluating its coverage is inherently hard, as this would require exhaustively annotated corpora as gold standard.
contrasting
train_17283
(2018) fed external world knowledge (ConceptNet relations and coreferences) explicitly into MAGE-GRU (Dhingra et al., 2017) and achieved improvements compared to only using the relational arguments.
we here show that it works even better when we learn this knowledge implicit through next sentence prediction task.
contrasting
train_17284
The fine-tuning is then done as in ProposedRU.
to the above single-step pre-training, we also attempted to start from the pre-trained model for ProposedRU web , which was pre-trained using general web text, and then additionally pretrain it with causality-rich texts to make the model suitable for causality-event recognition 4 .
contrasting
train_17285
As the neural phrase structure parser is trained on increasingly larger corpora, the regression model of EEG amplitudes fits better and better.
this pattern only obtains with in-domain training data that is lexically-similar to the first chapter of Alice in Wonderland.
contrasting
train_17286
In contrast, the development of distributional modelling techniques and the availability of vast text corpora have allowed researchers to construct effective vector space models of word meaning over large lexicons.
this comes at the cost of interpretable, human-like information about word meaning.
contrasting
train_17287
A criticism of these real-valued embedding vectors is the opaqueness of their representational dimensions and their lack of cognitive plausibility and interpretability (Murphy et al., 2012;Ş enel et al., 2018).
human conceptual property knowledge is often modelled in terms of relatively sparse and interpretable vectors, based on verbalizable, human-elicited features collected in property knowledge surveys (McRae et al., 2005;.
contrasting
train_17288
Geographical SQA has posed great challenges to NLP and related research, ranging from scenario understanding to cross-modal knowledge integration and reasoning.
there is a lack of large datasets and benchmarking efforts for this task.
contrasting
train_17289
Existing SQA datasets for other domains include the TREC Precision Medicine track (Roberts et al., 2018) for the medical domain, and CAIL for the legal domain.
sQA in the geography domain requires different forms of knowledge and different reasoning capabilities, and has posed different research challenges.
contrasting
train_17290
We chose this approach since synthetic data generation can be especially useful in novel and early-stage research efforts, as it provides a controlled environment and allows for detailed analyses of a model's ability to solve a task.
while methods such as Recurrent Entity Networks have shown promise for keeping track of the state-of-the-world in our experiments, this is still in scenarios where the complexity of natural language is relatively simple.
contrasting
train_17291
Second, previous work defines passages as articles, paragraphs, or sentences.
the question of proper granularity of passages is still underexplored.
contrasting
train_17292
Almost all previous state-of-the-art QA and RC models find answers by matching pas- DrQA (Chen et al., 2017) 37.7 44.5 41.9 48.7 32.3 38.3 29.8 -R 3 (Wang et al., 2018a) 35.3 41.7 49.0 55.3 47.3 53.7 29.1 37.5 OpenQA (Lin et al., 2018) 42.2 49.3 58.8 64.5 48.7 56.3 28.7 36.6 TraCRNet (Dehghani et al., 2019) 43.2 54.0 52.9 65.1 ----HAS-QA (Pang et al., 2019) 43.2 48.9 62.7 68.7 63.6 68.9 --BERT (Large) (Nogueira et al., sages with questions, aka inter-sentence matching (Wang and Jiang, 2017;Wang et al., 2016;Seo et al., 2017;Song et al., 2017).
bERT model simply concatenates a passage with a question, and differentiates them by separating them with a delimiter token [SEP], and assigning different segment ids for them.
contrasting
train_17293
This variance is especially high for local IDF, which serves as a strong signal as consistently observed in (Blair-Goldensohn et al., 2003).
in InsuranceQA, the variance 5894 of word signals are low.
contrasting
train_17294
Previous work usually utilizes knowledge graphs such as ConceptNet as external knowledge, and extracts triples from them to enhance the initial representation of the machine comprehension context.
such method cannot capture the structural information in the knowledge graph.
contrasting
train_17295
In semantic parsing, there is extensive work on using deep neural networks for training models over manually created logical forms in a supervised learning setup (Jia and Liang, 2016;Ling et al., 2016;Dong and Lapata, 2018).
creating labeled data for this task can be expensive and time-consuming.
contrasting
train_17296
Neural techniques question answering have improved (Devlin et al., 2018) machine reading comprehension (Rajpurkar et al., 2016, MRC): computers can take a single question and extract answers from datasets like Wikipedia.
qA models struggle to generalize when questions do not look like the standalone questions systems in training data: e.g., new genres, languages, or closely-related tasks (Yogatama et al., 2019).
contrasting
train_17297
A natural question is whether reinforcement learning could learn to retain the necessary context to rewrite questions in CQA.
our dataset could be used to pretrain a question rewriter that can further be refined using reinforcement learning.
contrasting
train_17298
You can consider any question it can answer to be too easy.
please note that the AI system incorrectly answering a question does not necessarily mean that it is good.
contrasting
train_17299
On standard answer selection datasets such as TrecQA (Wang et al., 2007) or WikiQA (Yang et al., 2015), Compare-Aggregate approaches achieve very competitive performance.
they still have some limitations.
contrasting