sentence1
stringlengths 16
446
| sentence2
stringlengths 14
436
|
---|---|
for all methods , the tweets were tokenized with the cmu twitter nlp tool . | the tweets were tokenized and part-ofspeech tagged with the cmu ark twitter nlp tool and stanford corenlp . |
it was shown by nederhof et al that prefix probabilities can also be effectively computed for probabilistic tree adjoining grammars . | nederhof et al , for instance , show that prefix probabilities , and therefore surprisal , can be estimated from tree adjoining grammars . |
first , kikuchi et al proposed a new long short-term memory network to control the length of the sentence generated by an encoder-decoder model in a text summarization task . | first , kikuchi et al tried to control the length of the sentence generated by an encoder-decoder model in a text summarization task . |
with word confusion networks further improves performance . | the complexity is dominated by the word confusion network construction and parsing . |
fofe can model the word order in a sequence based on a simple ordinally-forgetting mechanism , which uses the position of each word in the sequence . | fofe can model the word order in a sequence using a simple ordinally-forgetting mechanism according to the positions of words . |
we ’ ve demonstrated that the benefits of unsupervised multilingual learning increase steadily with the number of available languages . | we found that performance improves steadily as the number of available languages increases . |
dependency parsing consists of finding the structure of a sentence as expressed by a set of directed links ( dependencies ) between words . | dependency parsing is a way of structurally analyzing a sentence from the viewpoint of modification . |
for each task , we provide separate training , development , and test datasets for english , arabic , and spanish tweets . | for each task , we provided training , development , and test datasets for english , arabic , and spanish tweets . |
a 3-gram language model was trained from the target side of the training data for chinese and arabic , using the srilm toolkit . | a 5-gram language model with kneser-ney smoothing was trained with srilm on monolingual english data . |
c . ~ = { ( subj , 0 ) , < n , 0 ) , < v , 0 ) , < comp , 0 ) , ( bar , 0 ) , and a type 1feature successor to the feature system and . . . < agr , 1 ) , < slash , 1 ) } . | we add a type 0 feature 0e ( with p ( 0e ) = { 0 } ) c. ~= { ( subj , 0 ) , < n , 0 ) , < v , 0 ) , < comp,0 ) , ( bar , 0 ) , and a type 1feature successor to the feature system and ... < agr , 1 ) , < slash , 1 ) } use this to build the set of indices . |
shared task is a new approach to time normalization based on the semantically compositional annotation of time expressions . | the parsing time normalization task is the first effort to extend time normalization to richer and more complex time expressions . |
we derive 100-dimensional word vectors using word2vec skip-gram model trained over the domain corpus . | we use the word2vec cbow model with a window size of 5 and a minimum frequency of 5 to generate 200-dimensional vectors . |
syntactic language models can become intolerantly slow to train . | in contrast , syntactic language models can be much slower to train due to rich features . |
the learning rule was adam with default tensorflow parameters . | the learning rule was adam with standard parameters . |
we embed all words and characters into low-dimensional real-value vectors which can be learned by language model . | we derive 100-dimensional word vectors using word2vec skip-gram model trained over the domain corpus . |
semantic knowledge ( e . g . word-senses ) has been defined at the ibm scientific center . | semantic knowledge is represented in a very detailed form ( word_sense pragmatics ) . |
we used the target side of the parallel corpus and the srilm toolkit to train a 5-gram language model . | we use sri language model toolkit to train a 5-gram model with modified kneser-ney smoothing on the target-side training corpus . |
to obtain this , we used mcut proposed by ding et al which is a type of spectral clustering . | to obtain this , we perform min-max cut proposed by ding et al , which is a spectral clustering method . |
part-of-speech tagging is a crucial preliminary process in many natural language processing applications . | part-of-speech tagging is a key process for various tasks such as ` information extraction , text-to-speech synthesis , word sense disambiguation and machine translation . |
information extraction ( ie ) is a main nlp aspects for analyzing scientific papers , which includes named entity recognition ( ner ) and relation extraction ( re ) . | information extraction ( ie ) is the process of finding relevant entities and their relationships within textual documents . |
all systems are evaluated using case-insensitive bleu . | we adopted the case-insensitive bleu-4 as the evaluation metric . |
automatic image captioning is a fundamental task that couples visual and linguistic learning . | automatic image captioning is a much studied topic in both the natural language processing ( nlp ) and computer vision ( cv ) areas of research . |
in particular , the recent shared tasks of conll 2008 tackled joint parsing of syntactic and semantic dependencies . | the recent conll shared tasks have been focusing on semantic dependency parsing along with the traditional syntactic dependency parsing . |
conditional random fields are undirected graphical models trained to maximize the conditional probability of the desired outputs given the corresponding inputs . | conditional random fields are discriminatively-trained undirected graphical models that find the globally optimal labeling for a given configuration of random variables . |
additionally , a back-off 2-gram model with goodturing discounting and no lexical classes was built from the same training data , using the srilm toolkit . | a 5-gram language model was created with the sri language modeling toolkit and trained using the gigaword corpus and english sentences from the parallel data . |
johnson and charniak proposed a tag-based noisy channel model , which showed great improvement over a boosting-based classifier . | johnson and charniak , 2004 ) proposed a tag-based noisy channel model for disfluency detection . |
this is also in line with what has been previously observed in that a person may express the same stance towards a target by using negative or positive language . | as previously reported in , a person may express the same stance towards a target by using negative or positive language . |
relation extraction ( re ) is a task of identifying typed relations between known entity mentions in a sentence . | relation extraction is a fundamental task in information extraction . |
semantic difference is a ternary relation between two concepts ( apple , banana ) and a discriminative attribute ( red ) that characterizes the first concept but not the other . | semantic difference is a ternary relation between two concepts ( apple , banana ) and a discriminative feature ( red ) that characterizes the first concept but not the other . |
ding and palmer propose a syntax-based translation model based on a probabilistic synchronous dependency insert grammar , a version of synchronous grammars defined on dependency trees . | ding and palmer introduce the notion of a synchronous dependency insertion grammar as a tree substitution grammar defined on dependency trees . |
sentiment classification is a very domain-specific problem ; training a classifier using the data from one domain may fail when testing against data from another . | sentiment classification is a task to predict a sentiment label , such as positive/negative , for a given text and has been applied to many domains such as movie/product reviews , customer surveys , news comments , and social media . |
bansal et al show the benefits of such modified-context embeddings in dependency parsing task . | bansal et al show that deps context is preferable to linear context on parsing task . |
they have been useful as features in many nlp tasks . | others have found them useful in parsing and other tasks . |
for example , faruqui and dyer use canonical component analysis to align the two embedding spaces . | more concretely , faruqui and dyer use canonical correlation analysis to project the word embeddings in both languages to a shared vector space . |
the log-lineal combination weights were optimized using mert . | the minimum error rate training was used to tune the feature weights . |
we train a secondorder crf model using marmot , an efficient higher-order crf implementation . | we model the sequence of morphological tags using marmot , a pruned higher-order crf . |
word alignment is the process of identifying wordto-word links between parallel sentences . | word alignment is a fundamental problem in statistical machine translation . |
sentiment analysis is a natural language processing ( nlp ) task ( cite-p-10-3-0 ) which aims at classifying documents according to the opinion expressed about a given subject ( federici and dragoni , 2016a , b ) . | sentiment analysis is a much-researched area that deals with identification of positive , negative and neutral opinions in text . |
we use pre-trained glove embeddings to represent the words . | we use pre-trained vectors from glove for word-level embeddings . |
so in most cases of irony , such features will be useful for detection . | given much of the irony in tweets is sarcasm , looking at some of these features may be useful . |
that considers a word type and its allowed pos tags as a primary element of the model . | in this work , we take a more direct approach and treat a word type and its allowed pos tags as a primary element of the model . |
we use wordsim-353 , which contains 353 english word pairs with human similarity ratings . | specifically , we used wordsim353 , a benchmark dataset , consisting of relatedness judgments for 353 word pairs . |
mccarthy instead compares two semantic profiles in wordnet that contain the concepts corresponding to the nouns from the two argument positions . | in contrast to comparing head nouns directly , mccarthy instead compares the selectional preferences for each of the two slots . |
the 50-dimensional pre-trained word embeddings are provided by glove , which are fixed during our model training . | we use the glove pre-trained word embeddings for the vectors of the content words . |
mann and yarowsky use semantic information that is extracted from documents to inform a hierarchical agglomerative clustering algorithm . | mann and yarowsky used semantic information extracted from documents referring to the target person in an hierarchical agglomerative clustering algorithm . |
twitter is a very popular micro blogging site . | twitter is a well-known social network service that allows users to post short 140 character status update which is called “ tweet ” . |
the feature weights are tuned to optimize bleu using the minimum error rate training algorithm . | the parameter weights are optimized with minimum error rate training . |
in this paper , we propose a forest-based tree sequence to string model , which is designed to integrate the strengths of the forest-based and the tree . | to integrate their strengths , in this paper , we propose a forest-based tree sequence to string translation model . |
transliteration is a subtask in ne translation , which translates nes based on the phonetic similarity . | transliteration is often defined as phonetic translation ( cite-p-21-3-2 ) . |
in this paper , we discuss methods for automatically creating models of dialog structure . | in future work , we will assess the performance of dialog structure prediction on recognized speech . |
as a statistical significance test , we used bootstrap resampling . | we used bleu as our evaluation criteria and the bootstrapping method for significance testing . |
we use the sri language model toolkit to train a 5-gram model with modified kneser-ney smoothing on the target-side training corpus . | we have used the srilm with kneser-ney smoothing for training a language model for the first stage of decoding . |
the words in the document , question and answer are represented using pre-trained word embeddings . | the word embeddings are identified using the standard glove representations . |
relation extraction is the task of detecting and classifying relationships between two entities from text . | relation extraction is a fundamental task in information extraction . |
kobayashi et al identified opinion relations by searching for useful syntactic contextual clues . | kobayashi et al adopted a supervised learning technique to search for useful syntactic patterns as contextual clues . |
neural models have shown great success on a variety of tasks , including machine translation , image caption generation , and language modeling . | various models for learning word embeddings have been proposed , including neural net language models and spectral models . |
morphological disambiguation is the process of assigning one set of morphological features to each individual word in a text . | morphological disambiguation is the task of selecting the correct morphological parse for a given word in a given context . |
case-insensitive bleu4 was used as the evaluation metric . | all systems are evaluated using case-insensitive bleu . |
evaluation results show that our model clearly outperforms a number of baseline models in terms of both clustering posts . | the results show that our model can clearly outperform the baselines in terms of three evaluation metrics . |
modified kneser-ney trigram models are trained using srilm upon the chinese portion of the training data . | gram language models are trained over the target-side of the training data , using srilm with modified kneser-ney discounting . |
there are techniques for analyzing agreement when annotations involve segment boundaries , but our focus in this article is on words . | there are techniques for analyzing agreement when annotations involve segment boundaries , but our focus in this paper is on words . |
for the tree-based system , we applied a 4-gram language model with kneserney smoothing using srilm toolkit trained on the whole monolingual corpus . | further , we apply a 4-gram language model trained with the srilm toolkit on the target side of the training corpus . |
to reduce overfitting , we apply the dropout method to regularize our model . | to mitigate overfitting , we apply the dropout method to the inputs and outputs of the network . |
we train a kn-smoothed 5-gram language model on the target side of the parallel training data with srilm . | for language model , we use a trigram language model trained with the srilm toolkit on the english side of the training corpus . |
twitter is the medium where people post real time messages to discuss on the different topics , and express their sentiments . | twitter is a rich resource for information about everyday events – people post their tweets to twitter publicly in real-time as they conduct their activities throughout the day , resulting in a significant amount of mundane information about common events . |
the neural embeddings were created using the word2vec software 3 accompanying . | those models were trained using word2vec skip-gram and cbow . |
in this paper , we investigate unsupervised learning of field segmentation models . | in this work , we have examined the task of learning field segmentation models using unsupervised learning . |
neural machine translation is currently the state-of-the art paradigm for machine translation . | neural machine translation has recently become the dominant approach to machine translation . |
we use srilm toolkit to build a 5-gram language model with modified kneser-ney smoothing . | these language models were built up to an order of 5 with kneser-ney smoothing using the srilm toolkit . |
the constituent context model for inducing constituency parses was the first unsupervised approach to surpass a right-branching baseline . | the constituent-context model is the first unsupervised constituency grammar induction system that achieves better performance than the trivial right branching baseline for english . |
in this paper , we propose a procedure to train multi-domain , recurrent neural network-based ( rnn ) language generators via multiple adaptation . | the paper presents an incremental recipe for training multi-domain language generators based on a purely data-driven , rnn-based generation model . |
we use pre-trained word2vec word vectors and vector representations by tilk et al to obtain word-level similarity information . | we also used word2vec to generate dense word vectors for all word types in our learning corpus . |
the stochastic gradient descent with back-propagation is performed using adadelta update rule . | training is done through stochastic gradient descent over shuffled mini-batches with adadelta update rule . |
in the n-coalescent , every pair of lineages merges independently with rate 1 , with parents chosen uniformly at random from the set of possible parents . | in the n-coalescent , every pair of lineages merges independently with rate 1 , with parents chosen uniformly at random from the set of possible parents at the previous time step . |
in our approach is to allow highly flexible reordering operations , in combination with a discriminative model that can condition on rich features of the source-language input . | a critical difference in our work is to allow arbitrary reorderings of the source language sentence ( as in phrase-based systems ) , through the use of flexible parsing operations . |
we measure the translation quality using a single reference bleu . | we evaluated the translation quality of the system using the bleu metric . |
we use srilm toolkit to build a 5-gram language model with modified kneser-ney smoothing . | a trigram language model with modified kneser-ney discounting and interpolation was used as produced by the srilm toolkit . |
sentence compression is a standard nlp task where the goal is to generate a shorter paraphrase of a sentence . | sentence compression is the task of compressing long , verbose sentences into short , concise ones . |
wu presents a better-constrained grammar designed to only produce tail-recursive parses . | wu proposes a bilingual segmentation grammar extending the terminal rules by including phrase pairs . |
although coreference resolution is a subproblem of natural language understanding , coreference resolution evaluation metrics have predominately been discussed in terms of abstract entities and hypothetical system errors . | coreference resolution is the process of determining whether two expressions in natural language refer to the same entity in the world . |
we trained a tri-gram hindi word language model with the srilm tool . | we used the srilm toolkit to generate the scores with no smoothing . |
sentiment analysis is a research area where does a computational analysis of people ’ s feelings or beliefs expressed in texts such as emotions , opinions , attitudes , appraisals , etc . ( cite-p-12-1-3 ) . | sentiment analysis is the task of automatically identifying the valence or polarity of a piece of text . |
we adopt two standard metrics rouge and bleu for evaluation . | for the evaluation of the results we use the bleu score . |
in this paper , we focus on designing a review generation model that is able to leverage both user and item information . | in this paper , we focus on the problem of building assistive systems that can help users to write reviews . |
the syntactic feature set is extracted after dependency parsing using the maltparser . | all data is automatically annotated with syntactic tags using maltparser . |
we used a bitext projection technique to transfer dependency-based opinion frames . | we propose a cross-lingual framework for fine-grained opinion mining using bitext projection . |
knowledge of our native language provides an initial foundation for second language learning . | our native language ( l1 ) plays an essential role in the process of lexical choice . |
semantic roles are approximated by propbank argument roles . | direction , manner , and purpose are propbank adjunctive argument labels . |
in this paper , we study the problem of sentiment analysis on product reviews . | in this paper , we propose a novel and effective approach to sentiment analysis on product reviews . |
in this paper we present an algorithmic framework which allows an automated acquisition of map-like information from the web , based on surface patterns . | in this paper we utilize a pattern-based lexical acquisition framework for the discovery of geographical information . |
circles denote events , squares denote arguments , solid arrows represent event-event relations , and dashed arrows represent event-argument relations . | the circles denote fixations , and the lines are saccades . |
semantic parsing is the task of mapping natural language sentences to a formal representation of meaning . | semantic parsing is the task of translating natural language utterances into a machine-interpretable meaning representation . |
each context consists of approximately a paragraph of surrounding text , where the word to be discriminated ( the target word ) is found approximately in the middle of the context . | 1 a context consists of all the patterns of n-grams within a certain window around the corresponding entity mention . |
we used the pre-trained google embedding to initialize the word embedding matrix . | in this baseline , we applied the word embedding trained by skipgram on wiki2014 . |
the word embeddings used in each neural network is initialized with the pre-trained glove with the dimension of 300 . | the word embeddings are initialized using the pre-trained glove , and the embedding size is 300 . |
in spite of this broad attention , the open ie task definition has been lacking . | in spite of this wide attention , open ie ’ s formal definition is lacking . |
neural network models have been exploited to learn dense feature representation for a variety of nlp tasks . | interestingly convolutional neural networks , widely used for image processing , have recently emerged as a strong class of models for nlp tasks . |
reordering is a difficult task in translating between widely different languages such as japanese and english . | reordering is a common problem observed in language pairs of distant language origins . |
we initialize our word vectors with 300-dimensional word2vec word embeddings . | our cdsm feature is based on word vectors derived using a skip-gram model . |
we define a conditional random field for this task . | our model is a first order linear chain conditional random field . |
End of preview. Expand
in Dataset Viewer.
YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/datasets-cards)
Reformatted version of the ParaSCI dataset from ParaSCI: A Large Scientific Paraphrase Dataset for Longer Paraphrase Generation. Data retrieved from dqxiu/ParaSCI.
- Downloads last month
- 118