Unnamed: 0
int64
0
3.22k
text
stringlengths
49
577
id
int64
0
3.22k
label
int64
0
6
2,800
[[ It ]] also gives significant speedups in << inference >> for several datasets with varying degrees of spatio-temporal continuity .
2,800
3
2,801
<< It >> also gives significant speedups in inference for several [[ datasets ]] with varying degrees of spatio-temporal continuity .
2,801
6
2,802
It also gives significant speedups in inference for several << datasets >> with varying degrees of [[ spatio-temporal continuity ]] .
2,802
1
2,803
We also discuss the strengths and weaknesses of our [[ strategy ]] relative to existing << hierarchical approaches >> , and the kinds of image and video data that provide the best speedups .
2,803
5
2,804
Motivated by the success of [[ ensemble methods ]] in << machine learning >> and other areas of natural language processing , we developed a multi-strategy and multi-source approach to question answering which is based on combining the results from different answering agents searching for answers in multiple corpora .
2,804
3
2,805
Motivated by the success of [[ ensemble methods ]] in machine learning and other areas of << natural language processing >> , we developed a multi-strategy and multi-source approach to question answering which is based on combining the results from different answering agents searching for answers in multiple corpora .
2,805
3
2,806
Motivated by the success of ensemble methods in machine learning and other areas of natural language processing , we developed a [[ multi-strategy and multi-source approach ]] to << question answering >> which is based on combining the results from different answering agents searching for answers in multiple corpora .
2,806
3
2,807
The << answering agents >> adopt fundamentally different [[ strategies ]] , one utilizing primarily knowledge-based mechanisms and the other adopting statistical techniques .
2,807
3
2,808
The answering agents adopt fundamentally different << strategies >> , [[ one ]] utilizing primarily knowledge-based mechanisms and the other adopting statistical techniques .
2,808
2
2,809
The answering agents adopt fundamentally different strategies , << one >> utilizing primarily [[ knowledge-based mechanisms ]] and the other adopting statistical techniques .
2,809
3
2,810
The answering agents adopt fundamentally different << strategies >> , one utilizing primarily knowledge-based mechanisms and the [[ other ]] adopting statistical techniques .
2,810
2
2,811
The answering agents adopt fundamentally different strategies , one utilizing primarily knowledge-based mechanisms and the << other >> adopting [[ statistical techniques ]] .
2,811
3
2,812
We present our << multi-level answer resolution algorithm >> that combines results from the [[ answering agents ]] at the question , passage , and/or answer levels .
2,812
3
2,813
Experiments evaluating the effectiveness of our [[ answer resolution algorithm ]] show a 35.0 % relative improvement over our << baseline system >> in the number of questions correctly answered , and a 32.8 % improvement according to the average precision metric .
2,813
5
2,814
Experiments evaluating the effectiveness of our << answer resolution algorithm >> show a 35.0 % relative improvement over our baseline system in the number of questions correctly answered , and a 32.8 % improvement according to the [[ average precision metric ]] .
2,814
6
2,815
Experiments evaluating the effectiveness of our answer resolution algorithm show a 35.0 % relative improvement over our << baseline system >> in the number of questions correctly answered , and a 32.8 % improvement according to the [[ average precision metric ]] .
2,815
6
2,816
[[ Word Identification ]] has been an important and active issue in << Chinese Natural Language Processing >> .
2,816
2
2,817
In this paper , a new [[ mechanism ]] , based on the concept of sublanguage , is proposed for identifying << unknown words >> , especially personal names , in Chinese newspapers .
2,817
3
2,818
In this paper , a new << mechanism >> , based on the concept of [[ sublanguage ]] , is proposed for identifying unknown words , especially personal names , in Chinese newspapers .
2,818
3
2,819
In this paper , a new mechanism , based on the concept of sublanguage , is proposed for identifying << unknown words >> , especially [[ personal names ]] , in Chinese newspapers .
2,819
2
2,820
In this paper , a new << mechanism >> , based on the concept of sublanguage , is proposed for identifying unknown words , especially personal names , in [[ Chinese newspapers ]] .
2,820
3
2,821
The proposed << mechanism >> includes [[ title-driven name recognition ]] , adaptive dynamic word formation , identification of 2-character and 3-character Chinese names without title .
2,821
4
2,822
The proposed mechanism includes [[ title-driven name recognition ]] , << adaptive dynamic word formation >> , identification of 2-character and 3-character Chinese names without title .
2,822
0
2,823
The proposed << mechanism >> includes title-driven name recognition , [[ adaptive dynamic word formation ]] , identification of 2-character and 3-character Chinese names without title .
2,823
4
2,824
The proposed mechanism includes title-driven name recognition , [[ adaptive dynamic word formation ]] , << identification of 2-character and 3-character Chinese names without title >> .
2,824
0
2,825
The proposed << mechanism >> includes title-driven name recognition , adaptive dynamic word formation , [[ identification of 2-character and 3-character Chinese names without title ]] .
2,825
4
2,826
This report describes [[ Paul ]] , a << computer text generation system >> designed to create cohesive text through the use of lexical substitutions .
2,826
2
2,827
This report describes Paul , a [[ computer text generation system ]] designed to create << cohesive text >> through the use of lexical substitutions .
2,827
3
2,828
This report describes << Paul >> , a computer text generation system designed to create cohesive text through the use of [[ lexical substitutions ]] .
2,828
3
2,829
Specifically , this system is designed to deterministically choose between [[ pronominalization ]] , << superordinate substitution >> , and definite noun phrase reiteration .
2,829
5
2,830
Specifically , this system is designed to deterministically choose between pronominalization , [[ superordinate substitution ]] , and << definite noun phrase reiteration >> .
2,830
5
2,831
The [[ system ]] identifies a strength of << antecedence recovery >> for each of the lexical substitutions .
2,831
3
2,832
The system identifies a strength of [[ antecedence recovery ]] for each of the << lexical substitutions >> .
2,832
3
2,833
It describes the automated training and evaluation of an Optimal Position Policy , a [[ method ]] of locating the likely << positions of topic-bearing sentences >> based on genre-specific regularities of discourse structure .
2,833
3
2,834
It describes the automated training and evaluation of an Optimal Position Policy , a << method >> of locating the likely positions of topic-bearing sentences based on [[ genre-specific regularities of discourse structure ]] .
2,834
3
2,835
This [[ method ]] can be used in << applications >> such as information retrieval , routing , and text summarization .
2,835
3
2,836
This method can be used in << applications >> such as [[ information retrieval ]] , routing , and text summarization .
2,836
2
2,837
This method can be used in applications such as [[ information retrieval ]] , << routing >> , and text summarization .
2,837
0
2,838
This method can be used in << applications >> such as information retrieval , [[ routing ]] , and text summarization .
2,838
2
2,839
This method can be used in applications such as information retrieval , [[ routing ]] , and << text summarization >> .
2,839
0
2,840
This method can be used in << applications >> such as information retrieval , routing , and [[ text summarization ]] .
2,840
2
2,841
We describe a general [[ framework ]] for << online multiclass learning >> based on the notion of hypothesis sharing .
2,841
3
2,842
We describe a general << framework >> for online multiclass learning based on the [[ notion of hypothesis sharing ]] .
2,842
3
2,843
We generalize the [[ multiclass Perceptron ]] to our << framework >> and derive a unifying mistake bound analysis .
2,843
3
2,844
We demonstrate the merits of our approach by comparing [[ it ]] to previous << methods >> on both synthetic and natural datasets .
2,844
5
2,845
We demonstrate the merits of our approach by comparing << it >> to previous methods on both [[ synthetic and natural datasets ]] .
2,845
6
2,846
We demonstrate the merits of our approach by comparing it to previous << methods >> on both [[ synthetic and natural datasets ]] .
2,846
6
2,847
We describe a set of [[ supervised machine learning ]] experiments centering on the construction of << statistical models of WH-questions >> .
2,847
3
2,848
These << models >> , which are built from [[ shallow linguistic features of questions ]] , are employed to predict target variables which represent a user 's informational goals .
2,848
3
2,849
We argue in favor of the the use of [[ labeled directed graph ]] to represent various types of << linguistic structures >> , and illustrate how this allows one to view NLP tasks as graph transformations .
2,849
3
2,850
We argue in favor of the the use of [[ labeled directed graph ]] to represent various types of linguistic structures , and illustrate how this allows one to view << NLP tasks >> as graph transformations .
2,850
3
2,851
We argue in favor of the the use of labeled directed graph to represent various types of linguistic structures , and illustrate how [[ this ]] allows one to view << NLP tasks >> as graph transformations .
2,851
3
2,852
We present a general [[ method ]] for learning such << transformations >> from an annotated corpus and describe experiments with two applications of the method : identification of non-local depenencies -LRB- using Penn Treebank data -RRB- and semantic role labeling -LRB- using Proposition Bank data -RRB- .
2,852
3
2,853
We present a general << method >> for learning such transformations from an [[ annotated corpus ]] and describe experiments with two applications of the method : identification of non-local depenencies -LRB- using Penn Treebank data -RRB- and semantic role labeling -LRB- using Proposition Bank data -RRB- .
2,853
3
2,854
We present a general method for learning such transformations from an annotated corpus and describe experiments with two << applications >> of the [[ method ]] : identification of non-local depenencies -LRB- using Penn Treebank data -RRB- and semantic role labeling -LRB- using Proposition Bank data -RRB- .
2,854
3
2,855
We present a general method for learning such transformations from an annotated corpus and describe experiments with two << applications >> of the method : [[ identification of non-local depenencies ]] -LRB- using Penn Treebank data -RRB- and semantic role labeling -LRB- using Proposition Bank data -RRB- .
2,855
2
2,856
We present a general method for learning such transformations from an annotated corpus and describe experiments with two applications of the method : << identification of non-local depenencies >> -LRB- using [[ Penn Treebank data ]] -RRB- and semantic role labeling -LRB- using Proposition Bank data -RRB- .
2,856
3
2,857
We present a general method for learning such transformations from an annotated corpus and describe experiments with two << applications >> of the method : identification of non-local depenencies -LRB- using Penn Treebank data -RRB- and [[ semantic role labeling ]] -LRB- using Proposition Bank data -RRB- .
2,857
2
2,858
We present a general method for learning such transformations from an annotated corpus and describe experiments with two applications of the method : identification of non-local depenencies -LRB- using Penn Treebank data -RRB- and << semantic role labeling >> -LRB- using [[ Proposition Bank data ]] -RRB- .
2,858
3
2,859
We describe a generative probabilistic model of natural language , which we call HBG , that takes advantage of detailed [[ linguistic information ]] to resolve << ambiguity >> .
2,859
3
2,860
[[ HBG ]] incorporates lexical , syntactic , semantic , and structural information from the parse tree into the << disambiguation process >> in a novel way .
2,860
3
2,861
<< HBG >> incorporates [[ lexical , syntactic , semantic , and structural information ]] from the parse tree into the disambiguation process in a novel way .
2,861
3
2,862
We use a [[ corpus of bracketed sentences ]] , called a Treebank , in combination with << decision tree building >> to tease out the relevant aspects of a parse tree that will determine the correct parse of a sentence .
2,862
0
2,863
We use a [[ corpus of bracketed sentences ]] , called a Treebank , in combination with decision tree building to tease out the relevant aspects of a << parse tree >> that will determine the correct parse of a sentence .
2,863
3
2,864
We use a corpus of bracketed sentences , called a Treebank , in combination with [[ decision tree building ]] to tease out the relevant aspects of a << parse tree >> that will determine the correct parse of a sentence .
2,864
3
2,865
We use a corpus of bracketed sentences , called a Treebank , in combination with decision tree building to tease out the relevant aspects of a [[ parse tree ]] that will determine the correct << parse >> of a sentence .
2,865
3
2,866
This stands in contrast to the usual approach of further [[ grammar tailoring ]] via the usual linguistic introspection in the hope of generating the correct << parse >> .
2,866
3
2,867
This stands in contrast to the usual approach of further << grammar tailoring >> via the usual [[ linguistic introspection ]] in the hope of generating the correct parse .
2,867
3
2,868
In head-to-head tests against one of the best existing << robust probabilistic parsing models >> , which we call [[ P-CFG ]] , the HBG model significantly outperforms P-CFG , increasing the parsing accuracy rate from 60 % to 75 % , a 37 % reduction in error .
2,868
2
2,869
In head-to-head tests against one of the best existing robust probabilistic parsing models , which we call P-CFG , the [[ HBG model ]] significantly outperforms << P-CFG >> , increasing the parsing accuracy rate from 60 % to 75 % , a 37 % reduction in error .
2,869
5
2,870
In head-to-head tests against one of the best existing robust probabilistic parsing models , which we call P-CFG , the << HBG model >> significantly outperforms P-CFG , increasing the [[ parsing accuracy rate ]] from 60 % to 75 % , a 37 % reduction in error .
2,870
6
2,871
The framework of the << analysis >> is [[ model-theoretic semantics ]] .
2,871
3
2,872
This paper addresses the issue of << word-sense ambiguity >> in extraction from [[ machine-readable resources ]] for the construction of large-scale knowledge sources .
2,872
3
2,873
This paper addresses the issue of word-sense ambiguity in extraction from [[ machine-readable resources ]] for the << construction of large-scale knowledge sources >> .
2,873
3
2,874
We describe two experiments : one which ignored word-sense distinctions , resulting in 6.3 % [[ accuracy ]] for << semantic classification >> of verbs based on -LRB- Levin , 1993 -RRB- ; and one which exploited word-sense distinctions , resulting in 97.9 % accuracy .
2,874
6
2,875
These experiments were dual purpose : -LRB- 1 -RRB- to validate the central thesis of the work of -LRB- Levin , 1993 -RRB- , i.e. , that [[ verb semantics ]] and << syntactic behavior >> are predictably related ; -LRB- 2 -RRB- to demonstrate that a 15-fold improvement can be achieved in deriving semantic information from syntactic cues if we first divide the syntactic cues into distinct groupings that correlate with different word senses .
2,875
0
2,876
These experiments were dual purpose : -LRB- 1 -RRB- to validate the central thesis of the work of -LRB- Levin , 1993 -RRB- , i.e. , that verb semantics and syntactic behavior are predictably related ; -LRB- 2 -RRB- to demonstrate that a 15-fold improvement can be achieved in deriving << semantic information >> from [[ syntactic cues ]] if we first divide the syntactic cues into distinct groupings that correlate with different word senses .
2,876
3
2,877
Finally , we show that we can provide effective acquisition [[ techniques ]] for novel << word senses >> using a combination of online sources .
2,877
3
2,878
Finally , we show that we can provide effective acquisition << techniques >> for novel word senses using a combination of [[ online sources ]] .
2,878
3
2,879
The [[ TIPSTER Architecture ]] has been designed to enable a variety of different << text applications >> to use a set of common text processing modules .
2,879
3
2,880
The TIPSTER Architecture has been designed to enable a variety of different << text applications >> to use a set of [[ common text processing modules ]] .
2,880
3
2,881
Since [[ user interfaces ]] work best when customized for particular << applications >> , it is appropriator that no particular user interface styles or conventions are described in the TIPSTER Architecture specification .
2,881
3
2,882
However , the Computing Research Laboratory -LRB- CRL -RRB- has constructed several << TIPSTER applications >> that use a common set of configurable [[ Graphical User Interface -LRB- GUI -RRB- functions ]] .
2,882
3
2,883
These << GUIs >> were constructed using [[ CRL 's TIPSTER User Interface Toolkit -LRB- TUIT -RRB- ]] .
2,883
3
2,884
[[ TUIT ]] is a << software library >> that can be used to construct multilingual TIPSTER user interfaces for a set of common user tasks .
2,884
2
2,885
[[ TUIT ]] is a software library that can be used to construct << multilingual TIPSTER user interfaces >> for a set of common user tasks .
2,885
3
2,886
CRL developed [[ TUIT ]] to support their work to integrate << TIPSTER modules >> for the 6 and 12 month TIPSTER II demonstrations as well as their Oleada and Temple demonstration projects .
2,886
3
2,887
While such decoding is an essential underpinning , much recent work suggests that natural language interfaces will never appear cooperative or graceful unless << they >> also incorporate numerous [[ non-literal aspects of communication ]] , such as robust communication procedures .
2,887
4
2,888
While such decoding is an essential underpinning , much recent work suggests that natural language interfaces will never appear cooperative or graceful unless they also incorporate numerous << non-literal aspects of communication >> , such as [[ robust communication procedures ]] .
2,888
2
2,889
This paper defends that view , but claims that direct imitation of human performance is not the best way to implement many of these non-literal aspects of communication ; that the new technology of powerful << personal computers >> with integral [[ graphics displays ]] offers techniques superior to those of humans for these aspects , while still satisfying human communication needs .
2,889
4
2,890
This paper proposes a framework in which [[ Lagrangian Particle Dynamics ]] is used for the << segmentation of high density crowd flows >> and detection of flow instabilities .
2,890
3
2,891
This paper proposes a framework in which [[ Lagrangian Particle Dynamics ]] is used for the segmentation of high density crowd flows and << detection of flow instabilities >> .
2,891
3
2,892
This paper proposes a framework in which Lagrangian Particle Dynamics is used for the [[ segmentation of high density crowd flows ]] and << detection of flow instabilities >> .
2,892
0
2,893
For this purpose , a << flow field >> generated by a [[ moving crowd ]] is treated as an aperiodic dynamical system .
2,893
3
2,894
For this purpose , a << flow field >> generated by a moving crowd is treated as an [[ aperiodic dynamical system ]] .
2,894
3
2,895
A [[ grid of particles ]] is overlaid on the << flow field >> , and is advected using a numerical integration scheme .
2,895
3
2,896
A << grid of particles >> is overlaid on the flow field , and is advected using a [[ numerical integration scheme ]] .
2,896
3
2,897
The << evolution of particles >> through the flow is tracked using a [[ Flow Map ]] , whose spatial gradients are subsequently used to setup a Cauchy Green Deformation tensor for quantifying the amount by which the neighboring particles have diverged over the length of the integration .
2,897
3
2,898
The evolution of particles through the flow is tracked using a Flow Map , whose [[ spatial gradients ]] are subsequently used to setup a << Cauchy Green Deformation tensor >> for quantifying the amount by which the neighboring particles have diverged over the length of the integration .
2,898
3
2,899
The [[ maximum eigenvalue ]] of the << tensor >> is used to construct a Finite Time Lyapunov Exponent -LRB- FTLE -RRB- field , which reveals the Lagrangian Coherent Structures -LRB- LCS -RRB- present in the underlying flow .
2,899
1