Unnamed: 0
int64 0
3.22k
| text
stringlengths 49
577
| id
int64 0
3.22k
| label
int64 0
6
|
---|---|---|---|
2,800 | [[ It ]] also gives significant speedups in << inference >> for several datasets with varying degrees of spatio-temporal continuity . | 2,800 | 3 |
2,801 | << It >> also gives significant speedups in inference for several [[ datasets ]] with varying degrees of spatio-temporal continuity . | 2,801 | 6 |
2,802 | It also gives significant speedups in inference for several << datasets >> with varying degrees of [[ spatio-temporal continuity ]] . | 2,802 | 1 |
2,803 | We also discuss the strengths and weaknesses of our [[ strategy ]] relative to existing << hierarchical approaches >> , and the kinds of image and video data that provide the best speedups . | 2,803 | 5 |
2,804 | Motivated by the success of [[ ensemble methods ]] in << machine learning >> and other areas of natural language processing , we developed a multi-strategy and multi-source approach to question answering which is based on combining the results from different answering agents searching for answers in multiple corpora . | 2,804 | 3 |
2,805 | Motivated by the success of [[ ensemble methods ]] in machine learning and other areas of << natural language processing >> , we developed a multi-strategy and multi-source approach to question answering which is based on combining the results from different answering agents searching for answers in multiple corpora . | 2,805 | 3 |
2,806 | Motivated by the success of ensemble methods in machine learning and other areas of natural language processing , we developed a [[ multi-strategy and multi-source approach ]] to << question answering >> which is based on combining the results from different answering agents searching for answers in multiple corpora . | 2,806 | 3 |
2,807 | The << answering agents >> adopt fundamentally different [[ strategies ]] , one utilizing primarily knowledge-based mechanisms and the other adopting statistical techniques . | 2,807 | 3 |
2,808 | The answering agents adopt fundamentally different << strategies >> , [[ one ]] utilizing primarily knowledge-based mechanisms and the other adopting statistical techniques . | 2,808 | 2 |
2,809 | The answering agents adopt fundamentally different strategies , << one >> utilizing primarily [[ knowledge-based mechanisms ]] and the other adopting statistical techniques . | 2,809 | 3 |
2,810 | The answering agents adopt fundamentally different << strategies >> , one utilizing primarily knowledge-based mechanisms and the [[ other ]] adopting statistical techniques . | 2,810 | 2 |
2,811 | The answering agents adopt fundamentally different strategies , one utilizing primarily knowledge-based mechanisms and the << other >> adopting [[ statistical techniques ]] . | 2,811 | 3 |
2,812 | We present our << multi-level answer resolution algorithm >> that combines results from the [[ answering agents ]] at the question , passage , and/or answer levels . | 2,812 | 3 |
2,813 | Experiments evaluating the effectiveness of our [[ answer resolution algorithm ]] show a 35.0 % relative improvement over our << baseline system >> in the number of questions correctly answered , and a 32.8 % improvement according to the average precision metric . | 2,813 | 5 |
2,814 | Experiments evaluating the effectiveness of our << answer resolution algorithm >> show a 35.0 % relative improvement over our baseline system in the number of questions correctly answered , and a 32.8 % improvement according to the [[ average precision metric ]] . | 2,814 | 6 |
2,815 | Experiments evaluating the effectiveness of our answer resolution algorithm show a 35.0 % relative improvement over our << baseline system >> in the number of questions correctly answered , and a 32.8 % improvement according to the [[ average precision metric ]] . | 2,815 | 6 |
2,816 | [[ Word Identification ]] has been an important and active issue in << Chinese Natural Language Processing >> . | 2,816 | 2 |
2,817 | In this paper , a new [[ mechanism ]] , based on the concept of sublanguage , is proposed for identifying << unknown words >> , especially personal names , in Chinese newspapers . | 2,817 | 3 |
2,818 | In this paper , a new << mechanism >> , based on the concept of [[ sublanguage ]] , is proposed for identifying unknown words , especially personal names , in Chinese newspapers . | 2,818 | 3 |
2,819 | In this paper , a new mechanism , based on the concept of sublanguage , is proposed for identifying << unknown words >> , especially [[ personal names ]] , in Chinese newspapers . | 2,819 | 2 |
2,820 | In this paper , a new << mechanism >> , based on the concept of sublanguage , is proposed for identifying unknown words , especially personal names , in [[ Chinese newspapers ]] . | 2,820 | 3 |
2,821 | The proposed << mechanism >> includes [[ title-driven name recognition ]] , adaptive dynamic word formation , identification of 2-character and 3-character Chinese names without title . | 2,821 | 4 |
2,822 | The proposed mechanism includes [[ title-driven name recognition ]] , << adaptive dynamic word formation >> , identification of 2-character and 3-character Chinese names without title . | 2,822 | 0 |
2,823 | The proposed << mechanism >> includes title-driven name recognition , [[ adaptive dynamic word formation ]] , identification of 2-character and 3-character Chinese names without title . | 2,823 | 4 |
2,824 | The proposed mechanism includes title-driven name recognition , [[ adaptive dynamic word formation ]] , << identification of 2-character and 3-character Chinese names without title >> . | 2,824 | 0 |
2,825 | The proposed << mechanism >> includes title-driven name recognition , adaptive dynamic word formation , [[ identification of 2-character and 3-character Chinese names without title ]] . | 2,825 | 4 |
2,826 | This report describes [[ Paul ]] , a << computer text generation system >> designed to create cohesive text through the use of lexical substitutions . | 2,826 | 2 |
2,827 | This report describes Paul , a [[ computer text generation system ]] designed to create << cohesive text >> through the use of lexical substitutions . | 2,827 | 3 |
2,828 | This report describes << Paul >> , a computer text generation system designed to create cohesive text through the use of [[ lexical substitutions ]] . | 2,828 | 3 |
2,829 | Specifically , this system is designed to deterministically choose between [[ pronominalization ]] , << superordinate substitution >> , and definite noun phrase reiteration . | 2,829 | 5 |
2,830 | Specifically , this system is designed to deterministically choose between pronominalization , [[ superordinate substitution ]] , and << definite noun phrase reiteration >> . | 2,830 | 5 |
2,831 | The [[ system ]] identifies a strength of << antecedence recovery >> for each of the lexical substitutions . | 2,831 | 3 |
2,832 | The system identifies a strength of [[ antecedence recovery ]] for each of the << lexical substitutions >> . | 2,832 | 3 |
2,833 | It describes the automated training and evaluation of an Optimal Position Policy , a [[ method ]] of locating the likely << positions of topic-bearing sentences >> based on genre-specific regularities of discourse structure . | 2,833 | 3 |
2,834 | It describes the automated training and evaluation of an Optimal Position Policy , a << method >> of locating the likely positions of topic-bearing sentences based on [[ genre-specific regularities of discourse structure ]] . | 2,834 | 3 |
2,835 | This [[ method ]] can be used in << applications >> such as information retrieval , routing , and text summarization . | 2,835 | 3 |
2,836 | This method can be used in << applications >> such as [[ information retrieval ]] , routing , and text summarization . | 2,836 | 2 |
2,837 | This method can be used in applications such as [[ information retrieval ]] , << routing >> , and text summarization . | 2,837 | 0 |
2,838 | This method can be used in << applications >> such as information retrieval , [[ routing ]] , and text summarization . | 2,838 | 2 |
2,839 | This method can be used in applications such as information retrieval , [[ routing ]] , and << text summarization >> . | 2,839 | 0 |
2,840 | This method can be used in << applications >> such as information retrieval , routing , and [[ text summarization ]] . | 2,840 | 2 |
2,841 | We describe a general [[ framework ]] for << online multiclass learning >> based on the notion of hypothesis sharing . | 2,841 | 3 |
2,842 | We describe a general << framework >> for online multiclass learning based on the [[ notion of hypothesis sharing ]] . | 2,842 | 3 |
2,843 | We generalize the [[ multiclass Perceptron ]] to our << framework >> and derive a unifying mistake bound analysis . | 2,843 | 3 |
2,844 | We demonstrate the merits of our approach by comparing [[ it ]] to previous << methods >> on both synthetic and natural datasets . | 2,844 | 5 |
2,845 | We demonstrate the merits of our approach by comparing << it >> to previous methods on both [[ synthetic and natural datasets ]] . | 2,845 | 6 |
2,846 | We demonstrate the merits of our approach by comparing it to previous << methods >> on both [[ synthetic and natural datasets ]] . | 2,846 | 6 |
2,847 | We describe a set of [[ supervised machine learning ]] experiments centering on the construction of << statistical models of WH-questions >> . | 2,847 | 3 |
2,848 | These << models >> , which are built from [[ shallow linguistic features of questions ]] , are employed to predict target variables which represent a user 's informational goals . | 2,848 | 3 |
2,849 | We argue in favor of the the use of [[ labeled directed graph ]] to represent various types of << linguistic structures >> , and illustrate how this allows one to view NLP tasks as graph transformations . | 2,849 | 3 |
2,850 | We argue in favor of the the use of [[ labeled directed graph ]] to represent various types of linguistic structures , and illustrate how this allows one to view << NLP tasks >> as graph transformations . | 2,850 | 3 |
2,851 | We argue in favor of the the use of labeled directed graph to represent various types of linguistic structures , and illustrate how [[ this ]] allows one to view << NLP tasks >> as graph transformations . | 2,851 | 3 |
2,852 | We present a general [[ method ]] for learning such << transformations >> from an annotated corpus and describe experiments with two applications of the method : identification of non-local depenencies -LRB- using Penn Treebank data -RRB- and semantic role labeling -LRB- using Proposition Bank data -RRB- . | 2,852 | 3 |
2,853 | We present a general << method >> for learning such transformations from an [[ annotated corpus ]] and describe experiments with two applications of the method : identification of non-local depenencies -LRB- using Penn Treebank data -RRB- and semantic role labeling -LRB- using Proposition Bank data -RRB- . | 2,853 | 3 |
2,854 | We present a general method for learning such transformations from an annotated corpus and describe experiments with two << applications >> of the [[ method ]] : identification of non-local depenencies -LRB- using Penn Treebank data -RRB- and semantic role labeling -LRB- using Proposition Bank data -RRB- . | 2,854 | 3 |
2,855 | We present a general method for learning such transformations from an annotated corpus and describe experiments with two << applications >> of the method : [[ identification of non-local depenencies ]] -LRB- using Penn Treebank data -RRB- and semantic role labeling -LRB- using Proposition Bank data -RRB- . | 2,855 | 2 |
2,856 | We present a general method for learning such transformations from an annotated corpus and describe experiments with two applications of the method : << identification of non-local depenencies >> -LRB- using [[ Penn Treebank data ]] -RRB- and semantic role labeling -LRB- using Proposition Bank data -RRB- . | 2,856 | 3 |
2,857 | We present a general method for learning such transformations from an annotated corpus and describe experiments with two << applications >> of the method : identification of non-local depenencies -LRB- using Penn Treebank data -RRB- and [[ semantic role labeling ]] -LRB- using Proposition Bank data -RRB- . | 2,857 | 2 |
2,858 | We present a general method for learning such transformations from an annotated corpus and describe experiments with two applications of the method : identification of non-local depenencies -LRB- using Penn Treebank data -RRB- and << semantic role labeling >> -LRB- using [[ Proposition Bank data ]] -RRB- . | 2,858 | 3 |
2,859 | We describe a generative probabilistic model of natural language , which we call HBG , that takes advantage of detailed [[ linguistic information ]] to resolve << ambiguity >> . | 2,859 | 3 |
2,860 | [[ HBG ]] incorporates lexical , syntactic , semantic , and structural information from the parse tree into the << disambiguation process >> in a novel way . | 2,860 | 3 |
2,861 | << HBG >> incorporates [[ lexical , syntactic , semantic , and structural information ]] from the parse tree into the disambiguation process in a novel way . | 2,861 | 3 |
2,862 | We use a [[ corpus of bracketed sentences ]] , called a Treebank , in combination with << decision tree building >> to tease out the relevant aspects of a parse tree that will determine the correct parse of a sentence . | 2,862 | 0 |
2,863 | We use a [[ corpus of bracketed sentences ]] , called a Treebank , in combination with decision tree building to tease out the relevant aspects of a << parse tree >> that will determine the correct parse of a sentence . | 2,863 | 3 |
2,864 | We use a corpus of bracketed sentences , called a Treebank , in combination with [[ decision tree building ]] to tease out the relevant aspects of a << parse tree >> that will determine the correct parse of a sentence . | 2,864 | 3 |
2,865 | We use a corpus of bracketed sentences , called a Treebank , in combination with decision tree building to tease out the relevant aspects of a [[ parse tree ]] that will determine the correct << parse >> of a sentence . | 2,865 | 3 |
2,866 | This stands in contrast to the usual approach of further [[ grammar tailoring ]] via the usual linguistic introspection in the hope of generating the correct << parse >> . | 2,866 | 3 |
2,867 | This stands in contrast to the usual approach of further << grammar tailoring >> via the usual [[ linguistic introspection ]] in the hope of generating the correct parse . | 2,867 | 3 |
2,868 | In head-to-head tests against one of the best existing << robust probabilistic parsing models >> , which we call [[ P-CFG ]] , the HBG model significantly outperforms P-CFG , increasing the parsing accuracy rate from 60 % to 75 % , a 37 % reduction in error . | 2,868 | 2 |
2,869 | In head-to-head tests against one of the best existing robust probabilistic parsing models , which we call P-CFG , the [[ HBG model ]] significantly outperforms << P-CFG >> , increasing the parsing accuracy rate from 60 % to 75 % , a 37 % reduction in error . | 2,869 | 5 |
2,870 | In head-to-head tests against one of the best existing robust probabilistic parsing models , which we call P-CFG , the << HBG model >> significantly outperforms P-CFG , increasing the [[ parsing accuracy rate ]] from 60 % to 75 % , a 37 % reduction in error . | 2,870 | 6 |
2,871 | The framework of the << analysis >> is [[ model-theoretic semantics ]] . | 2,871 | 3 |
2,872 | This paper addresses the issue of << word-sense ambiguity >> in extraction from [[ machine-readable resources ]] for the construction of large-scale knowledge sources . | 2,872 | 3 |
2,873 | This paper addresses the issue of word-sense ambiguity in extraction from [[ machine-readable resources ]] for the << construction of large-scale knowledge sources >> . | 2,873 | 3 |
2,874 | We describe two experiments : one which ignored word-sense distinctions , resulting in 6.3 % [[ accuracy ]] for << semantic classification >> of verbs based on -LRB- Levin , 1993 -RRB- ; and one which exploited word-sense distinctions , resulting in 97.9 % accuracy . | 2,874 | 6 |
2,875 | These experiments were dual purpose : -LRB- 1 -RRB- to validate the central thesis of the work of -LRB- Levin , 1993 -RRB- , i.e. , that [[ verb semantics ]] and << syntactic behavior >> are predictably related ; -LRB- 2 -RRB- to demonstrate that a 15-fold improvement can be achieved in deriving semantic information from syntactic cues if we first divide the syntactic cues into distinct groupings that correlate with different word senses . | 2,875 | 0 |
2,876 | These experiments were dual purpose : -LRB- 1 -RRB- to validate the central thesis of the work of -LRB- Levin , 1993 -RRB- , i.e. , that verb semantics and syntactic behavior are predictably related ; -LRB- 2 -RRB- to demonstrate that a 15-fold improvement can be achieved in deriving << semantic information >> from [[ syntactic cues ]] if we first divide the syntactic cues into distinct groupings that correlate with different word senses . | 2,876 | 3 |
2,877 | Finally , we show that we can provide effective acquisition [[ techniques ]] for novel << word senses >> using a combination of online sources . | 2,877 | 3 |
2,878 | Finally , we show that we can provide effective acquisition << techniques >> for novel word senses using a combination of [[ online sources ]] . | 2,878 | 3 |
2,879 | The [[ TIPSTER Architecture ]] has been designed to enable a variety of different << text applications >> to use a set of common text processing modules . | 2,879 | 3 |
2,880 | The TIPSTER Architecture has been designed to enable a variety of different << text applications >> to use a set of [[ common text processing modules ]] . | 2,880 | 3 |
2,881 | Since [[ user interfaces ]] work best when customized for particular << applications >> , it is appropriator that no particular user interface styles or conventions are described in the TIPSTER Architecture specification . | 2,881 | 3 |
2,882 | However , the Computing Research Laboratory -LRB- CRL -RRB- has constructed several << TIPSTER applications >> that use a common set of configurable [[ Graphical User Interface -LRB- GUI -RRB- functions ]] . | 2,882 | 3 |
2,883 | These << GUIs >> were constructed using [[ CRL 's TIPSTER User Interface Toolkit -LRB- TUIT -RRB- ]] . | 2,883 | 3 |
2,884 | [[ TUIT ]] is a << software library >> that can be used to construct multilingual TIPSTER user interfaces for a set of common user tasks . | 2,884 | 2 |
2,885 | [[ TUIT ]] is a software library that can be used to construct << multilingual TIPSTER user interfaces >> for a set of common user tasks . | 2,885 | 3 |
2,886 | CRL developed [[ TUIT ]] to support their work to integrate << TIPSTER modules >> for the 6 and 12 month TIPSTER II demonstrations as well as their Oleada and Temple demonstration projects . | 2,886 | 3 |
2,887 | While such decoding is an essential underpinning , much recent work suggests that natural language interfaces will never appear cooperative or graceful unless << they >> also incorporate numerous [[ non-literal aspects of communication ]] , such as robust communication procedures . | 2,887 | 4 |
2,888 | While such decoding is an essential underpinning , much recent work suggests that natural language interfaces will never appear cooperative or graceful unless they also incorporate numerous << non-literal aspects of communication >> , such as [[ robust communication procedures ]] . | 2,888 | 2 |
2,889 | This paper defends that view , but claims that direct imitation of human performance is not the best way to implement many of these non-literal aspects of communication ; that the new technology of powerful << personal computers >> with integral [[ graphics displays ]] offers techniques superior to those of humans for these aspects , while still satisfying human communication needs . | 2,889 | 4 |
2,890 | This paper proposes a framework in which [[ Lagrangian Particle Dynamics ]] is used for the << segmentation of high density crowd flows >> and detection of flow instabilities . | 2,890 | 3 |
2,891 | This paper proposes a framework in which [[ Lagrangian Particle Dynamics ]] is used for the segmentation of high density crowd flows and << detection of flow instabilities >> . | 2,891 | 3 |
2,892 | This paper proposes a framework in which Lagrangian Particle Dynamics is used for the [[ segmentation of high density crowd flows ]] and << detection of flow instabilities >> . | 2,892 | 0 |
2,893 | For this purpose , a << flow field >> generated by a [[ moving crowd ]] is treated as an aperiodic dynamical system . | 2,893 | 3 |
2,894 | For this purpose , a << flow field >> generated by a moving crowd is treated as an [[ aperiodic dynamical system ]] . | 2,894 | 3 |
2,895 | A [[ grid of particles ]] is overlaid on the << flow field >> , and is advected using a numerical integration scheme . | 2,895 | 3 |
2,896 | A << grid of particles >> is overlaid on the flow field , and is advected using a [[ numerical integration scheme ]] . | 2,896 | 3 |
2,897 | The << evolution of particles >> through the flow is tracked using a [[ Flow Map ]] , whose spatial gradients are subsequently used to setup a Cauchy Green Deformation tensor for quantifying the amount by which the neighboring particles have diverged over the length of the integration . | 2,897 | 3 |
2,898 | The evolution of particles through the flow is tracked using a Flow Map , whose [[ spatial gradients ]] are subsequently used to setup a << Cauchy Green Deformation tensor >> for quantifying the amount by which the neighboring particles have diverged over the length of the integration . | 2,898 | 3 |
2,899 | The [[ maximum eigenvalue ]] of the << tensor >> is used to construct a Finite Time Lyapunov Exponent -LRB- FTLE -RRB- field , which reveals the Lagrangian Coherent Structures -LRB- LCS -RRB- present in the underlying flow . | 2,899 | 1 |