,text,id,label 0,"This paper presents an [[ algorithm ]] for << computing optical flow , shape , motion , lighting , and albedo >> from an image sequence of a rigidly-moving Lambertian object under distant illumination .",0,3 1,"This paper presents an << algorithm >> for computing optical flow , shape , motion , lighting , and albedo from an [[ image sequence ]] of a rigidly-moving Lambertian object under distant illumination .",1,3 2,"This paper presents an algorithm for computing optical flow , shape , motion , lighting , and albedo from an << image sequence >> of a [[ rigidly-moving Lambertian object ]] under distant illumination .",2,1 3,"This paper presents an algorithm for computing optical flow , shape , motion , lighting , and albedo from an image sequence of a << rigidly-moving Lambertian object >> under [[ distant illumination ]] .",3,1 4,"The problem is formulated in a manner that subsumes structure from [[ motion ]] , << multi-view stereo >> , and photo-metric stereo as special cases .",4,0 5,"The problem is formulated in a manner that subsumes structure from motion , [[ multi-view stereo ]] , and << photo-metric stereo >> as special cases .",5,0 6,The << algorithm >> utilizes both [[ spatial and temporal intensity variation ]] as cues : the former constrains flow and the latter constrains surface orientation ; combining both cues enables dense reconstruction of both textured and texture-less surfaces .,6,3 7,The algorithm utilizes both spatial and temporal intensity variation as << cues >> : the [[ former ]] constrains flow and the latter constrains surface orientation ; combining both cues enables dense reconstruction of both textured and texture-less surfaces .,7,2 8,The algorithm utilizes both spatial and temporal intensity variation as cues : the [[ former ]] constrains << flow >> and the latter constrains surface orientation ; combining both cues enables dense reconstruction of both textured and texture-less surfaces .,8,3 9,The algorithm utilizes both spatial and temporal intensity variation as cues : the [[ former ]] constrains flow and the << latter >> constrains surface orientation ; combining both cues enables dense reconstruction of both textured and texture-less surfaces .,9,0 10,The algorithm utilizes both spatial and temporal intensity variation as << cues >> : the former constrains flow and the [[ latter ]] constrains surface orientation ; combining both cues enables dense reconstruction of both textured and texture-less surfaces .,10,2 11,The algorithm utilizes both spatial and temporal intensity variation as cues : the former constrains flow and the [[ latter ]] constrains << surface orientation >> ; combining both cues enables dense reconstruction of both textured and texture-less surfaces .,11,3 12,The algorithm utilizes both spatial and temporal intensity variation as cues : the former constrains flow and the latter constrains surface orientation ; combining both [[ cues ]] enables << dense reconstruction of both textured and texture-less surfaces >> .,12,3 13,"The << algorithm >> works by iteratively [[ estimating affine camera parameters , illumination , shape , and albedo ]] in an alternating fashion .",13,3 14,An [[ entity-oriented approach ]] to << restricted-domain parsing >> is proposed .,14,3 15,"Like semantic grammar , [[ this ]] allows easy exploitation of << limited domain semantics >> .",15,3 16,"In addition , [[ it ]] facilitates << fragmentary recognition >> and the use of multiple parsing strategies , and so is particularly useful for robust recognition of extra-grammatical input .",16,3 17,"In addition , [[ it ]] facilitates fragmentary recognition and the use of << multiple parsing strategies >> , and so is particularly useful for robust recognition of extra-grammatical input .",17,3 18,"In addition , it facilitates fragmentary recognition and the use of [[ multiple parsing strategies ]] , and so is particularly useful for robust << recognition of extra-grammatical input >> .",18,3 19,"Representative samples from an entity-oriented language definition are presented , along with a [[ control structure ]] for an << entity-oriented parser >> , some parsing strategies that use the control structure , and worked examples of parses .",19,3 20,"Representative samples from an entity-oriented language definition are presented , along with a control structure for an entity-oriented parser , some << parsing strategies >> that use the [[ control structure ]] , and worked examples of parses .",20,3 21,A << parser >> incorporating the [[ control structure ]] and the parsing strategies is currently under implementation .,21,4 22,This paper summarizes the formalism of Category Cooccurrence Restrictions -LRB- CCRs -RRB- and describes two [[ parsing algorithms ]] that interpret << it >> .,22,3 23,The use of CCRs leads to << syntactic descriptions >> formulated entirely with [[ restrictive statements ]] .,23,1 24,The paper shows how conventional [[ algorithms ]] for the analysis of context free languages can be adapted to the << CCR formalism >> .,24,3 25,The paper shows how conventional << algorithms >> for the analysis of [[ context free languages ]] can be adapted to the CCR formalism .,25,3 26,Special attention is given to the part of the parser that checks the fulfillment of [[ logical well-formedness conditions ]] on << trees >> .,26,1 27,We present a [[ text mining method ]] for finding << synonymous expressions >> based on the distributional hypothesis in a set of coherent corpora .,27,3 28,We present a << text mining method >> for finding synonymous expressions based on the [[ distributional hypothesis ]] in a set of coherent corpora .,28,3 29,This paper proposes a new methodology to improve the [[ accuracy ]] of a << term aggregation system >> using each author 's text as a coherent corpus .,29,6 30,This paper proposes a new << methodology >> to improve the accuracy of a [[ term aggregation system ]] using each author 's text as a coherent corpus .,30,6 31,"Our proposed method improves the [[ accuracy ]] of our << term aggregation system >> , showing that our approach is successful .",31,6 32,"Our proposed << method >> improves the accuracy of our [[ term aggregation system ]] , showing that our approach is successful .",32,6 33,"In this work , we present a [[ technique ]] for << robust estimation >> , which by explicitly incorporating the inherent uncertainty of the estimation procedure , results in a more efficient robust estimation algorithm .",33,3 34,"In this work , we present a [[ technique ]] for robust estimation , which by explicitly incorporating the inherent uncertainty of the estimation procedure , results in a more << efficient robust estimation algorithm >> .",34,3 35,"In this work , we present a << technique >> for robust estimation , which by explicitly incorporating the [[ inherent uncertainty of the estimation procedure ]] , results in a more efficient robust estimation algorithm .",35,3 36,"The combination of these two [[ strategies ]] results in a << robust estimation procedure >> that provides a significant speed-up over existing RANSAC techniques , while requiring no prior information to guide the sampling process .",36,3 37,"The combination of these two strategies results in a << robust estimation procedure >> that provides a significant speed-up over existing [[ RANSAC techniques ]] , while requiring no prior information to guide the sampling process .",37,5 38,"In particular , our [[ algorithm ]] requires , on average , 3-10 times fewer samples than standard << RANSAC >> , which is in close agreement with theoretical predictions .",38,5 39,The efficiency of the << algorithm >> is demonstrated on a selection of [[ geometric estimation problems ]] .,39,6 40,An attempt has been made to use an [[ Augmented Transition Network ]] as a procedural << dialog model >> .,40,2 41,The development of such a model appears to be important in several respects : as a << device >> to represent and to use different [[ dialog schemata ]] proposed in empirical conversation analysis ; as a device to represent and to use models of verbal interaction ; as a device combining knowledge about dialog schemata and about verbal interaction with knowledge about task-oriented and goal-directed dialogs .,41,3 42,The development of such a model appears to be important in several respects : as a device to represent and to use different [[ dialog schemata ]] proposed in empirical << conversation analysis >> ; as a device to represent and to use models of verbal interaction ; as a device combining knowledge about dialog schemata and about verbal interaction with knowledge about task-oriented and goal-directed dialogs .,42,3 43,The development of such a model appears to be important in several respects : as a device to represent and to use different dialog schemata proposed in empirical conversation analysis ; as a << device >> to represent and to use [[ models ]] of verbal interaction ; as a device combining knowledge about dialog schemata and about verbal interaction with knowledge about task-oriented and goal-directed dialogs .,43,3 44,The development of such a model appears to be important in several respects : as a device to represent and to use different dialog schemata proposed in empirical conversation analysis ; as a device to represent and to use [[ models ]] of << verbal interaction >> ; as a device combining knowledge about dialog schemata and about verbal interaction with knowledge about task-oriented and goal-directed dialogs .,44,3 45,The development of such a model appears to be important in several respects : as a device to represent and to use different dialog schemata proposed in empirical conversation analysis ; as a device to represent and to use models of verbal interaction ; as a device combining knowledge about [[ dialog schemata ]] and about << verbal interaction >> with knowledge about task-oriented and goal-directed dialogs .,45,0 46,A standard [[ ATN ]] should be further developed in order to account for the << verbal interactions >> of task-oriented dialogs .,46,3 47,A standard ATN should be further developed in order to account for the [[ verbal interactions ]] of << task-oriented dialogs >> .,47,1 48,We present a practically [[ unsupervised learning method ]] to produce << single-snippet answers >> to definition questions in question answering systems that supplement Web search engines .,48,3 49,We present a practically unsupervised learning method to produce single-snippet answers to definition questions in [[ question answering systems ]] that supplement << Web search engines >> .,49,3 50,"The [[ method ]] exploits << on-line encyclopedias and dictionaries >> to generate automatically an arbitrarily large number of positive and negative definition examples , which are then used to train an svm to separate the two classes .",50,3 51,"The method exploits [[ on-line encyclopedias and dictionaries ]] to generate automatically an arbitrarily large number of << positive and negative definition examples >> , which are then used to train an svm to separate the two classes .",51,3 52,"The method exploits on-line encyclopedias and dictionaries to generate automatically an arbitrarily large number of [[ positive and negative definition examples ]] , which are then used to train an << svm >> to separate the two classes .",52,3 53,"We show experimentally that the proposed method is viable , that [[ it ]] outperforms the << alternative >> of training the system on questions and news articles from trec , and that it helps the search engine handle definition questions significantly better .",53,5 54,"We show experimentally that the proposed method is viable , that it outperforms the alternative of training the << system >> on questions and [[ news articles ]] from trec , and that it helps the search engine handle definition questions significantly better .",54,3 55,"We show experimentally that the proposed method is viable , that it outperforms the alternative of training the system on questions and [[ news articles ]] from << trec >> , and that it helps the search engine handle definition questions significantly better .",55,4 56,"We show experimentally that the proposed method is viable , that it outperforms the alternative of training the system on questions and news articles from trec , and that [[ it ]] helps the << search engine >> handle definition questions significantly better .",56,3 57,We revisit the << classical decision-theoretic problem of weighted expert voting >> from a [[ statistical learning perspective ]] .,57,3 58,"In the case of known expert competence levels , we give [[ sharp error estimates ]] for the << optimal rule >> .",58,3 59,We analyze a [[ reweighted version of the Kikuchi approximation ]] for estimating the << log partition function of a product distribution >> defined over a region graph .,59,3 60,We analyze a reweighted version of the Kikuchi approximation for estimating the [[ log partition function of a product distribution ]] defined over a << region graph >> .,60,1 61,"We establish sufficient conditions for the [[ concavity ]] of our << reweighted objective function >> in terms of weight assignments in the Kikuchi expansion , and show that a reweighted version of the sum product algorithm applied to the Kikuchi region graph will produce global optima of the Kikuchi approximation whenever the algorithm converges .",61,1 62,"We establish sufficient conditions for the concavity of our reweighted objective function in terms of weight assignments in the Kikuchi expansion , and show that a [[ reweighted version of the sum product algorithm ]] applied to the << Kikuchi region graph >> will produce global optima of the Kikuchi approximation whenever the algorithm converges .",62,3 63,"We establish sufficient conditions for the concavity of our reweighted objective function in terms of weight assignments in the Kikuchi expansion , and show that a reweighted version of the sum product algorithm applied to the Kikuchi region graph will produce [[ global optima ]] of the << Kikuchi approximation >> whenever the algorithm converges .",63,1 64,"Finally , we provide an explicit characterization of the polytope of concavity in terms of the [[ cycle structure ]] of the << region graph >> .",64,1 65,We apply a [[ decision tree based approach ]] to << pronoun resolution >> in spoken dialogue .,65,3 66,We apply a decision tree based approach to [[ pronoun resolution ]] in << spoken dialogue >> .,66,3 67,Our [[ system ]] deals with << pronouns >> with NP - and non-NP-antecedents .,67,3 68,Our system deals with << pronouns >> with [[ NP - and non-NP-antecedents ]] .,68,3 69,We present a set of [[ features ]] designed for << pronoun resolution >> in spoken dialogue and determine the most promising features .,69,3 70,We present a set of features designed for [[ pronoun resolution ]] in << spoken dialogue >> and determine the most promising features .,70,3 71,We evaluate the << system >> on twenty [[ Switchboard dialogues ]] and show that it compares well to Byron 's -LRB- 2002 -RRB- manually tuned system .,71,6 72,We evaluate the system on twenty Switchboard dialogues and show that [[ it ]] compares well to << Byron 's -LRB- 2002 -RRB- manually tuned system >> .,72,5 73,"We present a new [[ approach ]] for building an efficient and robust << classifier >> for the two class problem , that localizes objects that may appear in the image under different orien-tations .",73,3 74,"We present a new approach for building an efficient and robust [[ classifier ]] for the two << class problem >> , that localizes objects that may appear in the image under different orien-tations .",74,3 75,"In contrast to other works that address this problem using multiple classifiers , each one specialized for a specific orientation , we propose a simple two-step << approach >> with an [[ estimation stage ]] and a classification stage .",75,4 76,"In contrast to other works that address this problem using multiple classifiers , each one specialized for a specific orientation , we propose a simple two-step approach with an [[ estimation stage ]] and a << classification stage >> .",76,0 77,"In contrast to other works that address this problem using multiple classifiers , each one specialized for a specific orientation , we propose a simple two-step << approach >> with an estimation stage and a [[ classification stage ]] .",77,4 78,The estimator yields an initial set of potential << object poses >> that are then validated by the [[ classifier ]] .,78,3 79,This methodology allows reducing the [[ time complexity ]] of the << algorithm >> while classification results remain high .,79,6 80,"The << classifier >> we use in both stages is based on a [[ boosted combination of Random Ferns ]] over local histograms of oriented gradients -LRB- HOGs -RRB- , which we compute during a pre-processing step .",80,3 81,"The classifier we use in both stages is based on a << boosted combination of Random Ferns >> over [[ local histograms of oriented gradients -LRB- HOGs -RRB- ]] , which we compute during a pre-processing step .",81,1 82,"The classifier we use in both stages is based on a boosted combination of Random Ferns over << local histograms of oriented gradients -LRB- HOGs -RRB- >> , which we compute during a [[ pre-processing step ]] .",82,3 83,Both the use of [[ supervised learning ]] and working on the gradient space makes our << approach >> robust while being efficient at run-time .,83,3 84,Both the use of supervised learning and working on the [[ gradient space ]] makes our << approach >> robust while being efficient at run-time .,84,3 85,"We show these properties by thorough testing on standard databases and on a new << database >> made of [[ motorbikes under planar rotations ]] , and with challenging conditions such as cluttered backgrounds , changing illumination conditions and partial occlusions .",85,1 86,"We show these properties by thorough testing on standard databases and on a new << database >> made of motorbikes under planar rotations , and with challenging [[ conditions ]] such as cluttered backgrounds , changing illumination conditions and partial occlusions .",86,1 87,"We show these properties by thorough testing on standard databases and on a new database made of motorbikes under planar rotations , and with challenging << conditions >> such as [[ cluttered backgrounds ]] , changing illumination conditions and partial occlusions .",87,2 88,"We show these properties by thorough testing on standard databases and on a new database made of motorbikes under planar rotations , and with challenging conditions such as [[ cluttered backgrounds ]] , << changing illumination conditions >> and partial occlusions .",88,0 89,"We show these properties by thorough testing on standard databases and on a new database made of motorbikes under planar rotations , and with challenging << conditions >> such as cluttered backgrounds , [[ changing illumination conditions ]] and partial occlusions .",89,2 90,"We show these properties by thorough testing on standard databases and on a new database made of motorbikes under planar rotations , and with challenging conditions such as cluttered backgrounds , [[ changing illumination conditions ]] and << partial occlusions >> .",90,0 91,"We show these properties by thorough testing on standard databases and on a new database made of motorbikes under planar rotations , and with challenging << conditions >> such as cluttered backgrounds , changing illumination conditions and [[ partial occlusions ]] .",91,2 92,A very simple improved [[ duration model ]] has reduced the error rate by about 10 % in both << triphone and semiphone systems >> .,92,3 93,A very simple improved duration model has reduced the [[ error rate ]] by about 10 % in both << triphone and semiphone systems >> .,93,6 94,"A new << training strategy >> has been tested which , by itself , did not provide useful improvements but suggests that improvements can be obtained by a related [[ rapid adaptation technique ]] .",94,3 95,"Finally , the << recognizer >> has been modified to use [[ bigram back-off language models ]] .",95,3 96,The [[ system ]] was then transferred from the << RM task >> to the ATIS CSR task and a limited number of development tests performed .,96,3 97,The [[ system ]] was then transferred from the RM task to the << ATIS CSR task >> and a limited number of development tests performed .,97,3 98,The system was then transferred from the [[ RM task ]] to the << ATIS CSR task >> and a limited number of development tests performed .,98,0 99,A new [[ approach ]] for << Interactive Machine Translation >> where the author interacts during the creation or the modification of the document is proposed .,99,3 100,This paper presents a new << interactive disambiguation scheme >> based on the [[ paraphrasing ]] of a parser 's multiple output .,100,3 101,We describe a novel [[ approach ]] to << statistical machine translation >> that combines syntactic information in the source language with recent advances in phrasal translation .,101,3 102,We describe a novel << approach >> to statistical machine translation that combines [[ syntactic information ]] in the source language with recent advances in phrasal translation .,102,4 103,We describe a novel approach to statistical machine translation that combines [[ syntactic information ]] in the source language with recent advances in << phrasal translation >> .,103,0 104,We describe a novel << approach >> to statistical machine translation that combines syntactic information in the source language with recent advances in [[ phrasal translation ]] .,104,4 105,"This << method >> requires a [[ source-language dependency parser ]] , target language word segmentation and an unsupervised word alignment component .",105,3 106,"This method requires a [[ source-language dependency parser ]] , << target language word segmentation >> and an unsupervised word alignment component .",106,0 107,"This << method >> requires a source-language dependency parser , [[ target language word segmentation ]] and an unsupervised word alignment component .",107,3 108,"This method requires a source-language dependency parser , [[ target language word segmentation ]] and an << unsupervised word alignment component >> .",108,0 109,"This << method >> requires a source-language dependency parser , target language word segmentation and an [[ unsupervised word alignment component ]] .",109,3 110,We describe an efficient decoder and show that using these [[ tree-based models ]] in combination with conventional << SMT models >> provides a promising approach that incorporates the power of phrasal SMT with the linguistic generality available in a parser .,110,0 111,We describe an efficient decoder and show that using these [[ tree-based models ]] in combination with conventional SMT models provides a promising << approach >> that incorporates the power of phrasal SMT with the linguistic generality available in a parser .,111,3 112,We describe an efficient decoder and show that using these tree-based models in combination with conventional [[ SMT models ]] provides a promising << approach >> that incorporates the power of phrasal SMT with the linguistic generality available in a parser .,112,3 113,We describe an efficient decoder and show that using these tree-based models in combination with conventional SMT models provides a promising approach that incorporates the power of [[ phrasal SMT ]] with the << linguistic generality >> available in a parser .,113,0 114,We describe an efficient decoder and show that using these tree-based models in combination with conventional SMT models provides a promising approach that incorporates the power of [[ phrasal SMT ]] with the linguistic generality available in a << parser >> .,114,3 115,We describe an efficient decoder and show that using these tree-based models in combination with conventional SMT models provides a promising approach that incorporates the power of phrasal SMT with the [[ linguistic generality ]] available in a << parser >> .,115,1 116,"<< Video >> provides not only rich [[ visual cues ]] such as motion and appearance , but also much less explored long-range temporal interactions among objects .",116,1 117,"Video provides not only rich << visual cues >> such as [[ motion ]] and appearance , but also much less explored long-range temporal interactions among objects .",117,2 118,"Video provides not only rich visual cues such as [[ motion ]] and << appearance >> , but also much less explored long-range temporal interactions among objects .",118,0 119,"Video provides not only rich << visual cues >> such as motion and [[ appearance ]] , but also much less explored long-range temporal interactions among objects .",119,2 120,We aim to capture such interactions and to construct a powerful [[ intermediate-level video representation ]] for subsequent << recognition >> .,120,3 121,"First , we develop an efficient << spatio-temporal video segmentation algorithm >> , which naturally incorporates [[ long-range motion cues ]] from the past and future frames in the form of clusters of point tracks with coherent motion .",121,3 122,"First , we develop an efficient spatio-temporal video segmentation algorithm , which naturally incorporates << long-range motion cues >> from the past and future frames in the form of [[ clusters of point tracks ]] with coherent motion .",122,3 123,"Second , we devise a new << track clustering cost function >> that includes [[ occlusion reasoning ]] , in the form of depth ordering constraints , as well as motion similarity along the tracks .",123,4 124,"Second , we devise a new track clustering cost function that includes << occlusion reasoning >> , in the form of [[ depth ordering constraints ]] , as well as motion similarity along the tracks .",124,1 125,"Second , we devise a new << track clustering cost function >> that includes occlusion reasoning , in the form of depth ordering constraints , as well as [[ motion similarity ]] along the tracks .",125,4 126,We evaluate the proposed << approach >> on a challenging set of [[ video sequences of office scenes ]] from feature length movies .,126,6 127,"In this paper , we introduce [[ KAZE features ]] , a novel << multiscale 2D feature detection and description algorithm >> in nonlinear scale spaces .",127,2 128,"In this paper , we introduce KAZE features , a novel << multiscale 2D feature detection and description algorithm >> in [[ nonlinear scale spaces ]] .",128,1 129,"In contrast , we detect and describe << 2D features >> in a [[ nonlinear scale space ]] by means of nonlinear diffusion filtering .",129,1 130,"In contrast , we detect and describe << 2D features >> in a nonlinear scale space by means of [[ nonlinear diffusion filtering ]] .",130,3 131,The << nonlinear scale space >> is built using efficient [[ Additive Operator Splitting -LRB- AOS -RRB- techniques ]] and variable con-ductance diffusion .,131,3 132,The nonlinear scale space is built using efficient [[ Additive Operator Splitting -LRB- AOS -RRB- techniques ]] and << variable con-ductance diffusion >> .,132,0 133,The << nonlinear scale space >> is built using efficient Additive Operator Splitting -LRB- AOS -RRB- techniques and [[ variable con-ductance diffusion ]] .,133,3 134,"Even though our [[ features ]] are somewhat more expensive to compute than << SURF >> due to the construction of the nonlinear scale space , but comparable to SIFT , our results reveal a step forward in performance both in detection and description against previous state-of-the-art methods .",134,5 135,"Even though our [[ features ]] are somewhat more expensive to compute than SURF due to the construction of the nonlinear scale space , but comparable to << SIFT >> , our results reveal a step forward in performance both in detection and description against previous state-of-the-art methods .",135,5 136,"Even though our features are somewhat more expensive to compute than SURF due to the construction of the nonlinear scale space , but comparable to SIFT , our [[ results ]] reveal a step forward in performance both in detection and description against previous << state-of-the-art methods >> .",136,5 137,"Even though our features are somewhat more expensive to compute than SURF due to the construction of the nonlinear scale space , but comparable to SIFT , our << results >> reveal a step forward in performance both in [[ detection ]] and description against previous state-of-the-art methods .",137,6 138,"Even though our features are somewhat more expensive to compute than SURF due to the construction of the nonlinear scale space , but comparable to SIFT , our results reveal a step forward in performance both in [[ detection ]] and << description >> against previous state-of-the-art methods .",138,0 139,"Even though our features are somewhat more expensive to compute than SURF due to the construction of the nonlinear scale space , but comparable to SIFT , our results reveal a step forward in performance both in [[ detection ]] and description against previous << state-of-the-art methods >> .",139,6 140,"Even though our features are somewhat more expensive to compute than SURF due to the construction of the nonlinear scale space , but comparable to SIFT , our << results >> reveal a step forward in performance both in detection and [[ description ]] against previous state-of-the-art methods .",140,6 141,"Even though our features are somewhat more expensive to compute than SURF due to the construction of the nonlinear scale space , but comparable to SIFT , our results reveal a step forward in performance both in detection and [[ description ]] against previous << state-of-the-art methods >> .",141,6 142,[[ Creating summaries ]] on lengthy Semantic Web documents for quick << identification of the corresponding entity >> has been of great contemporary interest .,142,3 143,<< Creating summaries >> on [[ lengthy Semantic Web documents ]] for quick identification of the corresponding entity has been of great contemporary interest .,143,3 144,"Specifically , we highlight the importance of << diversified -LRB- faceted -RRB- summaries >> by combining three dimensions : [[ diversity ]] , uniqueness , and popularity .",144,1 145,"Specifically , we highlight the importance of diversified -LRB- faceted -RRB- summaries by combining three dimensions : [[ diversity ]] , << uniqueness >> , and popularity .",145,0 146,"Specifically , we highlight the importance of << diversified -LRB- faceted -RRB- summaries >> by combining three dimensions : diversity , [[ uniqueness ]] , and popularity .",146,1 147,"Specifically , we highlight the importance of diversified -LRB- faceted -RRB- summaries by combining three dimensions : diversity , [[ uniqueness ]] , and << popularity >> .",147,0 148,"Specifically , we highlight the importance of << diversified -LRB- faceted -RRB- summaries >> by combining three dimensions : diversity , uniqueness , and [[ popularity ]] .",148,1 149,"Our novel << diversity-aware entity summarization approach >> mimics [[ human conceptual clustering techniques ]] to group facts , and picks representative facts from each group to form concise -LRB- i.e. , short -RRB- and comprehensive -LRB- i.e. , improved coverage through diversity -RRB- summaries .",149,3 150,We evaluate our [[ approach ]] against the state-of-the-art techniques and show that our work improves both the quality and the efficiency of << entity summarization >> .,150,3 151,We evaluate our << approach >> against the [[ state-of-the-art techniques ]] and show that our work improves both the quality and the efficiency of entity summarization .,151,5 152,We evaluate our approach against the [[ state-of-the-art techniques ]] and show that our work improves both the quality and the efficiency of << entity summarization >> .,152,3 153,We evaluate our approach against the state-of-the-art techniques and show that our work improves both the [[ quality ]] and the efficiency of << entity summarization >> .,153,6 154,We evaluate our approach against the state-of-the-art techniques and show that our work improves both the quality and the [[ efficiency ]] of << entity summarization >> .,154,6 155,We present a [[ framework ]] for the << fast computation of lexical affinity models >> .,155,3 156,"The << framework >> is composed of a novel [[ algorithm ]] to efficiently compute the co-occurrence distribution between pairs of terms , an independence model , and a parametric affinity model .",156,4 157,"The framework is composed of a novel [[ algorithm ]] to efficiently compute the << co-occurrence distribution >> between pairs of terms , an independence model , and a parametric affinity model .",157,3 158,"The framework is composed of a novel [[ algorithm ]] to efficiently compute the co-occurrence distribution between pairs of terms , an << independence model >> , and a parametric affinity model .",158,0 159,"The << framework >> is composed of a novel algorithm to efficiently compute the co-occurrence distribution between pairs of terms , an [[ independence model ]] , and a parametric affinity model .",159,4 160,"The framework is composed of a novel algorithm to efficiently compute the co-occurrence distribution between pairs of terms , an [[ independence model ]] , and a << parametric affinity model >> .",160,0 161,"The << framework >> is composed of a novel algorithm to efficiently compute the co-occurrence distribution between pairs of terms , an independence model , and a [[ parametric affinity model ]] .",161,4 162,"In comparison with previous models , which either use arbitrary windows to compute similarity between words or use [[ lexical affinity ]] to create << sequential models >> , in this paper we focus on models intended to capture the co-occurrence patterns of any pair of words or phrases at any distance in the corpus .",162,3 163,"In comparison with previous << models >> , which either use arbitrary windows to compute similarity between words or use lexical affinity to create sequential models , in this paper we focus on [[ models ]] intended to capture the co-occurrence patterns of any pair of words or phrases at any distance in the corpus .",163,5 164,"In comparison with previous models , which either use arbitrary windows to compute similarity between words or use lexical affinity to create sequential models , in this paper we focus on [[ models ]] intended to capture the << co-occurrence patterns >> of any pair of words or phrases at any distance in the corpus .",164,3 165,"We apply [[ it ]] in combination with a terabyte corpus to answer << natural language tests >> , achieving encouraging results .",165,3 166,"We apply << it >> in combination with a [[ terabyte corpus ]] to answer natural language tests , achieving encouraging results .",166,6 167,This paper introduces a [[ system ]] for << categorizing unknown words >> .,167,3 168,The << system >> is based on a [[ multi-component architecture ]] where each component is responsible for identifying one class of unknown words .,168,3 169,The system is based on a << multi-component architecture >> where each [[ component ]] is responsible for identifying one class of unknown words .,169,4 170,The system is based on a multi-component architecture where each [[ component ]] is responsible for identifying one class of << unknown words >> .,170,3 171,The focus of this paper is the [[ components ]] that identify << names >> and spelling errors .,171,3 172,The focus of this paper is the [[ components ]] that identify names and << spelling errors >> .,172,3 173,The focus of this paper is the components that identify [[ names ]] and << spelling errors >> .,173,0 174,Each << component >> uses a [[ decision tree architecture ]] to combine multiple types of evidence about the unknown word .,174,3 175,The << system >> is evaluated using data from [[ live closed captions ]] - a genre replete with a wide variety of unknown words .,175,6 176,"At MIT Lincoln Laboratory , we have been developing a << Korean-to-English machine translation system >> [[ CCLINC -LRB- Common Coalition Language System at Lincoln Laboratory -RRB- ]] .",176,2 177,"The << CCLINC Korean-to-English translation system >> consists of two [[ core modules ]] , language understanding and generation modules mediated by a language neutral meaning representation called a semantic frame .",177,4 178,"The CCLINC Korean-to-English translation system consists of two core modules , << language understanding and generation modules >> mediated by a [[ language neutral meaning representation ]] called a semantic frame .",178,3 179,"The CCLINC Korean-to-English translation system consists of two core modules , language understanding and generation modules mediated by a << language neutral meaning representation >> called a [[ semantic frame ]] .",179,2 180,"The key features of the system include : -LRB- i -RRB- Robust efficient parsing of [[ Korean ]] -LRB- a << verb final language >> with overt case markers , relatively free word order , and frequent omissions of arguments -RRB- .",180,2 181,"The key features of the system include : -LRB- i -RRB- Robust efficient parsing of Korean -LRB- a << verb final language >> with [[ overt case markers ]] , relatively free word order , and frequent omissions of arguments -RRB- .",181,1 182,-LRB- ii -RRB- High quality << translation >> via [[ word sense disambiguation ]] and accurate word order generation of the target language .,182,3 183,-LRB- ii -RRB- High quality translation via [[ word sense disambiguation ]] and accurate << word order generation >> of the target language .,183,0 184,-LRB- ii -RRB- High quality << translation >> via word sense disambiguation and accurate [[ word order generation ]] of the target language .,184,3 185,"Having been trained on [[ Korean newspaper articles ]] on missiles and chemical biological warfare , the << system >> produces the translation output sufficient for content understanding of the original document .",185,3 186,"Having been trained on << Korean newspaper articles >> on [[ missiles and chemical biological warfare ]] , the system produces the translation output sufficient for content understanding of the original document .",186,1 187,"The [[ JAVELIN system ]] integrates a flexible , planning-based architecture with a variety of language processing modules to provide an << open-domain question answering capability >> on free text .",187,3 188,"The << JAVELIN system >> integrates a flexible , [[ planning-based architecture ]] with a variety of language processing modules to provide an open-domain question answering capability on free text .",188,4 189,"The << JAVELIN system >> integrates a flexible , planning-based architecture with a variety of [[ language processing modules ]] to provide an open-domain question answering capability on free text .",189,4 190,"The JAVELIN system integrates a flexible , << planning-based architecture >> with a variety of [[ language processing modules ]] to provide an open-domain question answering capability on free text .",190,0 191,We present the first application of the [[ head-driven statistical parsing model ]] of Collins -LRB- 1999 -RRB- as a << simultaneous language model >> and parser for large-vocabulary speech recognition .,191,3 192,We present the first application of the [[ head-driven statistical parsing model ]] of Collins -LRB- 1999 -RRB- as a simultaneous language model and << parser >> for large-vocabulary speech recognition .,192,3 193,We present the first application of the head-driven statistical parsing model of Collins -LRB- 1999 -RRB- as a [[ simultaneous language model ]] and << parser >> for large-vocabulary speech recognition .,193,0 194,We present the first application of the head-driven statistical parsing model of Collins -LRB- 1999 -RRB- as a [[ simultaneous language model ]] and parser for << large-vocabulary speech recognition >> .,194,3 195,We present the first application of the head-driven statistical parsing model of Collins -LRB- 1999 -RRB- as a simultaneous language model and [[ parser ]] for << large-vocabulary speech recognition >> .,195,3 196,"The [[ model ]] is adapted to an << online left to right chart-parser >> for word lattices , integrating acoustic , n-gram , and parser probabilities .",196,3 197,"The model is adapted to an [[ online left to right chart-parser ]] for << word lattices >> , integrating acoustic , n-gram , and parser probabilities .",197,3 198,"The model is adapted to an << online left to right chart-parser >> for word lattices , integrating [[ acoustic , n-gram , and parser probabilities ]] .",198,4 199,"The << parser >> uses [[ structural and lexical dependencies ]] not considered by n-gram models , conditioning recognition on more linguistically-grounded relationships .",199,3 200,Experiments on the [[ Wall Street Journal treebank ]] and << lattice corpora >> show word error rates competitive with the standard n-gram language model while extracting additional structural information useful for speech understanding .,200,0 201,Experiments on the [[ Wall Street Journal treebank ]] and lattice corpora show word error rates competitive with the standard << n-gram language model >> while extracting additional structural information useful for speech understanding .,201,6 202,Experiments on the Wall Street Journal treebank and [[ lattice corpora ]] show word error rates competitive with the standard << n-gram language model >> while extracting additional structural information useful for speech understanding .,202,6 203,Experiments on the Wall Street Journal treebank and lattice corpora show [[ word error rates ]] competitive with the standard << n-gram language model >> while extracting additional structural information useful for speech understanding .,203,6 204,Experiments on the Wall Street Journal treebank and lattice corpora show word error rates competitive with the standard n-gram language model while extracting additional [[ structural information ]] useful for << speech understanding >> .,204,3 205,[[ Image composition -LRB- or mosaicing -RRB- ]] has attracted a growing attention in recent years as one of the main elements in << video analysis and representation >> .,205,4 206,In this paper we deal with the problem of [[ global alignment ]] and << super-resolution >> .,206,0 207,We also propose to evaluate the quality of the resulting << mosaic >> by measuring the [[ amount of blurring ]] .,207,6 208,<< Global registration >> is achieved by combining a [[ graph-based technique ]] -- that exploits the topological structure of the sequence induced by the spatial overlap -- with a bundle adjustment which uses only the homographies computed in the previous steps .,208,3 209,Global registration is achieved by combining a [[ graph-based technique ]] -- that exploits the << topological structure >> of the sequence induced by the spatial overlap -- with a bundle adjustment which uses only the homographies computed in the previous steps .,209,3 210,Global registration is achieved by combining a [[ graph-based technique ]] -- that exploits the topological structure of the sequence induced by the spatial overlap -- with a << bundle adjustment >> which uses only the homographies computed in the previous steps .,210,0 211,<< Global registration >> is achieved by combining a graph-based technique -- that exploits the topological structure of the sequence induced by the spatial overlap -- with a [[ bundle adjustment ]] which uses only the homographies computed in the previous steps .,211,3 212,Global registration is achieved by combining a graph-based technique -- that exploits the topological structure of the sequence induced by the spatial overlap -- with a << bundle adjustment >> which uses only the [[ homographies ]] computed in the previous steps .,212,3 213,Experimental comparison with other << techniques >> shows the effectiveness of our [[ approach ]] .,213,5 214,The main of this project is << computer-assisted acquisition and morpho-syntactic description of verb-noun collocations >> in [[ Polish ]] .,214,3 215,"We present methodology and resources obtained in three main project << phases >> which are : [[ dictionary-based acquisition of collocation lexicon ]] , feasibility study for corpus-based lexicon enlargement phase , corpus-based lexicon enlargement and collocation description .",215,2 216,"We present methodology and resources obtained in three main project phases which are : [[ dictionary-based acquisition of collocation lexicon ]] , << feasibility study >> for corpus-based lexicon enlargement phase , corpus-based lexicon enlargement and collocation description .",216,0 217,"We present methodology and resources obtained in three main project << phases >> which are : dictionary-based acquisition of collocation lexicon , [[ feasibility study ]] for corpus-based lexicon enlargement phase , corpus-based lexicon enlargement and collocation description .",217,2 218,"We present methodology and resources obtained in three main project phases which are : dictionary-based acquisition of collocation lexicon , [[ feasibility study ]] for << corpus-based lexicon enlargement phase >> , corpus-based lexicon enlargement and collocation description .",218,3 219,"We present methodology and resources obtained in three main project << phases >> which are : dictionary-based acquisition of collocation lexicon , feasibility study for corpus-based lexicon enlargement phase , [[ corpus-based lexicon enlargement and collocation description ]] .",219,2 220,"We present methodology and resources obtained in three main project phases which are : dictionary-based acquisition of collocation lexicon , << feasibility study >> for corpus-based lexicon enlargement phase , [[ corpus-based lexicon enlargement and collocation description ]] .",220,0 221,The presented here [[ corpus-based approach ]] permitted us to triple the size the << verb-noun collocation dictionary >> for Polish .,221,3 222,The presented here corpus-based approach permitted us to triple the size the << verb-noun collocation dictionary >> for [[ Polish ]] .,222,1 223,"Along with the increasing requirements , the [[ hash-tag recommendation task ]] for << microblogs >> has been receiving considerable attention in recent years .",223,3 224,"Motivated by the successful use of [[ convolutional neural networks -LRB- CNNs -RRB- ]] for many << natural language processing tasks >> , in this paper , we adopt CNNs to perform the hashtag recommendation problem .",224,3 225,"To incorporate the << trigger words >> whose effectiveness have been experimentally evaluated in several previous works , we propose a novel [[ architecture ]] with an attention mechanism .",225,3 226,"To incorporate the trigger words whose effectiveness have been experimentally evaluated in several previous works , we propose a novel << architecture >> with an [[ attention mechanism ]] .",226,1 227,The results of experiments on the [[ data ]] collected from a real world microblogging service demonstrated that the proposed << model >> outperforms state-of-the-art methods .,227,6 228,The results of experiments on the data collected from a real world microblogging service demonstrated that the proposed [[ model ]] outperforms << state-of-the-art methods >> .,228,5 229,"By incorporating trigger words into the consideration , the relative improvement of the proposed [[ method ]] over the << state-of-the-art method >> is around 9.4 % in the F1-score .",229,5 230,"By incorporating trigger words into the consideration , the relative improvement of the proposed method over the << state-of-the-art method >> is around 9.4 % in the [[ F1-score ]] .",230,6 231,"In this paper , we improve an << unsupervised learning method >> using the [[ Expectation-Maximization -LRB- EM -RRB- algorithm ]] proposed by Nigam et al. for text classification problems in order to apply it to word sense disambiguation -LRB- WSD -RRB- problems .",231,3 232,"In this paper , we improve an unsupervised learning method using the [[ Expectation-Maximization -LRB- EM -RRB- algorithm ]] proposed by Nigam et al. for << text classification problems >> in order to apply it to word sense disambiguation -LRB- WSD -RRB- problems .",232,3 233,"In this paper , we improve an unsupervised learning method using the Expectation-Maximization -LRB- EM -RRB- algorithm proposed by Nigam et al. for text classification problems in order to apply [[ it ]] to << word sense disambiguation -LRB- WSD -RRB- problems >> .",233,3 234,"In experiments , we solved 50 noun WSD problems in the [[ Japanese Dictionary Task ]] in << SENSEVAL2 >> .",234,1 235,"Furthermore , our [[ methods ]] were confirmed to be effective also for << verb WSD problems >> .",235,3 236,"[[ Dividing sentences in chunks of words ]] is a useful preprocessing step for << parsing >> , information extraction and information retrieval .",236,3 237,"[[ Dividing sentences in chunks of words ]] is a useful preprocessing step for parsing , << information extraction >> and information retrieval .",237,3 238,"[[ Dividing sentences in chunks of words ]] is a useful preprocessing step for parsing , information extraction and << information retrieval >> .",238,3 239,"Dividing sentences in chunks of words is a useful preprocessing step for [[ parsing ]] , << information extraction >> and information retrieval .",239,0 240,"Dividing sentences in chunks of words is a useful preprocessing step for parsing , [[ information extraction ]] and << information retrieval >> .",240,0 241,"-LRB- Ramshaw and Marcus , 1995 -RRB- have introduced a `` convenient '' [[ data representation ]] for << chunking >> by converting it to a tagging task .",241,3 242,In this paper we will examine seven different [[ data representations ]] for the problem of << recognizing noun phrase chunks >> .,242,3 243,"However , equipped with the most suitable [[ data representation ]] , our << memory-based learning chunker >> was able to improve the best published chunking results for a standard data set .",243,3 244,"However , equipped with the most suitable data representation , our << memory-based learning chunker >> was able to improve the best published chunking results for a standard [[ data set ]] .",244,6 245,"We focus on << FAQ-like questions and answers >> , and build our [[ system ]] around a noisy-channel architecture which exploits both a language model for answers and a transformation model for answer/question terms , trained on a corpus of 1 million question/answer pairs collected from the Web .",245,3 246,"We focus on FAQ-like questions and answers , and build our << system >> around a [[ noisy-channel architecture ]] which exploits both a language model for answers and a transformation model for answer/question terms , trained on a corpus of 1 million question/answer pairs collected from the Web .",246,3 247,"We focus on FAQ-like questions and answers , and build our system around a [[ noisy-channel architecture ]] which exploits both a << language model >> for answers and a transformation model for answer/question terms , trained on a corpus of 1 million question/answer pairs collected from the Web .",247,3 248,"We focus on FAQ-like questions and answers , and build our system around a [[ noisy-channel architecture ]] which exploits both a language model for answers and a << transformation model >> for answer/question terms , trained on a corpus of 1 million question/answer pairs collected from the Web .",248,3 249,In this paper we evaluate four objective [[ measures of speech ]] with regards to << intelligibility prediction >> of synthesized speech in diverse noisy situations .,249,6 250,In this paper we evaluate four objective measures of speech with regards to << intelligibility prediction >> of [[ synthesized speech ]] in diverse noisy situations .,250,3 251,In this paper we evaluate four objective measures of speech with regards to intelligibility prediction of << synthesized speech >> in [[ diverse noisy situations ]] .,251,1 252,"We evaluated three [[ intel-ligibility measures ]] , the Dau measure , the glimpse proportion and the Speech Intelligibility Index -LRB- SII -RRB- and a << quality measure >> , the Perceptual Evaluation of Speech Quality -LRB- PESQ -RRB- .",252,0 253,"We evaluated three << intel-ligibility measures >> , the [[ Dau measure ]] , the glimpse proportion and the Speech Intelligibility Index -LRB- SII -RRB- and a quality measure , the Perceptual Evaluation of Speech Quality -LRB- PESQ -RRB- .",253,2 254,"We evaluated three intel-ligibility measures , the [[ Dau measure ]] , the << glimpse proportion >> and the Speech Intelligibility Index -LRB- SII -RRB- and a quality measure , the Perceptual Evaluation of Speech Quality -LRB- PESQ -RRB- .",254,0 255,"We evaluated three << intel-ligibility measures >> , the Dau measure , the [[ glimpse proportion ]] and the Speech Intelligibility Index -LRB- SII -RRB- and a quality measure , the Perceptual Evaluation of Speech Quality -LRB- PESQ -RRB- .",255,2 256,"We evaluated three intel-ligibility measures , the Dau measure , the [[ glimpse proportion ]] and the << Speech Intelligibility Index -LRB- SII -RRB- >> and a quality measure , the Perceptual Evaluation of Speech Quality -LRB- PESQ -RRB- .",256,0 257,"We evaluated three << intel-ligibility measures >> , the Dau measure , the glimpse proportion and the [[ Speech Intelligibility Index -LRB- SII -RRB- ]] and a quality measure , the Perceptual Evaluation of Speech Quality -LRB- PESQ -RRB- .",257,2 258,"We evaluated three intel-ligibility measures , the Dau measure , the glimpse proportion and the Speech Intelligibility Index -LRB- SII -RRB- and a << quality measure >> , the [[ Perceptual Evaluation of Speech Quality -LRB- PESQ -RRB- ]] .",258,2 259,For the << generation of synthesized speech >> we used a state of the art [[ HMM-based speech synthesis system ]] .,259,3 260,The << noisy conditions >> comprised four [[ additive noises ]] .,260,4 261,The [[ measures ]] were compared with << subjective intelligibility scores >> obtained in listening tests .,261,5 262,"The results show the [[ Dau ]] and the << glimpse measures >> to be the best predictors of intelligibility , with correlations of around 0.83 to subjective scores .",262,0 263,"The results show the [[ Dau ]] and the glimpse measures to be the best << predictors of intelligibility >> , with correlations of around 0.83 to subjective scores .",263,2 264,"The results show the [[ Dau ]] and the glimpse measures to be the best predictors of intelligibility , with correlations of around 0.83 to << subjective scores >> .",264,5 265,"The results show the Dau and the [[ glimpse measures ]] to be the best << predictors of intelligibility >> , with correlations of around 0.83 to subjective scores .",265,2 266,"The results show the Dau and the [[ glimpse measures ]] to be the best predictors of intelligibility , with correlations of around 0.83 to << subjective scores >> .",266,5 267,"The results show the << Dau >> and the glimpse measures to be the best predictors of intelligibility , with [[ correlations ]] of around 0.83 to subjective scores .",267,6 268,"The results show the Dau and the << glimpse measures >> to be the best predictors of intelligibility , with [[ correlations ]] of around 0.83 to subjective scores .",268,6 269,All [[ measures ]] gave less accurate << predictions of intelligibility >> for synthetic speech than have previously been found for natural speech ; in particular the SII measure .,269,6 270,All measures gave less accurate << predictions of intelligibility >> for [[ synthetic speech ]] than have previously been found for natural speech ; in particular the SII measure .,270,3 271,All measures gave less accurate predictions of intelligibility for [[ synthetic speech ]] than have previously been found for << natural speech >> ; in particular the SII measure .,271,5 272,All << measures >> gave less accurate predictions of intelligibility for synthetic speech than have previously been found for natural speech ; in particular the [[ SII measure ]] .,272,2 273,"In additional experiments , we processed the << synthesized speech >> by an [[ ideal binary mask ]] before adding noise .",273,3 274,The [[ Glimpse measure ]] gave the most accurate << intelligibility predictions >> in this situation .,274,3 275,"A [[ '' graphics for vision '' approach ]] is proposed to address the problem of << reconstruction >> from a large and imperfect data set : reconstruction on demand by tensor voting , or ROD-TV .",275,3 276,"A '' graphics for vision '' approach is proposed to address the problem of << reconstruction >> from a [[ large and imperfect data set ]] : reconstruction on demand by tensor voting , or ROD-TV .",276,3 277,"A '' graphics for vision '' approach is proposed to address the problem of reconstruction from a large and imperfect data set : << reconstruction >> on demand by [[ tensor voting ]] , or ROD-TV .",277,3 278,"A '' graphics for vision '' approach is proposed to address the problem of reconstruction from a large and imperfect data set : reconstruction on demand by [[ tensor voting ]] , or << ROD-TV >> .",278,0 279,"A '' graphics for vision '' approach is proposed to address the problem of reconstruction from a large and imperfect data set : << reconstruction >> on demand by tensor voting , or [[ ROD-TV ]] .",279,3 280,"<< ROD-TV >> simultaneously delivers good [[ efficiency ]] and robust-ness , by adapting to a continuum of primitive connectivity , view dependence , and levels of detail -LRB- LOD -RRB- .",280,6 281,"<< ROD-TV >> simultaneously delivers good efficiency and [[ robust-ness ]] , by adapting to a continuum of primitive connectivity , view dependence , and levels of detail -LRB- LOD -RRB- .",281,6 282,"ROD-TV simultaneously delivers good << efficiency >> and [[ robust-ness ]] , by adapting to a continuum of primitive connectivity , view dependence , and levels of detail -LRB- LOD -RRB- .",282,0 283,"ROD-TV simultaneously delivers good efficiency and robust-ness , by adapting to a continuum of << primitive connectivity >> , [[ view dependence ]] , and levels of detail -LRB- LOD -RRB- .",283,0 284,"ROD-TV simultaneously delivers good efficiency and robust-ness , by adapting to a continuum of primitive connectivity , << view dependence >> , and [[ levels of detail -LRB- LOD -RRB- ]] .",284,0 285,[[ Locally inferred surface elements ]] are robust to noise and better capture << local shapes >> .,285,3 286,"By inferring [[ per-vertex normals ]] at sub-voxel precision on the fly , we can achieve << interpolative shading >> .",286,3 287,"By inferring << per-vertex normals >> at [[ sub-voxel precision ]] on the fly , we can achieve interpolative shading .",287,1 288,"By relaxing the [[ mesh connectivity requirement ]] , we extend ROD-TV and propose a simple but effective << multiscale feature extraction algorithm >> .",288,3 289,"By relaxing the mesh connectivity requirement , we extend [[ ROD-TV ]] and propose a simple but effective << multiscale feature extraction algorithm >> .",289,3 290,<< ROD-TV >> consists of a [[ hierarchical data structure ]] that encodes different levels of detail .,290,4 291,The << local reconstruction algorithm >> is [[ tensor voting ]] .,291,2 292,"<< It >> is applied on demand to the visible subset of data at a desired level of detail , by [[ traversing the data hierarchy ]] and collecting tensorial support in a neighborhood .",292,3 293,"It is applied on demand to the visible subset of data at a desired level of detail , by [[ traversing the data hierarchy ]] and << collecting tensorial support >> in a neighborhood .",293,0 294,"<< It >> is applied on demand to the visible subset of data at a desired level of detail , by traversing the data hierarchy and [[ collecting tensorial support ]] in a neighborhood .",294,3 295,Both [[ rhetorical structure ]] and << punctuation >> have been helpful in discourse processing .,295,0 296,Both [[ rhetorical structure ]] and punctuation have been helpful in << discourse processing >> .,296,3 297,Both rhetorical structure and [[ punctuation ]] have been helpful in << discourse processing >> .,297,3 298,"Based on a corpus annotation project , this paper reports the discursive usage of 6 [[ Chinese punctuation marks ]] in << news commentary texts >> : Colon , Dash , Ellipsis , Exclamation Mark , Question Mark , and Semicolon .",298,4 299,"Based on a corpus annotation project , this paper reports the discursive usage of 6 << Chinese punctuation marks >> in news commentary texts : [[ Colon ]] , Dash , Ellipsis , Exclamation Mark , Question Mark , and Semicolon .",299,2 300,"Based on a corpus annotation project , this paper reports the discursive usage of 6 Chinese punctuation marks in news commentary texts : [[ Colon ]] , << Dash >> , Ellipsis , Exclamation Mark , Question Mark , and Semicolon .",300,0 301,"Based on a corpus annotation project , this paper reports the discursive usage of 6 << Chinese punctuation marks >> in news commentary texts : Colon , [[ Dash ]] , Ellipsis , Exclamation Mark , Question Mark , and Semicolon .",301,2 302,"Based on a corpus annotation project , this paper reports the discursive usage of 6 Chinese punctuation marks in news commentary texts : Colon , [[ Dash ]] , << Ellipsis >> , Exclamation Mark , Question Mark , and Semicolon .",302,0 303,"Based on a corpus annotation project , this paper reports the discursive usage of 6 << Chinese punctuation marks >> in news commentary texts : Colon , Dash , [[ Ellipsis ]] , Exclamation Mark , Question Mark , and Semicolon .",303,2 304,"Based on a corpus annotation project , this paper reports the discursive usage of 6 Chinese punctuation marks in news commentary texts : Colon , Dash , [[ Ellipsis ]] , << Exclamation Mark >> , Question Mark , and Semicolon .",304,0 305,"Based on a corpus annotation project , this paper reports the discursive usage of 6 << Chinese punctuation marks >> in news commentary texts : Colon , Dash , Ellipsis , [[ Exclamation Mark ]] , Question Mark , and Semicolon .",305,2 306,"Based on a corpus annotation project , this paper reports the discursive usage of 6 Chinese punctuation marks in news commentary texts : Colon , Dash , Ellipsis , [[ Exclamation Mark ]] , << Question Mark >> , and Semicolon .",306,0 307,"Based on a corpus annotation project , this paper reports the discursive usage of 6 << Chinese punctuation marks >> in news commentary texts : Colon , Dash , Ellipsis , Exclamation Mark , [[ Question Mark ]] , and Semicolon .",307,2 308,"Based on a corpus annotation project , this paper reports the discursive usage of 6 Chinese punctuation marks in news commentary texts : Colon , Dash , Ellipsis , Exclamation Mark , [[ Question Mark ]] , and << Semicolon >> .",308,0 309,"Based on a corpus annotation project , this paper reports the discursive usage of 6 << Chinese punctuation marks >> in news commentary texts : Colon , Dash , Ellipsis , Exclamation Mark , Question Mark , and [[ Semicolon ]] .",309,2 310,The [[ rhetorical patterns ]] of these << marks >> are compared against patterns around cue phrases in general .,310,1 311,The [[ rhetorical patterns ]] of these marks are compared against << patterns around cue phrases >> in general .,311,5 312,"Results show that these [[ Chinese punctuation marks ]] , though fewer in number than << cue phrases >> , are easy to identify , have strong correlation with certain relations , and can be used as distinctive indicators of nuclearity in Chinese texts .",312,5 313,"Results show that these [[ Chinese punctuation marks ]] , though fewer in number than cue phrases , are easy to identify , have strong correlation with certain relations , and can be used as distinctive << indicators of nuclearity >> in Chinese texts .",313,3 314,"Results show that these Chinese punctuation marks , though fewer in number than cue phrases , are easy to identify , have strong correlation with certain relations , and can be used as distinctive << indicators of nuclearity >> in [[ Chinese texts ]] .",314,1 315,The << features >> based on [[ Markov random field -LRB- MRF -RRB- models ]] are usually sensitive to the rotation of image textures .,315,3 316,This paper develops an [[ anisotropic circular Gaussian MRF -LRB- ACGMRF -RRB- model ]] for << modelling rotated image textures >> and retrieving rotation-invariant texture features .,316,3 317,This paper develops an [[ anisotropic circular Gaussian MRF -LRB- ACGMRF -RRB- model ]] for modelling rotated image textures and << retrieving rotation-invariant texture features >> .,317,3 318,This paper develops an anisotropic circular Gaussian MRF -LRB- ACGMRF -RRB- model for [[ modelling rotated image textures ]] and << retrieving rotation-invariant texture features >> .,318,0 319,"To overcome the [[ singularity problem ]] of the << least squares estimate -LRB- LSE -RRB- method >> , an approximate least squares estimate -LRB- ALSE -RRB- method is proposed to estimate the parameters of the ACGMRF model .",319,1 320,"To overcome the singularity problem of the least squares estimate -LRB- LSE -RRB- method , an [[ approximate least squares estimate -LRB- ALSE -RRB- method ]] is proposed to estimate the << parameters of the ACGMRF model >> .",320,3 321,The << rotation-invariant features >> can be obtained from the [[ parameters of the ACGMRF model ]] by the one-dimensional -LRB- 1-D -RRB- discrete Fourier transform -LRB- DFT -RRB- .,321,3 322,The << rotation-invariant features >> can be obtained from the parameters of the ACGMRF model by the [[ one-dimensional -LRB- 1-D -RRB- discrete Fourier transform -LRB- DFT -RRB- ]] .,322,3 323,Significantly improved accuracy can be achieved by applying the [[ rotation-invariant features ]] to classify << SAR -LRB- synthetic aperture radar >> -RRB- sea ice and Brodatz imagery .,323,3 324,"Despite much recent progress on accurate << semantic role labeling >> , previous work has largely used [[ independent classifiers ]] , possibly combined with separate label sequence models via Viterbi decoding .",324,3 325,"Despite much recent progress on accurate semantic role labeling , previous work has largely used [[ independent classifiers ]] , possibly combined with separate << label sequence models >> via Viterbi decoding .",325,0 326,"Despite much recent progress on accurate semantic role labeling , previous work has largely used independent classifiers , possibly combined with separate << label sequence models >> via [[ Viterbi decoding ]] .",326,3 327,"We show how to build a joint model of argument frames , incorporating novel [[ features ]] that model these interactions into << discriminative log-linear models >> .",327,4 328,This << system >> achieves an [[ error reduction ]] of 22 % on all arguments and 32 % on core arguments over a state-of-the art independent classifier for gold-standard parse trees on PropBank .,328,6 329,This system achieves an [[ error reduction ]] of 22 % on all arguments and 32 % on core arguments over a state-of-the art << independent classifier >> for gold-standard parse trees on PropBank .,329,6 330,This << system >> achieves an error reduction of 22 % on all arguments and 32 % on core arguments over a state-of-the art [[ independent classifier ]] for gold-standard parse trees on PropBank .,330,5 331,This << system >> achieves an error reduction of 22 % on all arguments and 32 % on core arguments over a state-of-the art independent classifier for [[ gold-standard parse trees ]] on PropBank .,331,6 332,This system achieves an error reduction of 22 % on all arguments and 32 % on core arguments over a state-of-the art << independent classifier >> for [[ gold-standard parse trees ]] on PropBank .,332,6 333,This system achieves an error reduction of 22 % on all arguments and 32 % on core arguments over a state-of-the art independent classifier for [[ gold-standard parse trees ]] on << PropBank >> .,333,4 334,"In order to deal with << ambiguity >> , the [[ MORphological PArser MORPA ]] is provided with a probabilistic context-free grammar -LRB- PCFG -RRB- , i.e. it combines a `` conventional '' context-free morphological grammar to filter out ungrammatical segmentations with a probability-based scoring function which determines the likelihood of each successful parse .",334,3 335,"In order to deal with ambiguity , the << MORphological PArser MORPA >> is provided with a [[ probabilistic context-free grammar -LRB- PCFG -RRB- ]] , i.e. it combines a `` conventional '' context-free morphological grammar to filter out ungrammatical segmentations with a probability-based scoring function which determines the likelihood of each successful parse .",335,3 336,"In order to deal with ambiguity , the MORphological PArser MORPA is provided with a probabilistic context-free grammar -LRB- PCFG -RRB- , i.e. << it >> combines a [[ `` conventional '' context-free morphological grammar ]] to filter out ungrammatical segmentations with a probability-based scoring function which determines the likelihood of each successful parse .",336,3 337,"In order to deal with ambiguity , the MORphological PArser MORPA is provided with a probabilistic context-free grammar -LRB- PCFG -RRB- , i.e. it combines a [[ `` conventional '' context-free morphological grammar ]] to filter out << ungrammatical segmentations >> with a probability-based scoring function which determines the likelihood of each successful parse .",337,3 338,"In order to deal with ambiguity , the MORphological PArser MORPA is provided with a probabilistic context-free grammar -LRB- PCFG -RRB- , i.e. << it >> combines a `` conventional '' context-free morphological grammar to filter out ungrammatical segmentations with a [[ probability-based scoring function ]] which determines the likelihood of each successful parse .",338,3 339,"In order to deal with ambiguity , the MORphological PArser MORPA is provided with a probabilistic context-free grammar -LRB- PCFG -RRB- , i.e. it combines a << `` conventional '' context-free morphological grammar >> to filter out ungrammatical segmentations with a [[ probability-based scoring function ]] which determines the likelihood of each successful parse .",339,0 340,"In order to deal with ambiguity , the MORphological PArser MORPA is provided with a probabilistic context-free grammar -LRB- PCFG -RRB- , i.e. it combines a `` conventional '' context-free morphological grammar to filter out ungrammatical segmentations with a [[ probability-based scoring function ]] which determines the likelihood of each successful << parse >> .",340,3 341,Test performance data will show that a [[ PCFG ]] yields good results in << morphological parsing >> .,341,3 342,[[ MORPA ]] is a fully implemented << parser >> developed for use in a text-to-speech conversion system .,342,2 343,[[ MORPA ]] is a fully implemented parser developed for use in a << text-to-speech conversion system >> .,343,3 344,MORPA is a fully implemented [[ parser ]] developed for use in a << text-to-speech conversion system >> .,344,3 345,This paper describes the framework of a << Korean phonological knowledge base system >> using the [[ unification-based grammar formalism ]] : Korean Phonology Structure Grammar -LRB- KPSG -RRB- .,345,3 346,This paper describes the framework of a Korean phonological knowledge base system using the << unification-based grammar formalism >> : [[ Korean Phonology Structure Grammar -LRB- KPSG -RRB- ]] .,346,2 347,The [[ approach ]] of << KPSG >> provides an explicit development model for constructing a computational phonological system : speech recognition and synthesis system .,347,3 348,The approach of [[ KPSG ]] provides an explicit development model for constructing a computational << phonological system >> : speech recognition and synthesis system .,348,3 349,We show that the proposed [[ approach ]] is more describable than other << approaches >> such as those employing a traditional generative phonological approach .,349,5 350,We show that the proposed approach is more describable than other approaches such as << those >> employing a traditional [[ generative phonological approach ]] .,350,3 351,"In this paper , we study the [[ design of core-selecting payment rules ]] for such << domains >> .",351,3 352,We design two [[ core-selecting rules ]] that always satisfy << IR >> in expectation .,352,3 353,To study the performance of our << rules >> we perform a [[ computational Bayes-Nash equilibrium analysis ]] .,353,3 354,"We show that , in equilibrium , our new [[ rules ]] have better incentives , higher efficiency , and a lower rate of ex-post IR violations than standard << core-selecting rules >> .",354,5 355,"We show that , in equilibrium , our new << rules >> have better incentives , higher efficiency , and a lower [[ rate of ex-post IR violations ]] than standard core-selecting rules .",355,6 356,"We show that , in equilibrium , our new rules have better incentives , higher efficiency , and a lower [[ rate of ex-post IR violations ]] than standard << core-selecting rules >> .",356,6 357,"In this paper , we will describe a [[ search tool ]] for a huge set of << ngrams >> .",357,3 358,This system can be a very useful [[ tool ]] for << linguistic knowledge discovery >> and other NLP tasks .,358,3 359,This system can be a very useful [[ tool ]] for linguistic knowledge discovery and other << NLP tasks >> .,359,3 360,This system can be a very useful tool for [[ linguistic knowledge discovery ]] and other << NLP tasks >> .,360,0 361,This paper explores the role of [[ user modeling ]] in such << systems >> .,361,4 362,"Since acquiring the knowledge for a [[ user model ]] is a fundamental problem in << user modeling >> , a section is devoted to this topic .",362,3 363,"Next , the benefits and costs of implementing a [[ user modeling component ]] for a << system >> are weighed in light of several aspects of the interaction requirements that may be imposed by the system .",363,4 364,"[[ Information extraction techniques ]] automatically create << structured databases >> from unstructured data sources , such as the Web or newswire documents .",364,3 365,"<< Information extraction techniques >> automatically create structured databases from [[ unstructured data sources ]] , such as the Web or newswire documents .",365,3 366,"Information extraction techniques automatically create structured databases from << unstructured data sources >> , such as the [[ Web ]] or newswire documents .",366,2 367,"Information extraction techniques automatically create structured databases from unstructured data sources , such as the [[ Web ]] or << newswire documents >> .",367,0 368,"Information extraction techniques automatically create structured databases from << unstructured data sources >> , such as the Web or [[ newswire documents ]] .",368,2 369,"Despite the successes of these << systems >> , [[ accuracy ]] will always be imperfect .",369,6 370,"The << information extraction system >> we evaluate is based on a [[ linear-chain conditional random field -LRB- CRF -RRB- ]] , a probabilistic model which has performed well on information extraction tasks because of its ability to capture arbitrary , overlapping features of the input in a Markov model .",370,3 371,"The information extraction system we evaluate is based on a [[ linear-chain conditional random field -LRB- CRF -RRB- ]] , a << probabilistic model >> which has performed well on information extraction tasks because of its ability to capture arbitrary , overlapping features of the input in a Markov model .",371,2 372,"The information extraction system we evaluate is based on a linear-chain conditional random field -LRB- CRF -RRB- , a [[ probabilistic model ]] which has performed well on << information extraction tasks >> because of its ability to capture arbitrary , overlapping features of the input in a Markov model .",372,3 373,"The information extraction system we evaluate is based on a linear-chain conditional random field -LRB- CRF -RRB- , a [[ probabilistic model ]] which has performed well on information extraction tasks because of its ability to capture << arbitrary , overlapping features >> of the input in a Markov model .",373,3 374,"The information extraction system we evaluate is based on a linear-chain conditional random field -LRB- CRF -RRB- , a probabilistic model which has performed well on information extraction tasks because of its ability to capture [[ arbitrary , overlapping features ]] of the << input >> in a Markov model .",374,1 375,"The information extraction system we evaluate is based on a linear-chain conditional random field -LRB- CRF -RRB- , a probabilistic model which has performed well on information extraction tasks because of its ability to capture [[ arbitrary , overlapping features ]] of the input in a << Markov model >> .",375,4 376,"We implement several techniques to estimate the confidence of both [[ extracted fields ]] and entire << multi-field records >> , obtaining an average precision of 98 % for retrieving correct fields and 87 % for multi-field records .",376,0 377,"We implement several << techniques >> to estimate the confidence of both extracted fields and entire multi-field records , obtaining an [[ average precision ]] of 98 % for retrieving correct fields and 87 % for multi-field records .",377,6 378,"In this paper , we use the [[ information redundancy in multilingual input ]] to correct errors in << machine translation >> and thus improve the quality of multilingual summaries .",378,3 379,"In this paper , we use the [[ information redundancy in multilingual input ]] to correct errors in machine translation and thus improve the quality of << multilingual summaries >> .",379,3 380,"We demonstrate how errors in the << machine translations >> of the input [[ Arabic documents ]] can be corrected by identifying and generating from such redundancy , focusing on noun phrases .",380,3 381,"In this paper , we propose a new [[ approach ]] to generate << oriented object proposals -LRB- OOPs -RRB- >> to reduce the detection error caused by various orientations of the object .",381,3 382,"In this paper , we propose a new approach to generate << oriented object proposals -LRB- OOPs -RRB- >> to reduce the [[ detection error ]] caused by various orientations of the object .",382,6 383,"To this end , we propose to efficiently locate << object regions >> according to [[ pixelwise object probability ]] , rather than measuring the objectness from a set of sampled windows .",383,3 384,"To this end , we propose to efficiently locate object regions according to [[ pixelwise object probability ]] , rather than measuring the << objectness >> from a set of sampled windows .",384,5 385,"We formulate the << proposal generation problem >> as a [[ generative proba-bilistic model ]] such that object proposals of different shapes -LRB- i.e. , sizes and orientations -RRB- can be produced by locating the local maximum likelihoods .",385,3 386,"We formulate the proposal generation problem as a generative proba-bilistic model such that << object proposals >> of different [[ shapes ]] -LRB- i.e. , sizes and orientations -RRB- can be produced by locating the local maximum likelihoods .",386,1 387,"We formulate the proposal generation problem as a generative proba-bilistic model such that object proposals of different << shapes >> -LRB- i.e. , [[ sizes ]] and orientations -RRB- can be produced by locating the local maximum likelihoods .",387,2 388,"We formulate the proposal generation problem as a generative proba-bilistic model such that object proposals of different shapes -LRB- i.e. , [[ sizes ]] and << orientations >> -RRB- can be produced by locating the local maximum likelihoods .",388,0 389,"We formulate the proposal generation problem as a generative proba-bilistic model such that object proposals of different << shapes >> -LRB- i.e. , sizes and [[ orientations ]] -RRB- can be produced by locating the local maximum likelihoods .",389,2 390,"We formulate the proposal generation problem as a generative proba-bilistic model such that << object proposals >> of different shapes -LRB- i.e. , sizes and orientations -RRB- can be produced by locating the [[ local maximum likelihoods ]] .",390,3 391,"First , it helps the [[ object detector ]] handle objects of different << orientations >> .",391,3 392,"Third , [[ it ]] avoids massive window sampling , and thereby reducing the << number of proposals >> while maintaining a high recall .",392,3 393,"Third , << it >> avoids massive window sampling , and thereby reducing the number of proposals while maintaining a high [[ recall ]] .",393,6 394,Experiments on the [[ PASCAL VOC 2007 dataset ]] show that the proposed << OOP >> outperforms the state-of-the-art fast methods .,394,6 395,Experiments on the PASCAL VOC 2007 dataset show that the proposed [[ OOP ]] outperforms the << state-of-the-art fast methods >> .,395,5 396,Further experiments show that the [[ rotation invariant property ]] helps a << class-specific object detector >> achieve better performance than the state-of-the-art proposal generation methods in either object rotation scenarios or general scenarios .,396,3 397,Further experiments show that the rotation invariant property helps a [[ class-specific object detector ]] achieve better performance than the state-of-the-art << proposal generation methods >> in either object rotation scenarios or general scenarios .,397,5 398,Further experiments show that the rotation invariant property helps a << class-specific object detector >> achieve better performance than the state-of-the-art proposal generation methods in either [[ object rotation scenarios ]] or general scenarios .,398,6 399,Further experiments show that the rotation invariant property helps a class-specific object detector achieve better performance than the state-of-the-art << proposal generation methods >> in either [[ object rotation scenarios ]] or general scenarios .,399,6 400,Further experiments show that the rotation invariant property helps a class-specific object detector achieve better performance than the state-of-the-art proposal generation methods in either [[ object rotation scenarios ]] or << general scenarios >> .,400,0 401,Further experiments show that the rotation invariant property helps a << class-specific object detector >> achieve better performance than the state-of-the-art proposal generation methods in either object rotation scenarios or [[ general scenarios ]] .,401,6 402,Further experiments show that the rotation invariant property helps a class-specific object detector achieve better performance than the state-of-the-art << proposal generation methods >> in either object rotation scenarios or [[ general scenarios ]] .,402,6 403,"This paper describes three relatively [[ domain-independent capabilities ]] recently added to the << Paramax spoken language understanding system >> : non-monotonic reasoning , implicit reference resolution , and database query paraphrase .",403,4 404,"This paper describes three relatively << domain-independent capabilities >> recently added to the Paramax spoken language understanding system : [[ non-monotonic reasoning ]] , implicit reference resolution , and database query paraphrase .",404,2 405,"This paper describes three relatively << domain-independent capabilities >> recently added to the Paramax spoken language understanding system : non-monotonic reasoning , [[ implicit reference resolution ]] , and database query paraphrase .",405,2 406,"This paper describes three relatively << domain-independent capabilities >> recently added to the Paramax spoken language understanding system : non-monotonic reasoning , implicit reference resolution , and [[ database query paraphrase ]] .",406,2 407,"Finally , we briefly describe an experiment which we have done in extending the << n-best speech/language integration architecture >> to improving [[ OCR accuracy ]] .",407,6 408,"We investigate the problem of fine-grained sketch-based image retrieval -LRB- SBIR -RRB- , where [[ free-hand human sketches ]] are used as queries to perform << instance-level retrieval of images >> .",408,3 409,"This is an extremely challenging task because -LRB- i -RRB- visual comparisons not only need to be fine-grained but also executed cross-domain , -LRB- ii -RRB- free-hand -LRB- finger -RRB- sketches are highly abstract , making fine-grained matching harder , and most importantly -LRB- iii -RRB- [[ annotated cross-domain sketch-photo datasets ]] required for training are scarce , challenging many state-of-the-art << machine learning techniques >> .",409,3 410,We then develop a [[ deep triplet-ranking model ]] for << instance-level SBIR >> with a novel data augmentation and staged pre-training strategy to alleviate the issue of insufficient fine-grained training data .,410,3 411,We then develop a [[ deep triplet-ranking model ]] for instance-level SBIR with a novel data augmentation and staged pre-training strategy to alleviate the issue of << insufficient fine-grained training data >> .,411,3 412,We then develop a << deep triplet-ranking model >> for instance-level SBIR with a novel [[ data augmentation ]] and staged pre-training strategy to alleviate the issue of insufficient fine-grained training data .,412,3 413,We then develop a deep triplet-ranking model for instance-level SBIR with a novel [[ data augmentation ]] and << staged pre-training strategy >> to alleviate the issue of insufficient fine-grained training data .,413,0 414,We then develop a << deep triplet-ranking model >> for instance-level SBIR with a novel data augmentation and [[ staged pre-training strategy ]] to alleviate the issue of insufficient fine-grained training data .,414,3 415,Extensive experiments are carried out to contribute a variety of insights into the challenges of [[ data sufficiency ]] and << over-fitting avoidance >> when training deep networks for fine-grained cross-domain ranking tasks .,415,0 416,Extensive experiments are carried out to contribute a variety of insights into the challenges of data sufficiency and over-fitting avoidance when training [[ deep networks ]] for << fine-grained cross-domain ranking tasks >> .,416,3 417,In this paper we target at generating << generic action proposals >> in [[ unconstrained videos ]] .,417,3 418,"Each action proposal corresponds to a << temporal series of spatial bounding boxes >> , i.e. , a [[ spatio-temporal video tube ]] , which has a good potential to locate one human action .",418,2 419,"Each action proposal corresponds to a temporal series of spatial bounding boxes , i.e. , a [[ spatio-temporal video tube ]] , which has a good potential to locate one << human action >> .",419,3 420,"Assuming each action is performed by a human with meaningful motion , both [[ appearance and motion cues ]] are utilized to measure the << ac-tionness >> of the video tubes .",420,3 421,"Assuming each action is performed by a human with meaningful motion , both appearance and motion cues are utilized to measure the [[ ac-tionness ]] of the << video tubes >> .",421,6 422,"After picking those spatiotem-poral paths of high actionness scores , our << action proposal generation >> is formulated as a [[ maximum set coverage problem ]] , where greedy search is performed to select a set of action proposals that can maximize the overall actionness score .",422,3 423,"After picking those spatiotem-poral paths of high actionness scores , our action proposal generation is formulated as a maximum set coverage problem , where [[ greedy search ]] is performed to select a set of << action proposals >> that can maximize the overall actionness score .",423,3 424,"After picking those spatiotem-poral paths of high actionness scores , our action proposal generation is formulated as a maximum set coverage problem , where greedy search is performed to select a set of << action proposals >> that can maximize the overall [[ actionness score ]] .",424,6 425,"Compared with existing [[ action proposal approaches ]] , our << action proposals >> do not rely on video segmentation and can be generated in nearly real-time .",425,5 426,"Experimental results on two challenging [[ datasets ]] , MSRII and UCF 101 , validate the superior performance of our << action proposals >> as well as competitive results on action detection and search .",426,6 427,"Experimental results on two challenging << datasets >> , [[ MSRII ]] and UCF 101 , validate the superior performance of our action proposals as well as competitive results on action detection and search .",427,2 428,"Experimental results on two challenging datasets , [[ MSRII ]] and << UCF 101 >> , validate the superior performance of our action proposals as well as competitive results on action detection and search .",428,0 429,"Experimental results on two challenging << datasets >> , MSRII and [[ UCF 101 ]] , validate the superior performance of our action proposals as well as competitive results on action detection and search .",429,2 430,"Experimental results on two challenging datasets , MSRII and UCF 101 , validate the superior performance of our << action proposals >> as well as competitive results on [[ action detection and search ]] .",430,6 431,This paper reports recent research into [[ methods ]] for << creating natural language text >> .,431,3 432,"<< KDS -LRB- Knowledge Delivery System -RRB- >> , which embodies this [[ paradigm ]] , has distinct parts devoted to creation of the propositional units , to organization of the text , to prevention of excess redundancy , to creation of combinations of units , to evaluation of these combinations as potential sentences , to selection of the best among competing combinations , and to creation of the final text .",432,4 433,The Fragment-and-Compose paradigm and the [[ computational methods ]] of << KDS >> are described .,433,3 434,This paper explores the issue of using different [[ co-occurrence similarities ]] between terms for separating << query terms >> that are useful for retrieval from those that are harmful .,434,3 435,This paper explores the issue of using different co-occurrence similarities between terms for separating [[ query terms ]] that are useful for << retrieval >> from those that are harmful .,435,3 436,This paper explores the issue of using different co-occurrence similarities between terms for separating << query terms >> that are useful for retrieval from [[ those ]] that are harmful .,436,5 437,The hypothesis under examination is that [[ useful terms ]] tend to be more similar to each other than to other << query terms >> .,437,5 438,Preliminary experiments with << similarities >> computed using [[ first-order and second-order co-occurrence ]] seem to confirm the hypothesis .,438,3 439,"We propose a new [[ phrase-based translation model ]] and << decoding algorithm >> that enables us to evaluate and compare several , previously proposed phrase-based translation models .",439,0 440,"Within our framework , we carry out a large number of experiments to understand better and explain why [[ phrase-based models ]] outperform << word-based models >> .",440,5 441,"Our empirical results , which hold for all examined language pairs , suggest that the highest levels of performance can be obtained through relatively simple << means >> : [[ heuristic learning of phrase translations ]] from word-based alignments and lexical weighting of phrase translations .",441,2 442,"Our empirical results , which hold for all examined language pairs , suggest that the highest levels of performance can be obtained through relatively simple means : << heuristic learning of phrase translations >> from [[ word-based alignments ]] and lexical weighting of phrase translations .",442,3 443,"Our empirical results , which hold for all examined language pairs , suggest that the highest levels of performance can be obtained through relatively simple << means >> : heuristic learning of phrase translations from word-based alignments and [[ lexical weighting of phrase translations ]] .",443,2 444,"Traditional [[ methods ]] for << color constancy >> can improve surface re-flectance estimates from such uncalibrated images , but their output depends significantly on the background scene .",444,3 445,"Traditional [[ methods ]] for color constancy can improve << surface re-flectance estimates >> from such uncalibrated images , but their output depends significantly on the background scene .",445,3 446,"Traditional methods for color constancy can improve << surface re-flectance estimates >> from such [[ uncalibrated images ]] , but their output depends significantly on the background scene .",446,3 447,"We introduce the multi-view color constancy problem , and present a [[ method ]] to recover << estimates of underlying surface re-flectance >> based on joint estimation of these surface properties and the illuminants present in multiple images .",447,3 448,"The [[ method ]] can exploit << image correspondences >> obtained by various alignment techniques , and we show examples based on matching local region features .",448,3 449,"The method can exploit << image correspondences >> obtained by various [[ alignment techniques ]] , and we show examples based on matching local region features .",449,3 450,Our results show that [[ multi-view constraints ]] can significantly improve << estimates of both scene illuminants and object color -LRB- surface reflectance -RRB- >> when compared to a baseline single-view method .,450,3 451,Our results show that << multi-view constraints >> can significantly improve estimates of both scene illuminants and object color -LRB- surface reflectance -RRB- when compared to a [[ baseline single-view method ]] .,451,5 452,"Our contributions include a [[ concise , modular architecture ]] with reversible processes of << understanding >> and generation , an information-state model of reference , and flexible links between semantics and collaborative problem solving .",452,3 453,"Our contributions include a [[ concise , modular architecture ]] with reversible processes of understanding and << generation >> , an information-state model of reference , and flexible links between semantics and collaborative problem solving .",453,3 454,"Our contributions include a concise , modular architecture with reversible processes of [[ understanding ]] and << generation >> , an information-state model of reference , and flexible links between semantics and collaborative problem solving .",454,0