Benjamin Aw
Add updated pkl file v3
6fa4bc9
raw
history blame contribute delete
No virus
54.1 kB
{
"paper_id": "M95-1016",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T03:12:52.004872Z"
},
"title": "DESCRIPTION OF THE SAIC DX SYSTE M AS USED FOR MUC -6",
"authors": [
{
"first": "Lance",
"middle": [
"A"
],
"last": "Miller",
"suffix": "",
"affiliation": {},
"email": "lamiller@cpva.saic.com"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "BACKGROUND This is a very young project, operational for only a few months. The focus of the effort is dataextraction, the identification of instances of data-classes in \"commercial\" text-e .g., newspapers, technical reports, business correspondence, intelligence briefs. Instances of data-classes in text ar e phrases which identify factual content, such as names of people or organizations, products, financia l amounts, quantities, and so forth. ' A number of applications would be very well served simply with automated and accurate identification of the data-classes that occurred in their texts of interest, with interpretations left to experts. Such applications include extraction of bibliographic information, document indexing, competitiv e analyses based on open sources. technical information retrieval, foreign technology and politica l assessments, tracking financial and other resource transactions in the written media, and various types of link analyses based on text correlations. Providing very accurate automated data-extraction capability for this type of application is th e objective of the DX project. Relative to the broad language understanding goals of many projects, thi s effort is clearly a niche one, but perhaps all the more ambitious in its accuracy goals for that reason. Last October, 1994, we started designing a new pattern-recognition language to be developed for just the purpose of data-extraction. In May, 1995, the first version of the language, DXL (for \"Data eXtraction In our view, data-classes comprise IDs and \"quantifications\". IDs are either proper names or technical descriptions (of a \"part-of\" or \"kind-of\" type). Quantifications are either measurements or mathematical or logical expressions .",
"pdf_parse": {
"paper_id": "M95-1016",
"_pdf_hash": "",
"abstract": [
{
"text": "BACKGROUND This is a very young project, operational for only a few months. The focus of the effort is dataextraction, the identification of instances of data-classes in \"commercial\" text-e .g., newspapers, technical reports, business correspondence, intelligence briefs. Instances of data-classes in text ar e phrases which identify factual content, such as names of people or organizations, products, financia l amounts, quantities, and so forth. ' A number of applications would be very well served simply with automated and accurate identification of the data-classes that occurred in their texts of interest, with interpretations left to experts. Such applications include extraction of bibliographic information, document indexing, competitiv e analyses based on open sources. technical information retrieval, foreign technology and politica l assessments, tracking financial and other resource transactions in the written media, and various types of link analyses based on text correlations. Providing very accurate automated data-extraction capability for this type of application is th e objective of the DX project. Relative to the broad language understanding goals of many projects, thi s effort is clearly a niche one, but perhaps all the more ambitious in its accuracy goals for that reason. Last October, 1994, we started designing a new pattern-recognition language to be developed for just the purpose of data-extraction. In May, 1995, the first version of the language, DXL (for \"Data eXtraction In our view, data-classes comprise IDs and \"quantifications\". IDs are either proper names or technical descriptions (of a \"part-of\" or \"kind-of\" type). Quantifications are either measurements or mathematical or logical expressions .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "Language\"), was available and training begun, but because of personnel unavailability, serious use of i t did not begin until September.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "Other elements of the DX system are similarly quite new . The underlying text-processing syste m was adapted from an ongoing SAIC-Navy project to convert legacy technical manuals into a particula r data-base form suitable for interactive electronic use on a portable computer . 2 Numerous modifications have been made to this system including a new data-base access capability developed in September fo r rapid access to many millions of words and facts . The fact contents of the data-base (called the knowledge-bank), such as names of places, persons, and organizations, were acquired only in the last fe w months and entered using a coding scheme which represents their many syntactic and semantic features. Also recent is the accumulation of the large number of critical (for our approach) \"signal-words\" whic h mark the presence of certain data-classes --for example, \"Mr .\", \"Ave .\", and \"Corp.\" suggest the nearby occurrence of person-names, street-names, and organization-names, respectively . Similarly, the needed linguistic infrastructure has also only recently been developed : the compilation from several sources of a large lexicon with parts-of-speech, the development of the capability to find the \"roots\" of words prior t o lexical look-up (e.g ., the root of \"placing\" is \"place\"), and the morphological analysis capability to permi t guessing at the part-of-speech of unknown words based on their suffixes .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "The life-cycle of the DX project is thus independent of the MUC, which has been a very useful, i f not the best-timed, experience.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "In this section we summarize the elements of our design objectives and provide extensive detail s of the general system implementation.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "SYSTEM DESCRIPTIO N",
"sec_num": null
},
{
"text": "Two sets of considerations determined our selected approach to the data-extraction problem : our set of performance objectives, and the results of our analyses of data-classes . Concerning performance, we wanted a system that would most of all be highly accurate --less than 5% misses and false alarms --ye t robust in the sense of failing seldom, failing soft, and restoring easily. We also wanted a system that was easily extendible, serviceable by programmers, supportive of informed guessing (but giving confidenc e and basis), and eventually capable of learning extensions . Extremely rapid processing is not a requirement, in the belief that achieving the quality goals first could be followed subsequent speed-u p enhancements (e .g ., parallelization) .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Design Approac h",
"sec_num": null
},
{
"text": "Our analysis of a wide variety of data-classes indicates that the majority of classes, and o f instances, do not require sentence-level linguistic information for their detection and identification . That is, it appears that having a good sentence parse would only infrequently be of value in determining the data-classes embedded in the sentences . The needed information seems to be primarily available withi n the data-class phrase itself, in the form of obligatory and optional elements . For the majority of \" difficult\" cases, it would appear that the semantic features of local adjacent contexts plus some global context (e.g. , the type of document) could resolve the identifications .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Design Approac h",
"sec_num": null
},
{
"text": "These considerations motivated the three-stage strategy we adopted for the DX project . The first is unabashedly a brute-force method and is to identify by look-up : for those data-classes like person-This conversion activity is part of the Navy program to develop the support technologies for producing and using Interactive Electroni c Technical Manuals (IETM) for new and existing systems, as governed by the MIL-M-87268 standard. In turn, DXL is now being used as the specification language for recognizing various components of existing Navy tech manuals in the SAIC IETM project . names and organizations that are amenable to this app ; oach, get as many instances as possible and enter in the knowledge-bank and then check for them in any new input . Secondly, identify by pattern, using a language specially developed to have all the functionality useful for this approach, especially very powerful pattern-matching capability coupled with the facility of placing arbitrary constraints on th e patterns or local contexts. The final step is to identify using semantics-based contextual reasoning, wherein potential remaining data-class targets are fuzzily classified as one of several related data-classe s (e .g ., proper names) on the basis of suggestive cues (e .g., capitalization) . Then, reasoning heuristics (e .g. , \"names of people often have appositives whose heads or modifiers are marked semantically as bein g characteristic of people\") are used to direct the search for discriminating contextual clues .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Design Approac h",
"sec_num": null
},
{
"text": "This three-fold approach can be argued as meeting most of the performance goals and certainl y fits the finding of data-classes being pattern-based. However, it is an empirical question whether the desired accuracy levels --particularly <5% misses --can truly be achieved without at least a reasonabl e identification of the case-roles of the surface constituents . For this reason, an additional design requirement for the new pattern language was that it would have the capability to represent very powerfu l sentence-level parsing grammars (e .g., context-sensitive rules, unrestricted look-ahead, eas y representation of discontinuous constituents).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Design Approac h",
"sec_num": null
},
{
"text": "A much-simplified view of the DX system is given in Figure 1 . The first step is to convert the character-stream input into a stream of \"tokens\", essentially words, based primarily on the presence o f blanks between character-strings . Upon identification, each token is annotated with a set of \"prime\" attribute-values including : the most important attribute \"type\" (\"word\", \"num\", \"sgml\", \"punc\", or \"mix\"), start/stop character positions, \"chars\" (the token's character string), capitalization (\"capital\" , \"upper\", \"lower\", or \"mixed\"), and token number . The value of the \"chars\" attribute is that string which has all beginning brackets (or parentheses or braces) removed as well as all ending brackets an d punctuation of all kinds ; appropriate attributes are set to indicate what was removed . In addition, the singular possessive marking \" `s \" is also removed and a possessive attribute set. In this way, the toke n \"chars\" value can be looked up in the knowledge-bank without worrying about punctuation of any kind .",
"cite_spans": [],
"ref_spans": [
{
"start": 52,
"end": 60,
"text": "Figure 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "System-level Overview.",
"sec_num": null
},
{
"text": "All of the subsequent system processing is accomplished with specific pattern-rules and involve s moving up and down the token stream, checking tokens for particular attribute values, and, whe n successful, either changing the attribute-values of single tokens or replacing several tokens with a new one (with a new \"type\" attribute, whose \"chars\" attribute is the concatenation of those replaced) . For any particular rule, the whole document is searched from the first token to the last .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "System-level Overview.",
"sec_num": null
},
{
"text": "The second major processing step is to check the knowledge-bank to see whether the \"chars\" value of tokens are known as an instance of a data-class of interest (e .g., known locations o r organizations) . If so, the tokens corresponding to the known data-class entry are replaced with a single token whose \"type\" is set to the name of the data-class . In addition, new attributes are added, includin g part-of-speech, the word-stem, number (singular or plural) . Token \"chars\" not recognized as data-classe s are analyzed morphologically to determine their part-of-speech, word-stem, and number .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "System-level Overview.",
"sec_num": null
},
{
"text": "In the third step, the major data-class recognition rules are applied . The knowledge-bank is again a major source of information but this time primarily for \"signal-words\" which usually preface o r terminate data-class patterns . For example, a number of organizations fit the canonical form of a \"prefix organizational title\" followed by \"of' or \"for\" followed by a location or unknown capitalized words (e .g. \"Bank of New York\") . Prefix titles like \"bank\" are characterized in the knowledge-bank by a number o f",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "System-level Overview.",
"sec_num": null
},
{
"text": "a) N Cb C a ) C 4-3 U) j C CO a) a) - O a) U a) C) L C O O .~c _ 4-l C C O O _ a_ .- -O U O L. a) 0_ O C LC D O .- CO \u2022--C D to N u) W CD U 4-3 CD a) 4-3 CD a) a . Ls) C CO a) _ U 0 a. U~4-3 C a ) cu cD W . 2 O _C C a) C3 03 Cn a) -C L C O a) L C o a) W CO U (n 4 - 4- 4- 4-3 L a) E a) a) E CO CL 4-3 O CO a CA CA U) C",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "System-level Overview.",
"sec_num": null
},
{
"text": "attributes, such as being capitalized, not a name, referring to an organization, whose activity i s commercial. A token whose \" chars\" value is \"bank\" will not have these attribute-values alread y associated with it. Rather, only when some rule asks whether the token has the attribute values associated with an organizational title will the knowledge-bank be checked for these . As a result of checking for such attributes, the values returned will be set on the token . This process of adding values to a token onl y when a rule has inquired of the attributes is known as \"lazy annotation\" and is much more economica l than attaching all possible attribute-values slavishly to all known knowledge-bank entries .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "System-level Overview.",
"sec_num": null
},
{
"text": "An example of these processing steps is given in Figure 2 for the input \"Dr . Joyce Smith lost money.\" The tokenization and associated attribute-annotation for the first two words is shown as lists of attribute-value pairs in the LISP-list format used by the Scheme programming language in which th e system is written (supplemented by a few C and C++ modules) . The next line on Figure 2 shows what th e results would be were the knowledge-bank looked-up for instances of various kinds of titles .' In this case the character-string \"dr\" is associated with two types of titles, one a prefix for person-names and one a suffix for in-city street references ; both values are returned for the attribute title_typ . For a person-name recognition rule such as is shown in Figure 3 below, the likely result would be the replacement of the firs t three tokens (\"dr\", \"Joyce\", and \"smith\") with a new single token of type \"person\", with start/stop value s reflecting the span of coverage in the input, and with a new attribute giving the name of the rule that was successfully applied (\"cap-person\") . Outputs include both a tagging of the target in the input and th e inclusion of the target characters in a list of other such person targets .",
"cite_spans": [],
"ref_spans": [
{
"start": 49,
"end": 57,
"text": "Figure 2",
"ref_id": null
},
{
"start": 380,
"end": 388,
"text": "Figure 2",
"ref_id": null
},
{
"start": 768,
"end": 776,
"text": "Figure 3",
"ref_id": null
}
],
"eq_spans": [],
"section": "System-level Overview.",
"sec_num": null
},
{
"text": "Five of the major features of DXL are : (1) extremely flexible pattern specification, (2 ) unrestricted rewrite capability, (3) the capability to put constraints on pattern elements or to invoke globa l constraints, (4) the capability to invoke, anywhere, the full power of the Scheme programming language , and (5) the ability to expand complex \"chars\" strings .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Data-Extraction Language, DXL",
"sec_num": null
},
{
"text": "The first feature is illustrated by Figure 3 , which shows a DXL rule for one of the PERSON canonical forms . DXL rules have three components : a left-hand-side (LHS) specifying the total pattern that is to be matched, a right-pointing arrow indicating that the LHS is to be rewritten, and a right-handside (RHS) indicating what is to be substituted for the LHS, with what actions . In Figure 2 , the LHS has four elements : an optional pattern followed by an obligatory one, followed by two optional patterns . This rule would match instances such as \"Dr . Harry Morgan Raffler, Jr ., Vice President\" and \"Frank\". \" The meaning of the prefix symbols is given in the following table : * An optional operator, zero or more occurrences, match the minimu m ?",
"cite_spans": [],
"ref_spans": [
{
"start": 36,
"end": 44,
"text": "Figure 3",
"ref_id": null
},
{
"start": 386,
"end": 394,
"text": "Figure 2",
"ref_id": null
}
],
"eq_spans": [],
"section": "The Data-Extraction Language, DXL",
"sec_num": null
},
{
"text": "An optional operator, zero or one occurrence, match the minimu m + Obligatory element, at least one, possibly more When placed after the above, forces a match of the maximum number of specified pattern s Indicates a defined pattern",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Data-Extraction Language, DXL",
"sec_num": null
},
{
"text": "The LHS therefore involves four defined patterns : 0 or more prefix titles for people (such as \"Prof.\"), at least one capitalized word (with no comma following it), an optional family title such as \"Jr .\" , and an optional suffix title, such as \"Director \" . Two of the four BNF pattern definitions are given at the bottom of the Figure . That for title_pref, for example, specifies that the \"chars \" value should have an initial capitalized letter, that the type of the title ought to be \"pref' (for prefix), and that the value of cl , the second hierarchical knowledge classification value (c0 is the first), should be \"person\" .",
"cite_spans": [],
"ref_spans": [
{
"start": 330,
"end": 338,
"text": "Figure .",
"ref_id": null
}
],
"eq_spans": [],
"section": "The Data-Extraction Language, DXL",
"sec_num": null
},
{
"text": "a) U) 7 cD a) :.-0 i i ' - CL3 Q C~a) 0 O 0 > .F\u0168 ) \u2022/ - C s Q CM_ u) 1 a) a) Q A - ' i T O 0 C p as o C' - 0 L - a) . 4-- a) 2 { 0 L 0 0 CO 0) + 0 v V 0- CZS C~ c a) - U) U) CD a) CI) u) = U U 0 7:3 -0 a) 0 0 II I I A C ii ~ ~0 i a) ca.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Data-Extraction Language, DXL",
"sec_num": null
},
{
"text": "Discontinuous constituents are easily specified using the \"*\" (Kleene star) operator, as in @first pattern * .any @second_pattern",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "U )",
"sec_num": null
},
{
"text": "where \" .any\" means a token with any \"type\" value, and the star indicates zero or arbitrarily many intervening tokens between the two patterns of interest . To rewrite these into new patterns requires use o f the dynamic variable-assignment facility which takes whatever tokens were matched by the pattern and assigns them, as a list, to the variable indicated, as shown b y @first.attern ->vi *.any->v2 @second_pattern ->v3",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "U )",
"sec_num": null
},
{
"text": "_> $v1 [ actions ] Sv2 Sv3 [ actions ]",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "U )",
"sec_num": null
},
{
"text": "On the LHS, variables vl-v3 are assigned to the three pattern elements; those binding to vl and v 3 should, in this example, be understood as binding to single tokens, and these are referenced on the RHS , via the \"$\" operator. Single tokens bound to vl and v3 are modified by attribute-changing actions, and al l the bound tokens are finally reinserted back where they were in the token stream .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "U )",
"sec_num": null
},
{
"text": "The second, rewrite, feature of DXL is illustrated by the abstract rule below, which, for clarity' s sake, omits the needed variable-assignments :",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "U )",
"sec_num": null
},
{
"text": "\\ A \\ B [X] C D (Y) /E/ _> C F B",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "U )",
"sec_num": null
},
{
"text": "This rule specifies a left-context of A after which three obligatory patterns are sought, B-D, where B has some additional condition X placed on it . There is also a global test Y which has to succeed, all this i n the right context of E . If all this LHS succeeds, then B and C are to be permuted, D is to be deleted, an d F is to be added (inserted between C and B), with neither A nor E being further involved . With LHS context-sensitivity and the ability to permute, add, and delete LHS elements --and also unrestricted lookahead --the rewrite power is essentially unconstrained, beyond recursively-enumerable context-sensitiv e grammars, having the power of a Turing machine.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "U )",
"sec_num": null
},
{
"text": "Local and global constraints (the third feature) have been illustrated, and they are often expressed in practice by dropping into full Scheme code (the fourth feature) . The fifth feature, expansion is very useful when the \"chars\" attribute is a mixture of letters/numbers/symbols . Expansion permits the components of a complex \"chars\" string to be broken apart and analyzed using the same rule formalisms employed for multiple tokens . For example, the following rule expands all tokens of \"type\" \"mix\" : type:\"mix\" ->vi => (expand vl )",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "U )",
"sec_num": null
},
{
"text": "The result of expanding a token whose \"chars\" are \"451bs ./sq.in .\", is to replace this token in the token stream with the following ones :",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "U )",
"sec_num": null
},
{
"text": ".sot 45 tbs . / sq. in. .eo t",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "U )",
"sec_num": null
},
{
"text": "The first and last tokens have special type values, \"start of (expanded) token\" and \"end of (expanded ) token\", but they have no start or stop attributes, being \"dimensionless\" with respect to characters . The inserted tokens can now be recognized as a measure by a rules such as the following :",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "U )",
"sec_num": null
},
{
"text": ".sot type :\"num\" @unit \"/\" @unit .eot => measure where the pattern @unit is appropriately defined as a single or multi-word reference to a unit of measure .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "U )",
"sec_num": null
},
{
"text": "The knowledge-bank has five major types of entries (with the approximate quantities presently being added given in parentheses) : (1) words or phrases which are given a part-of-speech and perhaps a word-stem attribute, but little else (64K) ; (2) words/phrases which are full instances of data-classes (suc h as person first-names, cities, organizations) (6M) ; (3) \"signal-words\", usually called \"titles\" whic h indicate the likely presence of data-classes (such as \"Ave.\", \"p .m.\", \"Bank\", \"Mlle .\") (1K) ; (4 ) \"clusters\" of ordinary words which share significant semantic features (such as \"communication acts\" ) (2k); and (5) isolated ordinary words which have particular significance of one kind or another (1K) .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Knowledge-Ban k",
"sec_num": null
},
{
"text": "The first type of entry, mostly common words, facilitates sentence parsing when needed . The second type implement the brute-force \"identify by look-up\" principle of the DX system . The third type contribute the most in supporting the pattern-based identification principle of the system, as facilitated b y the fourth and fifth types .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Knowledge-Ban k",
"sec_num": null
},
{
"text": "The last four types have a part-of-speech attribute plus several classification attributes plus a number of other features . It is this rich set of features that are used as conditions on the DXL pattern elements and largely underlie the potential for very high accuracy in target detection . A sampling of knowledge-bank entries illustrating these features is given in Table 1 .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Knowledge-Ban k",
"sec_num": null
},
{
"text": "In the first row of Table 1 the first \"key\" column indicates the lower-case look-up character s that will be matched against the \"chars\" values from the input text . The rest of the column headings in the first row are various attributes which are relevant to the examples givens The entry for \"susan \" shows that its highest knowledge-classification feature, c0, is \"hument\" (\"human entity\"), subcategorized by the next classification feature, cl, as \"person\", in contrast to the second entry, \"ibm\" , which is an organization. These two entries also are coded as being names and capitalized. The third entry, \"mr\" is also categorized as referring to a person and capitalized but is not a name, rather a title, o f type \"pref' (for \"prefix\") . The fourth entry, \"militant\", again refers to a person, but is neither a name nor a title (nor capitalized) but has a role value of either \"political\" or \"terrorist\" . The string \"texas\" i s identified as a place, sub-category \"center\" and sub-sub-category \"state\", and it is both a name an d capitalized . The last two examples both have the knowledge-classification features meas-time-date, bu t \"christmas\" is also categorized as a \"holiday\" and is both capitalized and a name ; the phrase \"pai d holiday\" is neither capitalized nor a name but fills the semantic role of referring to a \"holiday\" entity .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Knowledge-Ban k",
"sec_num": null
},
{
"text": "These examples suffice to illustrate the rich set of syntactic and semantic knowledge-ban k features that are available for use in the DXL rules to identify and discriminate among even very simila r data-classes . 6 These features would also provide the basis for a neural-net or deterministic classificatio n approach (e .g ., C4 .5) for a learning capability to be developed later.'",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Knowledge-Ban k",
"sec_num": null
},
{
"text": "The Table 1 attribute-by-entry structure is convenient for exposition, but given there are now well over a hundred total possibl e attributes, such a structure would be very space-inefficient . The knowledge-bank actually has a much different representation . Also, not all relevant attributes are shown. For example \"susan\" also has a name_type attribute with the multiple values of \"given I female\".",
"cite_spans": [],
"ref_spans": [
{
"start": 4,
"end": 11,
"text": "Table 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "SYSTEM PERFORMANC \u00c8",
"sec_num": null
},
{
"text": "6 For example, the existing features can discriminate among five types of capitalized references relating to people, as indicated by th e following: \"reagan\", \"american\", \"christian\", \"irish\", and \"mr . president\".",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "SYSTEM PERFORMANC \u00c8",
"sec_num": null
},
{
"text": "We envision examples of new data-classes first being clustered by hand into high-similarity groups, but this may not be necessary .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "SYSTEM PERFORMANC \u00c8",
"sec_num": null
},
{
"text": "Co U cu o a) E co U CO CO 0 L. _3 CO CO Cl) -0 -0 C- I I 0 1 i E Q O O 0 c--0 C C C L O .O O a) u) U) 4-3 a) a) L. L. c a) c a) a) a) E E 0 O 0.. 0 U . 0 U 4-D 4 4 C C C C CD U) a) U) ) E E E E C C C 0 L s CO O E c C E C O O N -o 4-3 co L. o N CD O U 0 . C",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "SYSTEM PERFORMANC \u00c8",
"sec_num": null
},
{
"text": "Nine major processing steps were employed for identification of the data-classes involved in the MUC-6 Named Entity task ., as described in the following table .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Sequence of Processing Used in the Named Entity Tas k",
"sec_num": null
},
{
"text": "Step Descriptio n 1",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "No.",
"sec_num": null
},
{
"text": "Multi-Word Lookp Identify the data-class of multi-word phrases found in the knowledgebank 2",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "No.",
"sec_num": null
},
{
"text": "Find-Root Check to see if remaining words have a part-of-speech by direct look-up first and then by stripping plural suffixe s 3 Morphology Guess at unknown words' part-of-speech based on word suffixes . Add number (and gender, person) attributes where appropriate . 4",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "No.",
"sec_num": null
},
{
"text": "Single-word Places Identify all instances of single-word places (such as \"Chicago\") .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "No.",
"sec_num": null
},
{
"text": "-\u2022 5",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "No.",
"sec_num": null
},
{
"text": "Level 1 Data-Classes",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "No.",
"sec_num": null
},
{
"text": "Apply DXL rules for data-classes which are non-interacting and do no t require other data-classes for their identification (also parallelizeable) . These were : time, date, percent, money (some references to time involved single-word places, e .g., \"4 p .m. Chicago Time\").",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "No.",
"sec_num": null
},
{
"text": "Level 2 Data-Classes",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "6",
"sec_num": null
},
{
"text": "Apply DXL rules for more complex interacting rule-sets in the followin g order : placel, place2, orgl, org2, personl, person2, person3. The numeric suffix on the rule-sets indicates that data-class rules were clustered into groups of canonical forms that were increasingly complex, as indicated by the number-value . 7",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "6",
"sec_num": null
},
{
"text": "Last-Pass A set of DXL rules which checked the token stream for any remainin g tokens which could be part of place, org, or person . 8 Adjust A set of DXL rules which adjusted the start/stop positions of the target s to include or exclude punctuation according to MUC-6 rules. 9",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "6",
"sec_num": null
},
{
"text": "Tag Insert the appropriate SGML tags around the targets in the input text as calculated by the Adjust rules .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "6",
"sec_num": null
},
{
"text": "In developing the rules for a particular data-class, the following five-step strategy was employed: (1) a large number of examples of the data-class were collected, and these were clustered into group s having high internal similarity ; (2) a canonical pattern-sequence of obligatory and optional element s (plus signifying contexts) was developed for each cluster ; (3) DXL rules were developed for eac h canonical form ; (4) the rules were ordered in a sequence which causes those with the most number o f fixed patterns to be run before those with fewer, and in the case of a tie, with those accounting for th e largest number of instances to be run first ; and (5) the rules are broken into two groups with the simple r reliable ones grouped in one rule-set, e .g ., placel for the location rules, with the complex rules placed in a second rule-set, e .g ., place2 . Some experimentation was performed to determine the final grouping of the rule-sets into the Level 1 and Level 2 sets, but that indicated in the table led to the best results .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "6",
"sec_num": null
},
{
"text": "In the Last-Pass step, there were several rules which attempted to use context-inferences an d other heuristics to identify token sequences which were likely place, org, or person instances. One such was to see whether a promising token (a capitalized unaccounted-for one, for example) was an element of any of the instances of the three types of data-classes that had previously been identified or was a n acronym thereof. Thus, both \"Consuela Washington\" and \"Ms . Washington\" were recognized by straight-forward person rules (the first by one which looks for known first-names ; the second by one which uses prefix titles). The subsequent reference simply to \"Washington\" was correctly identified by the previously-seen by being a sub-string of, in this case, the closest prior reference, that of \"Ms . Washington\" .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "6",
"sec_num": null
},
{
"text": "\u2022 The late release of MUC-6 task information and materials in August precluded advance study of the complex scoring, recognition, and tagging criteria",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "6",
"sec_num": null
},
{
"text": "The primary method of obtaining training materials was to extract a very large number o f instances of each of the seven Named Entity data-classes from the provided test materials and use thes e files for development of the canonical groups and rules . An enormous amount of time was spent learnin g how to use DXL (and, as the rule writers did not know UNIX, learning the Linux system) . An equa l amount of time was spent in learning effective strategies for writing reliable rules --especially, learnin g to avoid the temptation of trying to recognize too many variations of data-class instances with a singl e rule . It is only in the past month, after the contest, that the grammar writers have learned the \"right\" leve l of ambition for writing a rule, testing it with the limited debugging capabilities, and revising it modestly .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Training",
"sec_num": null
},
{
"text": "We note that the Wall Street Journal style guide was very useful in reducing the number o f training examples needed.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Training",
"sec_num": null
},
{
"text": "We did not do well . It did not magically all come together at the last moment . Our three-week flurry of rule-writing simply didn ' t cope.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Performance on the October 5 MUC-6 Test s",
"sec_num": null
},
{
"text": "In addition to the developmental pressures noted above, the difficulties of learning to program i n a completely new language in a few weeks, and the agonies of understanding MUC identification an d tagging criteria, we also made the mistake of attempting a much too complex approach to identificatio n of the proper name classes : using a \"fuzzy logic\" approach in which capitalized words had fuzz y membership in the three person/location/organization classes which were gradually and simultaneously t o be resolved . It didn't work.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Performance on the October 5 MUC-6 Test s",
"sec_num": null
},
{
"text": "We started over, and performance now is very high . On a recent test involving over a hundre d test-cases per class, we logged on average 2% misses, 6% correctly identified but incorrectly tagged, and 0 false alarms for the four simpler classes of time, date, percentage, and money . For locations and organizations, the numbers are 4%, 9%, and 3 respectively ; the person rules are still in flux but will be i n this ballpark by the time of the conference .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Performance One-Month Later!",
"sec_num": null
},
{
"text": "Aside from the fuzzy-logic approach, it took us a long time (several weeks at least) to learn that , for pattern-based rules, simple is terribly much more effective than comprehensive . We also suffered from the fact that, since the language DXL has only just been developed it quite reasonably still has a fe w bugs ; these just happened to be difficult ones : e .g ., coding that worked perfectly well in the main pattern s of a rule LHS, did not do so in the left-context or in the pattern definitions (e .g ., \"@title_suf) ; the knowledge-bank could get corrupted just a bit without failing in the main ; etc. And we had endles s problems with the fact that the knowledge-bank is incomplete : the fact that it has so very many entrie s usually meant that problems were with our rules, not with the knowledge-bank, but there were many times when expected entries were simply not there or had been miscoded in some way . Finally, the absence of user function libraries to provide very high-level scotch-guarded functions for easy use in the rules, mean t that too often we had to use very low-level programming functions or drop into Scheme, neither a forte' o f the rule-writers .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Things That Didn't Wor k",
"sec_num": null
},
{
"text": "Concerning specific target difficulties, we perhaps had the most trouble with organizations . Single-word organizations or ones without a prefix or suffix title (e .g., \"Pilsbury\", \"Birds Eye\") required context-sensitive semantics to pull in . Names with commas in the middle, commonly law firms (e .g., L . F. Rotchieff, Unterberg, Towbin), were difficult because we used the commas to suggest phrase boundaries. And organizations with \"and\" (in contrast to \"&\") as part of their name were difficult t o discriminate from conjoined organizations (e .g ., \"Hollis and Pergamon Holdings, Ltd .\", where \"Hollis\" i s a reference to a prior-mentioned company) .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Things That Didn't Wor k",
"sec_num": null
},
{
"text": "Person targets were often difficult, but they tended to be the default case : we had already done our best with our second set of place and organization rules, and what remained was most likely a person . The high frequency presence of appositives with person-names provide one powerful source of semanti c context resolution.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Things That Didn't Wor k",
"sec_num": null
},
{
"text": "The knowledge-classification feature hierarchy works termendously well in supportin g identification and discrimination of data-classes (e .g ., people having the main branches of HUMENT-PERSON, while cities have the main branches of PLACE-CENTER-CITY, and signal words associate d with time references have the branches of MEAS-TIME-DATE-HOUR) . The DXL language, now tha t we understand it, is wonderfully powerful and flexible . And the brute-force look-up approach handles over 40% of our instance recognition and is easily extensible . We are also quite encouraged by the success of some of the LastPass stray-pickup rules based on semantic contexts and other heuristics .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Things That Worked Wel l",
"sec_num": null
},
{
"text": "The leading candidate for this prize is the user function library, to keep the rule-writer out of th e bowels of programming . But, of course, nominations for the library can only come with experience whic h is only now maturing .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Thing That Most Needs Reworkin g",
"sec_num": null
},
{
"text": "The outstanding lesson for this project as a result of the MUC is that the DX system will in fac t be able to meet its performance objectives in the near future, that the three-part design basis (of look-up, pattern-match, and semantically resolve) is sound .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "LESSONS LEARNE D",
"sec_num": null
},
{
"text": "Lesson two is : start early on the MUCs.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "LESSONS LEARNE D",
"sec_num": null
},
{
"text": "This would not actually be a good way of approaching recognition of person-name instances, but it does illustrate what additiona l attributes would be added as a result of consulting the knowledge-bank. Such a rule is only illustrative and would actually not be used because so many optional elements beg for over-generalizations .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [
{
"text": "The DX project team for the MUC-6 participation involved seven people, most becomin g involved recently . Team participation and responsibilities are shown in the following table . Roughly a third of the 28 person-month effort was devoted to design and implementation of the dataextraction language, DXL; another third went for overall system and knowledge-bank development ; and the last third was focused on development of general and MUC-specific data-class recognition rules usin g DXL . Not counting the peculiarities of MUC requirements for tagging the identified data-classes, no par t of this effort has been spent on non-reusable MUC-specific activities .",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Development Effort",
"sec_num": null
},
{
"text": "There were a number of factors which made the timing of the MUC-6 contest inconvenient relative to the external factors determining the pacing of development of the DX system :",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Period of Participatio",
"sec_num": null
},
{
"text": "The knowledge-bank, upon which all processing relies, was designed and populated only i n the August-September period; its implementation needed to be radically modified i n September to permit handling of massive numbers of entries, and it is still not yet fully reliable \u2022 The rule writers and knowledge-base/system administrator were effectively not availabl e until early Septembe r \u2022 The language, DXL, while possessing all the needed functionality, had limited high-level function libraries which are only being developed gradually with experience in rule-writin g \u2022 The system processing stages (esp. tokenization) are still under development, and the system still has quite limited debugging facilitie s",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "\u2022",
"sec_num": null
}
],
"bib_entries": {},
"ref_entries": {}
}
}