File size: 59,215 Bytes
6fa4bc9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 |
{
"paper_id": "A88-1011",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T02:04:06.402252Z"
},
"title": "TRIPHONE ANALYSIS: A COMBINED METHOD FOR THE CORRECTION OF ORTHOGRAPHICAL AND TYPOGRAPHICAL ERRORS",
"authors": [
{
"first": "Brigitte",
"middle": [],
"last": "Van Berkelt",
"suffix": "",
"affiliation": {},
"email": ""
},
{
"first": "Koenraad",
"middle": [],
"last": "De Smedt",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "University of Nijmegen",
"location": {
"addrLine": "Montessofilaan 3",
"postCode": "6525 HR",
"settlement": "Nijmegen",
"country": "The Netherlands"
}
},
"email": ""
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "Most existing systems for the correction of word level errors are oriented toward either typographical or orthographical errors. Triphone analysis is a new correction strategy which combines phonemic transcription with trigram analysis.",
"pdf_parse": {
"paper_id": "A88-1011",
"_pdf_hash": "",
"abstract": [
{
"text": "Most existing systems for the correction of word level errors are oriented toward either typographical or orthographical errors. Triphone analysis is a new correction strategy which combines phonemic transcription with trigram analysis.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "Any method for the correction of word level errors in written texts must be carefully tuned. On the one hand, the number of probable corrections should be maximized; on the other hand, the number of unlikely corrections should be minimized. In order to achieve these goals, the characteristics of specific error types must be exploited as much as possible. In this article we distinguish two major types of word level errors: orthographical errors and typographical errors. They have some clearly different characteristics.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Error types",
"sec_num": "1.1"
},
{
"text": "Orthographical errors are cognitive errors consisting of the substitution of a deviant spelling for a correct one when the author either simply doesn't know the correct spelling for a correct spelling, forgot it, or misconceived it. An important characteristic of orthographical errors is that they generally result in a string which is phonologically identical or very similar to the correct string (e.g. indicies instead of indices1). As a consequence, orthographical errors are dependent on the correspondence between spelling and pronunciation in a particular language. Another characteristic is that proper names, infrequent words and foreign words are particularly prone to orthographical errors.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Error types",
"sec_num": "1.1"
},
{
"text": "1 All examples of errors given in this article were actually found by the authors in texts written by native speakers of the language in question.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Error types",
"sec_num": "1.1"
},
{
"text": "hitting the wrong sequence of keys. Hence their characteristics depend on the use of a particular keyboard rather than on a particular language. Roughly eighty percent of these errors can be described as single deletions (e.g. continous) insertions (e.g. explaination), substitutions (e.g. anyboby) or transpositions (e.g. autoamtically) while the remaining twenty percent are complex errors (Peterson, 1980) . Some statistical facts about typographical errors are that word-initial errors are rare, and doubling and undoubling (e.g. succeeed, discusion) are common. In general, typographical errors do not lead to a string which is homophonous with the correct string.",
"cite_spans": [
{
"start": 392,
"end": 408,
"text": "(Peterson, 1980)",
"ref_id": "BIBREF6"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Typographical errors are motoric errors caused by",
"sec_num": null
},
{
"text": "Most of the correction methods currently in use in spelling checkers are biased toward the correction of typographical errors. We argue that this is not the fight thing to do. Even if orthographical errors are not as frequent as typographical errors, they are not to be neglected for a number of good reasons. First, orthographical errors are cognitive errors, so they are more persistent than typographical errors: proofreading by the author himself will often fail to lead to correction. Second, orthographical errors leave a worse impression on the reader than typographical errors. Third, the use of orthographical correction for standardization purposes (e.g. consistent use of either British or American spelling) is an important application appreciated by editors. In this context, our research pays special attention to Dutch, which has a preferred standard spelling but allows alternatives for a great many foreign words, e.g. architect (preferred) vs. architekt (allowed and commonly used in Dutch).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Typographical errors are motoric errors caused by",
"sec_num": null
},
{
"text": "Editors of books generally prefer a consistent use of the standard spelling.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Typographical errors are motoric errors caused by",
"sec_num": null
},
{
"text": "Finally, we would like to point out that methods for orthographical error correction can not only be applied in text processing, but also in database retrieval. In fact, our research was prompted partly by a project proposal for a user interface to an electronic encyclopedia. One or our experiments involving a lists of some five thousand worldwide geographical names (mainly in Dutch spelling, e.g. Noordkorea, Nieuwzeeland) has yielded very positive results. In this context, the correction of orthographical errors is obviously more important than the correction of typographical errors. Daelemans, Bakker & Schotel (1984) distinguish between two basic kinds of strategies: statistical and linguistic strategies. Statistical strategies are based on string comparison techniques, often augmented by specific biases using statistical characteristics of some error types, such as the fact that typographical errors do not frequently occur in the beginning of a word. Since these strategies do not exploit any specific linguistic knowledge, they will generally work better for typographical errors than for orthographical errors.",
"cite_spans": [
{
"start": 592,
"end": 626,
"text": "Daelemans, Bakker & Schotel (1984)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Typographical errors are motoric errors caused by",
"sec_num": null
},
{
"text": "Linguistic strategies exploit the fact that orthographical errors often result in homophonous strings (sound-alikes. e.g. consistancy and consistency). They normally involve some kind of phonemic transcription. Typographical errors which do not severely affect the pronunciation, such as doubling and undoubling, may be covered as well, but in general, linguistic strategies will do a poor job on all other typographical errors.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Correction strategies",
"sec_num": "1.2"
},
{
"text": "Because each type of strategy is oriented toward one class of errors only, what is needed in our opinion is a combined method for orthographical and typographical errors. Our research has explored one approach to this problem, namely, the combination of a linguistic strategy with a statistical one.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Correction strategies",
"sec_num": "1.2"
},
{
"text": "The remainder of this document is structured as follows. First we will discuss and criticize some existing statistical and linguistic correction methods. Then we will introduce triphone analysis. Finally we will report some results of an experiment with this method.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Correction strategies",
"sec_num": "1.2"
},
{
"text": "In Peterson's SPELL (Peterson, 1980) , all probable corrections are directly generated from an incorrect string by considering the four major single error types. The program first makes a list of all strings from which the incorrect string can be derived by a single deletion, insertion, substitution or transposition. This list is then matched against the dictionary: all strings occuring in both the list and the dictionary are considered probable corrections.",
"cite_spans": [
{
"start": 20,
"end": 36,
"text": "(Peterson, 1980)",
"ref_id": "BIBREF6"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Spell",
"sec_num": "2.1"
},
{
"text": "Although the number of derivations is relatively small for short strings, they often lead to several probable corrections because many of them will actually occur in the dictionary. For longer strings, many possible derivations are considered but most of those will be non-existent words.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Spell",
"sec_num": "2.1"
},
{
"text": "An advantage of SPELL with respect to all other methods is that short words can be corrected equally well as long ones. A disadvantage is that all complex errors and many orthographical errors fall outside the scope of SPELL.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Spell",
"sec_num": "2.1"
},
{
"text": "SPEEDCOP (Pollock & Zamora, 1984 ) uses a special technique for searching and comparing strings. In order to allow a certain measure of similarity, strings are converted into similarity keys which intentionally blur the characteristics of the original strings. The key of the misspelling is looked up in a list of keys for all dictionary entries. The keys found in the list within a certain distance of the target key are considered probable corrections.",
"cite_spans": [
{
"start": 9,
"end": 32,
"text": "(Pollock & Zamora, 1984",
"ref_id": "BIBREF7"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Speedcop",
"sec_num": "2.2"
},
{
"text": "The blurring of the similarity keys must be carefully finetuned. On the one hand, if too much information is lost, too many words collate to the same key. If, on the other hand, too much information is retained, the key will be too sensitive to alterations by misspellings. Two similarity keys are used in SPEEDCOP: a skeleton key and an omission key. These keys are carefully designed in order to partially preserve the characters in a string and their interrelationships. The information contained in the key is ordered according to some characteristics of typographical errors, e.g. the fact that word-initial errors are infrequent and that the sequence of consonants is often undisturbed.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Speedcop",
"sec_num": "2.2"
},
{
"text": "The skeleton key contains the first letter of a string, then the remaining consonants and finally the remaining vowels (in order, without duplicates). E.g. the skeleton key of information would be infrmtoa.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Speedcop",
"sec_num": "2.2"
},
{
"text": "The advantage of using this key is that some frequent error types such as doubling and undoubling of characters as well as transpositions involving one consonant and one vowel (except for an initial vowel) results in keys which are identical to the keys of the original strings.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Speedcop",
"sec_num": "2.2"
},
{
"text": "The most vulnerable aspect of the skeleton key is its dependence on the first few consonants. This turned out to be a problem, especially for omissions.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Speedcop",
"sec_num": "2.2"
},
{
"text": "Therefore, a second key, the omission key, was developed. According to Pollock & Zamora (1984) , consonants are omitted in the following declining or-der of frequency: RSTNLCHDPGMFBYWVZXQKJ. The omission key is construed by first putting the consonants in increasing order of omission frequency and adding the vowels in order of occurrence. E.g. the omission key for information isfmntrioa.",
"cite_spans": [
{
"start": 71,
"end": 94,
"text": "Pollock & Zamora (1984)",
"ref_id": "BIBREF7"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Speedcop",
"sec_num": "2.2"
},
{
"text": "SPEEDCOP exploits the statistical properties of typographical errors well, so it deals better with frequent kinds of typographical errors than with infrequent ones. Because of this emphasis on typographical errors, its performance on orthographical errors will be poor. A specific disadvantage is its dependence on the correctness of initial characters. Even when the omission key is used, word-initial errors involving e.g. j or k do not lead to an appropriate correction.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Speedcop",
"sec_num": "2.2"
},
{
"text": "Trigram analysis, as used in FUZZIE (De Heer, 1982) and ACUTE (Angell, 1983) , uses a more general similarity measure. The idea behind this method is that a word can be divided in a set of small overlapping substrings, called n-grams, which each carry some information about the identity of a word. When a misspelling has at least one undisturbed ngram, the correct spelling spelling can still be traced. The difference between FUZZIE and ACUTE is mainly in the criteria which are used to restrict the number of possible corrections. FUZZIE emphasizes frequency as a selection criterium whereas ACUTE also uses word length. Low frequency trigrams are assumed to have a higher identifying value than high frequency trigrams. In FUZZIE, only the correction candidates associated with the n least frequent trigrams, which are called selective trigrams, are considered. ACUTE offers the choice between giving low frequency trigrams a higher value and giving all trigrams the same value.",
"cite_spans": [
{
"start": 36,
"end": 51,
"text": "(De Heer, 1982)",
"ref_id": "BIBREF5"
},
{
"start": 62,
"end": 76,
"text": "(Angell, 1983)",
"ref_id": "BIBREF0"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Trigram analysis: Fuzzie and Acute",
"sec_num": "2.3"
},
{
"text": "Taking trigram frequency into account has advantages as well as disadvantages. On the one hand, there is a favorable distribution of trigrams in natural languages in the sense that there is a large number of low frequency trigrams. Also, the majority of words contain at least one selective trigram. On the other hand, typographical errors may yield very low frequency trigrams which inevitably get a high information value.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Trigram analysis: Fuzzie and Acute",
"sec_num": "2.3"
},
{
"text": "In general, trigram analysis works better for long words than for short ones, because a single error may disturb all or virtually all trigrams in a short word. Some advantages of this method are that the error position is not important and that complex errors (e.g. differenent), and, to a certain extent, orthographical errors, can often be corrected. A disadvantage which is specific to this method is that transpositions disturb more trigrams than other types of errors and will thus be more difficult to correct.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Trigram analysis: Fuzzie and Acute",
"sec_num": "2.3"
},
{
"text": "Trigram analysis lends itself well to extensions. By first selecting a large group of intermediate solutions, i.e. all words which share at least one selective trigram with the misspelling, there is a lot of room for other factors to decide which words will eventually be chosen as probable corrections. ACUTE for example uses word length as an important criterium.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Trigram analysis: Fuzzie and Acute",
"sec_num": "2.3"
},
{
"text": "The PF-474 chip is a special-purpose VLSI circuit designed for very fast comparison of a string with every entry in a dictionary (Yianilos, 1983) . It consists of a DMA controller for handling input from a data base (the dictionary), a proximity computer for computing the proximity (similarity) of two strings, and a ranker for ranking the 16 best solutions according to their proximity values.",
"cite_spans": [
{
"start": 129,
"end": 145,
"text": "(Yianilos, 1983)",
"ref_id": "BIBREF9"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "The PF-474 chip",
"sec_num": "2.4"
},
{
"text": "The proximity value (PV) of two strings is a function of the number of corresponding characters of both strings counted in forward and backward directions. It is basically expressed as the following ratio:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The PF-474 chip",
"sec_num": "2.4"
},
{
"text": "2*(ABforwar d + ABbackward) PV-AAforw ard+AAbackward+BB forward +BBbackward",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The PF-474 chip",
"sec_num": "2.4"
},
{
"text": "This value can be influenced by manipulating the parameters weight, bias and compensation. The parameter weight makes some characters more important than others. This parameter can e.g. be manipulated to reflect the fact that consonants carry more information than vowels. The parameter bias may correct the weight of a character in either word-initial or word-final position. The parameter compensation determines the importance of an occurrence of a certain character within the word. By using a high compensation/weight ratio, for example, substitution of characters will be less severe than omission. One may force two characters to be considered identical by equalizing their compensation and weight values.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The PF-474 chip",
"sec_num": "2.4"
},
{
"text": "An advantage of the PF-474 chip, apart from its high speed, is that it is a general string comparison technique which is not biased to a particular kind of errors. By carefully manipulating the parameters, many orthographical errors may be corrected in addition to typographical errors.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The PF-474 chip",
"sec_num": "2.4"
},
{
"text": "SPELL THERAPIST (Van Berkel, 1986 ) is a linguistic method for the correction of orthographical errors. The misspelling is transcribed into a phonological code which is subsequently looked up in a dictionary consisting of phonological codes with associated spellings. The phonemic transcription, based on the GRAFON system (Daelemans, 1987) , is performed in three steps. First the character string is split into syllables. Then a rule-based system converts each syllable into a phoneme string by means of transliteration rules. These syllabic phoneme strings are further processed by phonological rules which take the surrounding syllable context into account and are finally concatenated.",
"cite_spans": [
{
"start": 6,
"end": 33,
"text": "THERAPIST (Van Berkel, 1986",
"ref_id": null
},
{
"start": 323,
"end": 340,
"text": "(Daelemans, 1987)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Spell Therapist",
"sec_num": "2.5"
},
{
"text": "The transliteration rules in SPELL THERAPIST are grouped into three ordered lists: one for the onset of the syllable, one for the nucleus, and one for the coda. Each rule consists of a graphemic selection pattern, a graphemic conversion pattern, and a phoneme string. The following rules are some examples for Dutch onsets:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Spell Therapist",
"sec_num": "2.5"
},
{
"text": "((sc(-h i e y)) c /k/) (( qu ) qu (/k//kw/)) (( a ( consonantp )) a/a/)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Spell Therapist",
"sec_num": "2.5"
},
{
"text": "The first rule indicates that in a graphemic pattern consisting of sc which is not followed by either h, i, e or y, the grapheme c is to be transcribed as the phoneme/k/.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Spell Therapist",
"sec_num": "2.5"
},
{
"text": "The transcription proceeds as follows. The onset of a syllable is matched with the graphemic selection patterns in the onset rule list. The first rule which matches is selected. Then the characters which match with the conversion pattern are converted into the phoneme string. The same procedure is then performed for the nucleus and coda of the syllable.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Spell Therapist",
"sec_num": "2.5"
},
{
"text": "The result of the transcription is then processed by means of phonological rules, which convert a sequence of phonemes into another sequence of phonemes in a certain phonological context on the level of the word. An example for Dutch is the cluster reduction rule which deletes a/t/in certain consonant clusters:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Spell Therapist",
"sec_num": "2.5"
},
{
"text": "((( obstruent-p ) /t/ ( obstruent-p )) /t/ //)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Spell Therapist",
"sec_num": "2.5"
},
{
"text": "Such rules account for much of the power of SPELL THERAPIST because many homophonous orthographic errors seem to be related to rules such as assimilation (e.g. inplementation) or cluster reduction and degemination (e.g. Dutch kunstof instead of kunststo]).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Spell Therapist",
"sec_num": "2.5"
},
{
"text": "This method is further enhanced by the following refinements. First, a spelling may be transcribed into more than one phonological code in order to account for possible pronunciation variants, especially those due to several possible stress patterns. Second, the phonological code itself is designed to intentionally blur some finer phonological distinctions. E.g. in order to account for the fact that short vowels in unstressed syllables are prone to misspellings (e.g. optomization, incoded) such vowels are always reduced to a schwa /3/. As a result, misspellings of this type will collocate.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Spell Therapist",
"sec_num": "2.5"
},
{
"text": "It is clear that this method is suited only for errors which result in completely homophonous spellings (e.g. issuing, inplementation). A somewhat less stringent similarity measure is created by using a coarse phonological coding, as mentioned above. Still, this method is not suitable for most typographical errors. Moreover, orthographical errors involving 'hard' phonological differences (e.g. managable, recommand) fail to lead to correction.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Spell Therapist",
"sec_num": "2.5"
},
{
"text": "Of the methods described in the previous chapter, no single method sufficiently covers the whole spectrum of errors. Because each method has its strengths and weaknesses, it is advantageous to combine two methods which supplement each other. Because orthographical errors are the most difficult and persistent, we chose to take a linguistic method as a starting point and added another method to cover its weaknesses. SPELL THERAPIST has two weak points. First, most typographical errors cannot be corrected. Second, even though the phonological codes are somewhat blurred, at least one possible transcription of the misspelling must match exactly with the phonological code of the intended word.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Combining methods",
"sec_num": "3.1"
},
{
"text": "A possible solution to both problems consists in applying a general string comparison technique to phonological codes rather than spellings. We decided to combine SPELL THERAPIST with trigram analysis by using sequences of three phonemes instead of three characters. We call such a sequence a triphone and the new strategy triphone analysis.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Combining methods",
"sec_num": "3.1"
},
{
"text": "Triphone analysis is a fast and efficient method for correcting orthographical and typographical errors. When carefully implemented, it is not significantly slower than trigram analysis. The new method uses only one dictionary in the form of an inverted file of triphones. Such a file is created by first computing phonological variants for each word, then splitting each code into triphones, and finally adding backpointers from each triphone in the file to each spelling in which it occurs. Also, a frequency value is associated with each triphone.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Trlphone analysis",
"sec_num": "3.2"
},
{
"text": "The way this inverted file is used during correction is virtually the same as in FUZZIE, except that f'trst all phonological variants of the misspelling have to be generated. The grapheme-to-phoneme conversion is similar to that of SPELL THERAPIST, except that the phonological code is made even coarser by means of various simplifications., e.g. by removing the distinction between tense and lax vowels and by not applying certain phonological rules.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Trlphone analysis",
"sec_num": "3.2"
},
{
"text": "The easiest way to select probable corrections from an inverted file is the method used by FUZZIE, because the similarity measure used by ACUTE requires that the number of triphones in the possible correction be known in advance. The problem with this requirement is that phonological variants may have different string lengths and hence a varying number of triphones.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Trlphone analysis",
"sec_num": "3.2"
},
{
"text": "Using the FUZZIE method, each phonological variant may select probable corrections by means of the following steps:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Trlphone analysis",
"sec_num": "3.2"
},
{
"text": "1. The phonological code is split into triphones.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Trlphone analysis",
"sec_num": "3.2"
},
{
"text": "depending on its frequency. The sum of all values is I.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Each triphone receives an information value",
"sec_num": "2."
},
{
"text": "3. The selective triphones (those with a frequency below a certain preset value) are looked up in the inverted file.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Each triphone receives an information value",
"sec_num": "2."
},
{
"text": "4. For all correction candidates found in this way, the similarity with the misspelling is determined by computing the sum of the information values of all triphones shared between the candidate and the misspelling.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Each triphone receives an information value",
"sec_num": "2."
},
{
"text": "If a certain candidate for correction is found by more than one phonological variant, only the highest information value for that candidate is retained. After candidates have been selected for all variants, they are ordered by their similarity values. A possible extension could be realized by also taking into account the difference in string length between the misspelling and each candidate.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Each triphone receives an information value",
"sec_num": "2."
},
{
"text": "Because processing time increases with each phonological variant, it is important to reduce the number of variants as much as possible. A considerable reduction is achieved by not generating a separate variant for each possible stress pattern. The resulting inaccuracy is largely compensated by the fact that a perfect match is no longer required by the new method.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Each triphone receives an information value",
"sec_num": "2."
},
{
"text": "Although this method yields very satisfactory results for both orthographical and typographical errors and for combinations of them, it does have some shortcomings for typographical errors in short words. One problem is that certain deletions cause two surrounding letters to be contracted into very different phonemes. Consider the deletion of the r in very: the pronunciation of the vowels in the resulting spelling, vey, changes substantially. Counting one surrounding space, the misspelling does not have a single triphone in common with the original and so it cannot be corrected.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Each triphone receives an information value",
"sec_num": "2."
},
{
"text": "A second problem is that a character (or character cluster) leading to several possible phonemes carries more information than a character leading to a single phoneme. Consequently, an error affecting such a character disturbs more triphones.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Each triphone receives an information value",
"sec_num": "2."
},
{
"text": "The triphone analysis method presented here has been implemented on a Symbolics LISP Machine and on an APOLLO workstation running Common LISP. After the programs had been completed, we decided to test the new method and compare its qualitative performance with that of the other methods.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "An experiment",
"sec_num": "3.3"
},
{
"text": "For a first, preliminary test we chose our domain carefully. The task domain had to be very error-prone, especially with respect to orthographical errors, so that we could elicit errors from human subjects under controlled circumstances. Given these requirements, we decided to choose Dutch surnames as the task domain. In Dutch, many surnames have very different spellings. For example, there are 32 different names with the same pronunciation as Theyse, and even 124 ways to spell Craeybeckx! When such a name is written in a dictation task (e.g. during a telephone conversation) the chance of the right spelling being chosen is quite small.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "An experiment",
"sec_num": "3.3"
},
{
"text": "For our experiment, we recorded deviant spellings of Dutch surnames generated by native speakers of Dutch in a writing-to-dictation task. A series of 123 Dutch surnames was randomly chosen from a telephone directory. The names were dictated to 10 subjects via a cassette tape recording. A comparison of the subjects' spelling with the intended spellings showed that on the average, subjects wrote down 37.6% of the names in a deviant way. The set of 463 tokens of misspellings contained 188 different types, which were subsequently given as input to implementations of each of the methods 2. The dictionary consisted of 254 names (the 123 names mentioned above plus I31 additional Dutch surnames randomly selected from a different source). The results of the correction are presented in Tables 1 and 2. The PF-474 method was simulated in software instead of using the special hardware.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "An experiment",
"sec_num": "3.3"
},
{
"text": "The experiment was designed in order to minimize typographical errors and to maximize orthographical errors. Hence it is not surprising that SPELL and SPEEDCOP, which are very much dependent on the characteristics of typographical errors, do very poorly. What is perhaps most surprising is that SPELL THERAPIST, a method primarily aiming at the correction of orthographical errors, shows worse results than FUZZIE, ACUTE and the PF-474 method, which are general string comparison methods. The reason is that a certain number of orthographical errors turned out to involve real phonological differences. These were probably caused by mishearings rather than misspellings. Poor sound quality of the cassette recorder and dialectal differences between speaker and hearer are possible causes. As expected, triphone analysis yielded the best results: not a single misspelling could not be corrected, and only about one out of twenty failed to be returned as the most likely correction.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "3.4"
},
{
"text": "We have demonstrated that an integration of complementary correction methods performs better than single methods. With respect to orthographical errors, triphone analysis performs better than either grapheme-to-phoneme conversion or trigram analysis alone. Its capacity to correct typographical errors is still to be evaluated, but it is already clear that it will be better than that of SPELL THERAPIST although somewhat worse than trigram analysis in those cases where a typographical error drastically alters the pronunciation. In practice, however, one always finds both kinds of errors. Therefore, it would be interesting to compare the various methods in actual use.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "CONCLUSION",
"sec_num": "4."
},
{
"text": "Future research will go into a number of variants on the basic ideas presented here. From a linguistic point of view, it is possible to make the phonological matching less stringent. One way to do this is to use a comparison at the level of phonological features rather than phonemes. However, greater emphasis on orthographical errors may deteriorate performance on the correction of typing errors.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "CONCLUSION",
"sec_num": "4."
},
{
"text": "An area of current research is the extension of triphone analysis toward the correction of compounds. In languages like Dutch and German, new compounds such as taaltechnologie (language technology) are normally written as one word. Correction of errors in such compounds is difficult because the constituting words should be corrected separately but there is no 82, easy way to find the right segmentation. We have developed some heuristics to solve this problem.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "CONCLUSION",
"sec_num": "4."
},
{
"text": "Of course, other combinations of methods are possible. One possibility which looks promising is to combine phonemic transcription with the PF-474 chip. Although triphone analysis is fairly fast, use of the PF-474 chip might further increase the speed. For the correction of large quantities of word material, speed is an essential factor. However, it should be kept in mind that there is a linear correlation between the size of the dictionary and the required processing time, and that the correlation curve is steeper for the PF-474 chip than for triphone analysis. This means that triphone analysis will still be faster for very large dictionaries.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "CONCLUSION",
"sec_num": "4."
},
{
"text": "With an eye to commercial applications, TNO-ITI is extending the basic method with data compression techniques and an improved formalism for graphemeto-phoneme conversion.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "CONCLUSION",
"sec_num": "4."
}
],
"back_matter": [
{
"text": "A prototype of triphone analysis was implemented at the Dept. of Psychology of the University of Nijmegen under ESPRIT project OS-82. Parts of the experiment and the port to the APOLLO were carried out at TNO-ITI, which also developed FUZZIE.We are indebted to Prof. Dr. Gerard Kempen (University of Nijmegen) and to Adriaan van Paassen (TNO-ITI) for their stimulation of the research and for the helpful comments, and to Hil Weber for typing the paper.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "ACKNOWLEDGEMENTS",
"sec_num": null
}
],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "Automatic spelling correction using a trigram similarity measure",
"authors": [
{
"first": "R",
"middle": [
"C"
],
"last": "Angell",
"suffix": ""
},
{
"first": "G",
"middle": [
"E"
],
"last": "Freund",
"suffix": ""
},
{
"first": "P",
"middle": [],
"last": "Willett",
"suffix": ""
}
],
"year": 1983,
"venue": "Information Processing & Management",
"volume": "19",
"issue": "",
"pages": "255--261",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Angell, R.C., Freund, G.E. & Willett, P. (1983) Automatic spelling correction using a trigram similarity measure. Information Processing & Management, 19,255-261.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Subjective trial of the performance of the information trace method",
"authors": [
{
"first": "L",
"middle": [
"G"
],
"last": "Barendregt",
"suffix": ""
},
{
"first": "C",
"middle": [
"A"
],
"last": "Benschop",
"suffix": ""
},
{
"first": "T",
"middle": [],
"last": "De Heer",
"suffix": ""
}
],
"year": 1985,
"venue": "Information Processing & Management",
"volume": "21",
"issue": "",
"pages": "103--111",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Barendregt, L.G., Benschop, C.A. & De Heer, T. (1985) Subjective trial of the performance of the information trace method. Information Processing & Management, 21,103-111.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "Studies in Language technology",
"authors": [
{
"first": "W",
"middle": [],
"last": "Daelemans",
"suffix": ""
}
],
"year": 1987,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Daelemans, W. (1987) Studies in Language technology. Ph.D. Dissertation, Linguistics Dept., University of Leuven.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Automatische detectie en correctie van spelfouten. lnformatie",
"authors": [
{
"first": "W",
"middle": [],
"last": "Daelemans",
"suffix": ""
},
{
"first": "D",
"middle": [],
"last": "Bakker",
"suffix": ""
},
{
"first": "H",
"middle": [],
"last": "Schotel",
"suffix": ""
}
],
"year": 1984,
"venue": "",
"volume": "26",
"issue": "",
"pages": "949--1024",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Daelemans, W., Bakker, D. & Schotel, H. (1984) Automatische detectie en correctie van spelfouten. lnformatie, 26, 949-1024.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "A technique for computer detection and correction of spelling errors",
"authors": [
{
"first": "F",
"middle": [
"J"
],
"last": "Damerau",
"suffix": ""
}
],
"year": 1964,
"venue": "CACM",
"volume": "7",
"issue": "",
"pages": "171--177",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Damerau, F.J. (1964) A technique for computer detection and correction of spelling errors. CACM, 7, 171-177.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "The application of the concept of homeosemy to natural language information retrieval",
"authors": [
{
"first": "T",
"middle": [],
"last": "De Heer",
"suffix": ""
}
],
"year": 1982,
"venue": "Information Processing & Management",
"volume": "18",
"issue": "",
"pages": "229--236",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "De Heer, T. (1982) The application of the concept of homeosemy to natural language information retrieval. Information Processing & Management, 18, 229-236.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Computer programs for detecting and correcting spelling errors",
"authors": [
{
"first": "J",
"middle": [
"L"
],
"last": "Peterson",
"suffix": ""
}
],
"year": 1980,
"venue": "CACM",
"volume": "23",
"issue": "",
"pages": "676--687",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Peterson, J.L. (1980) Computer programs for detecting and correcting spelling errors. CACM, 23,676-687.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "Automatic spelling correction in scientific and scholarly text",
"authors": [
{
"first": "J",
"middle": [
"J"
],
"last": "Pollock",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Zamora",
"suffix": ""
}
],
"year": 1984,
"venue": "CACM",
"volume": "27",
"issue": "",
"pages": "358--368",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Pollock, J.J. & Zamora, A. (1984) Automatic spelling correction in scientific and scholarly text. CACM, 27, 358-368.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "SPELTERAPUIT: een algoritme voor spel-en typefoutencorrectie gebaseerd op grafeem-foneemomzetting",
"authors": [
{
"first": "B",
"middle": [],
"last": "Van Berkel",
"suffix": ""
}
],
"year": 1986,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Van Berkel, B. (1986) SPELTERAPUIT: een algoritme voor spel-en typefoutencorrectie gebaseerd op grafeem-foneemomzetting. Master's thesis, Dept. of Psychology, University of Nijmegen.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "A dedicated comparator matches symbol strings fast and intelligently. Electronics",
"authors": [
{
"first": "P",
"middle": [
"N"
],
"last": "Yianilos",
"suffix": ""
}
],
"year": 1983,
"venue": "",
"volume": "",
"issue": "",
"pages": "113--117",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Yianilos, P.N. (1983) A dedicated comparator matches symbol strings fast and intelligently. Electronics, December 1983, 113-117.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"text": "For natural languages, trigrams seem to have the most suitable length. E.g., counting one surrounding space, the word trigram is represented by the trigrams #tr, tri, rig, igr, gra, ram, and am#. B/grams are in general too short to contain any useful identifying information while tetragrams and larger n-grams are already close to average word length. Correction using trigrams proceeds as follows. The trigrams in a misspelling are looked up in an inverted file consisting of all trigrams extracted from the dictionary. With each trigram in this inverted file, a list of all words containing the trigram is associated. The words retrieved by means of the trigrams in the misspelling are probable corrections.",
"uris": null,
"type_str": "figure",
"num": null
},
"TABREF0": {
"content": "<table><tr><td/><td/><td/><td>The</td></tr><tr><td colspan=\"4\">numbers refer to percentages of recognized</td></tr><tr><td colspan=\"4\">(first, second or third choice) or not</td></tr><tr><td colspan=\"3\">recognized surnames (n = 188).</td><td/></tr><tr><td/><td colspan=\"2\">Ist choice 2nd or 3rd</td><td>not found</td></tr><tr><td>SPELL</td><td>58.5</td><td>1.1</td><td>40.4</td></tr><tr><td>SPEEDCOP</td><td>53.7</td><td>1.1</td><td>45.2</td></tr><tr><td>FUT_7_,IE</td><td>86.2</td><td>9.6</td><td>4.2</td></tr><tr><td>ACUTE</td><td>89.9</td><td>6.9</td><td>3.2</td></tr><tr><td>PF-474</td><td>84.0</td><td>14.9</td><td>1.1</td></tr><tr><td>SPELL</td><td/><td/><td/></tr><tr><td colspan=\"2\">THERAPIST 86.2</td><td>1.1</td><td>12.8</td></tr><tr><td>TRIPHONE</td><td/><td/><td/></tr><tr><td>ANALYSIS</td><td>94.1</td><td>5.9</td><td>0.0</td></tr><tr><td colspan=\"4\">Table 2. Results of the evaluation study. The</td></tr><tr><td colspan=\"4\">numbers refer to percentages of recognized</td></tr><tr><td colspan=\"4\">(first, second or third choice) or not</td></tr><tr><td colspan=\"4\">recognized surnames multiplied by their</td></tr><tr><td colspan=\"2\">frequencies (n = 463).</td><td/><td/></tr><tr><td/><td colspan=\"2\">1st choice 2nd or 3rd</td><td>not found</td></tr><tr><td>SPELL</td><td>63.7</td><td>2.2</td><td>34.1</td></tr><tr><td>SPEEDCOP</td><td>55.7</td><td>2.2</td><td>42.1</td></tr><tr><td>FUZZIE</td><td>87.7</td><td>8.4</td><td>3.9</td></tr><tr><td>ACUTE</td><td>90.3</td><td>6.7</td><td>3.0</td></tr><tr><td>PF-474</td><td>85.5</td><td>14.1</td><td>0.4</td></tr><tr><td>SPELL</td><td/><td/><td/></tr><tr><td>THERAPIST</td><td>90.5</td><td>2.2</td><td>7.3</td></tr><tr><td>TRIPHONE</td><td/><td/><td/></tr><tr><td>ANALYSIS</td><td>95.2</td><td>4.8</td><td>0.0</td></tr><tr><td>2</td><td/><td/><td/></tr></table>",
"type_str": "table",
"num": null,
"text": "Results of the evaluation study.",
"html": null
}
}
}
} |