File size: 64,457 Bytes
6fa4bc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
{
    "paper_id": "A92-1004",
    "header": {
        "generated_with": "S2ORC 1.0.0",
        "date_generated": "2023-01-19T02:03:50.705006Z"
    },
    "title": "A PARSER FOR REAL-TIME SPEECH SYNTHESIS OF CONVERSATIONAL TEXTS",
    "authors": [
        {
            "first": "Joan",
            "middle": [],
            "last": "Bachenko",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "AT&T Bell Laboratories Murray Hill",
                "location": {
                    "postCode": "07974",
                    "region": "NJ"
                }
            },
            "email": ""
        },
        {
            "first": "Jeffrey",
            "middle": [],
            "last": "Daugherty",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "AT&T Bell Laboratories Murray Hill",
                "location": {
                    "postCode": "07974",
                    "region": "NJ"
                }
            },
            "email": ""
        },
        {
            "first": "Eileen",
            "middle": [],
            "last": "Fitzpatrick",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "AT&T Bell Laboratories Murray Hill",
                "location": {
                    "postCode": "07974",
                    "region": "NJ"
                }
            },
            "email": ""
        }
    ],
    "year": "",
    "venue": null,
    "identifiers": {},
    "abstract": "In this paper, we concern ourselves with an application of text-to-speech for speech-impaired, deaf, and hard of hearing people. The application is unusual because it requires real-time synthesis of unedited, spontaneously generated conversational texts transmitted via a Telecommunications Device for the Deaf (TDD). We describe a parser that we have implemented as a front end for a version of the Bell Laboratories text-to-speech synthesizer (Olive and Liberman 1985). The parser prepares TDD texts for synthesis by (a) performing lexical regularization of abbreviations and some non-standard forms, and (b) identifying prosodic phrase boundaries. Rules for identifying phrase boundaries are derived from the prosodic phrase grammar described in Bachenko and Fitzpatrick (1990). Following the parent analysis, these rules use a mix of syntactic and phonological factors to identify phrase boundaries but, unlike the parent system, they forgo building any hierarchical structure in order to bypass the need for a stacking mechanism; this permits the system to operate in near real time. As a component of the text-to-speech system, the parser has undergone rigorous testing during a successful three-month field trial at an AT&T telecommunications center in California. In addition, laboratory evaluations indicate that the parser's performance compares favorably with human judgments about phrasing.",
    "pdf_parse": {
        "paper_id": "A92-1004",
        "_pdf_hash": "",
        "abstract": [
            {
                "text": "In this paper, we concern ourselves with an application of text-to-speech for speech-impaired, deaf, and hard of hearing people. The application is unusual because it requires real-time synthesis of unedited, spontaneously generated conversational texts transmitted via a Telecommunications Device for the Deaf (TDD). We describe a parser that we have implemented as a front end for a version of the Bell Laboratories text-to-speech synthesizer (Olive and Liberman 1985). The parser prepares TDD texts for synthesis by (a) performing lexical regularization of abbreviations and some non-standard forms, and (b) identifying prosodic phrase boundaries. Rules for identifying phrase boundaries are derived from the prosodic phrase grammar described in Bachenko and Fitzpatrick (1990). Following the parent analysis, these rules use a mix of syntactic and phonological factors to identify phrase boundaries but, unlike the parent system, they forgo building any hierarchical structure in order to bypass the need for a stacking mechanism; this permits the system to operate in near real time. As a component of the text-to-speech system, the parser has undergone rigorous testing during a successful three-month field trial at an AT&T telecommunications center in California. In addition, laboratory evaluations indicate that the parser's performance compares favorably with human judgments about phrasing.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Abstract",
                "sec_num": null
            }
        ],
        "body_text": [
            {
                "text": "Text-to-speech researchers and developers tend to assume that applications of their technology will focus on edited text, either \"canned\" material such as name and address lists, or free text like the AP newswire. There has been much effort aimed at preparing text-to-speech for applications such as caller identification and newsreading services, in which texts are generally proofed and the primary challenges come from issues of name pronunciation, intonation contouring, etc. In this paper, we concern ourselves with an application of text-to-speech for speech-impaired, deaf, ~md hard of hearing people. The application is unusual because it requires text-to-speech synthesis of unedited, spontaneously generated conversational text. Moreover the synthesis must occur in near real time as the user is typing.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "INTRODUCTION",
                "sec_num": "1."
            },
            {
                "text": "We will describe a parser that prepares conversational texts for synthesis by first performing lexical regularization of nonstandard forms and then identifying prosodic phrase boundaries. The parser is derived from the prosodic phrase system presented in Bachenko and Fitzpatrick (1990) and has been implemented as the front end of a version of the Bell Laboratories text-to-speech synthesizer (Olive and Liberman 1985) . As a component of the text-to-speech system, the parser has undergone rigorous testing during a successful three-month field trial at an AT&T telecommunications center in California. In addition, laboratory evaluations indicate that the parser's performance compares favorably with human judgments about phrasing. In Section 2 of the paper we describe the application and the texts. Section 3 provides a technical description of the parser and Section 4 discusses evaluation of the parser's performance.",
                "cite_spans": [
                    {
                        "start": 255,
                        "end": 286,
                        "text": "Bachenko and Fitzpatrick (1990)",
                        "ref_id": null
                    },
                    {
                        "start": 394,
                        "end": 419,
                        "text": "(Olive and Liberman 1985)",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "INTRODUCTION",
                "sec_num": "1."
            },
            {
                "text": "Users of Telecommunications Devices for the Deaf (TDD's) can communicate with voice telephone users via services such as AT&T's Telecommunications Relay Service (TRS). During a TRS call, special operators read incoming TDD text to the voice telephone user and then type that person's spoken responses back to the TDD user, this makes for a three-way interaction in which the special operator is performing both text-to-speech and speech-to-text conversion. Text-to-speech synthesis I. AT&T Bell Laboratories, Naperville, Illinois.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "THE APPLICATION",
                "sec_num": "2."
            },
            {
                "text": "Expected texts (e.g. AP newswire) ................................................................... 1% errors standard upper and lower st. English case conventions dialect 5% errors little or single case only written none language of the deaf Figure I : TDD vs. Expected Text Input makes it possible to automate part of this arrangement by reading the TDD text over the telephone to the voice user. The synthesizer thus replaces an operator on the TDD half of the conversation, providing increased privacy and control to the TDD user and, presumably, cost savings to the provider of the service.",
                "cite_spans": [
                    {
                        "start": 34,
                        "end": 101,
                        "text": "...................................................................",
                        "ref_id": null
                    }
                ],
                "ref_spans": [
                    {
                        "start": 245,
                        "end": 253,
                        "text": "Figure I",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "THE APPLICATION",
                "sec_num": "2."
            },
            {
                "text": "Spelling Punctuation Case Syntax",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "TDD texts",
                "sec_num": null
            },
            {
                "text": "TDD texts present unusual challenges for text-tospeech. Except in laboratory experiments, large scale applications of text-to-speech have tended to focus on name pronunciation and \"canned text\" such as catalogue orders. To the best of our knowledge, the TRS text-tospeech field trial in California represents the first large scale attempt to use speech synthesis on spontaneously generated conversational texts, and also the first to use this technology on texts that are orthographically and linguistically non-standard. Unlike the written material that most text-to-speech systems are tested on, e.g. the AP newswire, TDD texts observe few of the writing conventions of English. All text is in upper case, and punctuation, even at major sentence boundaries, rarely occurs; spelling and typographical errors complicate the picture even further (Tsao 1990; Kukich, 1992) . In addition, nearly all texts employ special abbreviations and lingo, e.g., CU stands for see you, GA is the message terminator go ahead. The following example illustrates a typical TDD text: OH SURE PLS CAIJ. ME ANYTI/V[E AFTER SAT MORNING AND I WILL GIVE U THE NAMES AND PHONE NOS OK QGA (Oh sure, please call me anytime after Saturday morning and I will give you the names and phone numbers. OK? Go ahead.) Finally, many texts are written in a variety of English that departs from expected lexical and syntactic patterns of the standard dialect (Charrow 1974) . For example, WHEN DO I WIIJ. CAIJ. BACK U Q GA is a short TDD text that we believe most native speakers of English would recognize as When should I call you back? Go ahead. The (attested) example below is less clear, but interpretable: I WISH THAT DAY I COULD LIKE TO",
                "cite_spans": [
                    {
                        "start": 845,
                        "end": 856,
                        "text": "(Tsao 1990;",
                        "ref_id": "BIBREF10"
                    },
                    {
                        "start": 857,
                        "end": 870,
                        "text": "Kukich, 1992)",
                        "ref_id": "BIBREF5"
                    },
                    {
                        "start": 1268,
                        "end": 1282,
                        "text": "OK? Go ahead.)",
                        "ref_id": null
                    },
                    {
                        "start": 1421,
                        "end": 1435,
                        "text": "(Charrow 1974)",
                        "ref_id": "BIBREF2"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "TDD texts",
                "sec_num": null
            },
            {
                "text": "Syntactic variation in such texts is systematic and consistent (Bacbenko 1989 , Charrow 1974 . Although a complete account has yet to be formulated, Suri (1991) reports that aspects of the variation may be explained by the influence of a native language--ASL--on a second language--English. Figure 1 above summarizes the points about TDD texts. Spelling error estimates come from Kukich (1992) and Tsao (1990) .",
                "cite_spans": [
                    {
                        "start": 63,
                        "end": 77,
                        "text": "(Bacbenko 1989",
                        "ref_id": null
                    },
                    {
                        "start": 78,
                        "end": 92,
                        "text": ", Charrow 1974",
                        "ref_id": "BIBREF2"
                    },
                    {
                        "start": 149,
                        "end": 160,
                        "text": "Suri (1991)",
                        "ref_id": "BIBREF9"
                    },
                    {
                        "start": 380,
                        "end": 393,
                        "text": "Kukich (1992)",
                        "ref_id": "BIBREF5"
                    },
                    {
                        "start": 398,
                        "end": 409,
                        "text": "Tsao (1990)",
                        "ref_id": "BIBREF10"
                    }
                ],
                "ref_spans": [
                    {
                        "start": 291,
                        "end": 299,
                        "text": "Figure 1",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "MEETING DIFFERENT PEOPLE WHO DOES THIS JOB AND THE WAY I WANT TO SEE HOW THEY DO IT LIKE THAT BUT THIS PLACES WAS FROM SAN FRANCISCO I GUESS",
                "sec_num": null
            },
            {
                "text": "Timing creates an additional obstacle since we expect TRS text-to-speech to synthesize the text while it is being typed, much as an operator would read it at the TRS center. How to chunk the incoming text now becomes a critical question. Word by word synthesis, where the listener hears a pause after each word, is the easiest approach but one that many people find nervewracking. N-word synthesis, where the listener hears a pause after some arbitrary number of words, is nearly as simple but runs the risk of creating unacceptably high levels of ambiguity and, for long texts, may be as irritating as single-word synthesis. Our solution was to build a TDD parser that uses linguistic roles to break up the speech into short, natural-sounding phrases. With partial buffering of incoming text, the parser is able to work in near real time as well as to perform lexical regularization of abbreviations and a small number of non-standard forms.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "MEETING DIFFERENT PEOPLE WHO DOES THIS JOB AND THE WAY I WANT TO SEE HOW THEY DO IT LIKE THAT BUT THIS PLACES WAS FROM SAN FRANCISCO I GUESS",
                "sec_num": null
            },
            {
                "text": "In constructing the parser, our goal was to come up with a system that (a) substitutes non-standard and abbreviated items with standard, pronounceable words, and (b) produces the most plausible phrasing with the simplest possible mechanism. Extensive data collection has been the key to success in regularizing lexical material, e.g. the conversion of fwy (pronounced \"f-wee\") to freeway. Phrasing is accomplished by a collection of rules derived from the prosodic phrase grammar of Bacbenko and Fitzpatrick (1990) , with some important modifications. The most radical of these is that the TDD phrasing rules build no hierarchical structure. Instead they rely on string adjacency, part of speech, word subclass and length to make inferences about possible syntactic constituency and to create enough prosodic cohesion to determine the location of phrase boundaries.",
                "cite_spans": [
                    {
                        "start": 483,
                        "end": 514,
                        "text": "Bacbenko and Fitzpatrick (1990)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "PARSER STRUCTURE AND RULES",
                "sec_num": "3.1."
            },
            {
                "text": "The parser works deterministicaUy (Marcus 1980 , Hindle 1983 ). It uses a small three element buffer that can contain either words or structures; once a lexical or prosodic structure is built it cannot be undone. As TDD text is typed, incoming words are collected in the buffer where they are formed into structures by rules described below. Phrasing rules then scan buffer structures. If a phrasing rule applies, all text up to the element that triggered the rule is sent to the synthesizer while, during synthesis, the buffer is reset and the roles restart anew. Once a structure has moved out of the buffer it cannot be recovered for exatnination by later phrasing rules.",
                "cite_spans": [
                    {
                        "start": 34,
                        "end": 46,
                        "text": "(Marcus 1980",
                        "ref_id": "BIBREF6"
                    },
                    {
                        "start": 47,
                        "end": 60,
                        "text": ", Hindle 1983",
                        "ref_id": "BIBREF4"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "PARSER STRUCTURE AND RULES",
                "sec_num": "3.1."
            },
            {
                "text": "Our approach differs from other recent efforts to build small parsers for text-to-speech, e.g. O'Shaughnessy (1988) and Emorine and Martin (1988) , where savings are sought in the lexicon rather than in processing.",
                "cite_spans": [
                    {
                        "start": 120,
                        "end": 145,
                        "text": "Emorine and Martin (1988)",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "PARSER STRUCTURE AND RULES",
                "sec_num": "3.1."
            },
            {
                "text": "O'Shaughnessy (1988) (henceforth O.) describes a non-deterministic parser that builds sentencelevel structure using a dictionary of 300 entries and a medium sized grammar, which we guess to be slightly under 100 rules. The lexicon is augmented by a morphological component of 60 word suffixes used principally to derive part of speech; for example, .ship and -hess are considered good indicators that a word of two or more syllables has the category 'noun'. O. gives a thorough account of his parser. Much of his exposition focusses on technical details of the syntactic analysis, and supporting linguistic data are plentiful. However, evaluation of O.'s proposals for speech synthesis is difficult since he gives us only a vague indication of how the parsed sentences would be prosodically phrased in a text-to-speech system.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "PARSER STRUCTURE AND RULES",
                "sec_num": "3.1."
            },
            {
                "text": "Without an explicit description of the syntax/prosody relation, we cannot be sure how to assess the suitability of O.'s analysis for speech applications.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "PARSER STRUCTURE AND RULES",
                "sec_num": "3.1."
            },
            {
                "text": "The system described by Emorine and Martin (1988) (henceforth E&M) incorporates a 300-entry dictionary and approximately 50 rules for identifying syntactic constituents and marking prosodic phrase boundaries. The rules in this system build sentence-level structures that are syntactically simpler than those given in O. but more geared to the requirements of phrasing in that prosodic events (e.g. pause) are explicitly mentioned in the rules. Unfortunately, E&M share few technical details about their system and, like O., provide no examples of the prosodic phrasing produced by their system, making evaluation an elusive task.",
                "cite_spans": [
                    {
                        "start": 24,
                        "end": 49,
                        "text": "Emorine and Martin (1988)",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "PARSER STRUCTURE AND RULES",
                "sec_num": "3.1."
            },
            {
                "text": "Applications such as TRS, which requires near real time processing, make systems based on sentence-level analyses infeasible. In our parser, decisions about phrasing are necessarily local--they depend on lexical information and word adjacency but not upon relations among non-contiguous elements. This combined with the need for lexical regularization in TDD texts motivates a much stronger lexicon than that of O. or E&M. In addition, our parser incorporates a small number of part-of-speech disambiguation rules to make additional lexical information available to the phrasing rules. Let us briefly describe each of the three components that make up the grammar: lexicon, disambiguation rules, and phrasing rules.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "PARSER STRUCTURE AND RULES",
                "sec_num": "3.1."
            },
            {
                "text": "3.1.1. The lexicon contains 1029 entries consisting of words, abbreviations, and two-to three-word phrases. Each entry has four fields: the input word (e.g. u), the output orthography (you), lexical category (Noun), and a list of word subclasses (destress_pronoun short_subject). Word subclasses reflect co-occurrence patterns and may or may not have any relationship to lexical categories. For example, Interjectionl includes the phrase byebye for now, the adverb however, the noun phrase my goodness, and the verb smile, as in I APPRECIATE THE I-lFff.p SMILE THANK YOU SO MUCH. Both the lexical category and subclass fields are optional--either may be marked as NIL. Abbreviations and acronyms are usually nouns and make up 20% of the lexical entries. Nouns and verbs together make up about 50%. We expect that additions to the lexicon will consist mostly of new abbreviations and short phrases.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "PARSER STRUCTURE AND RULES",
                "sec_num": "3.1."
            },
            {
                "text": "identify part-ofspeech and expand ambiguous abbreviations. Currently, part-of-speech disambiguation is performed by ten rules. Most apply to words lexically marked for both noun and verb, e.g. act, call, need, assigning a single category, either noun or verb, when a rule's contextual tests are satisfied. For example, if the third term of the buffer contains a word that is lexically marked as 'noun+verb', the word will be assigned the category 'verb' when the second buffer element is the word to and the first buffer element is either a verb or adverb. When applied to the word string expect to call, tiffs rule correctly analyzes call as a verb. Other part-of-speech rules distinguish the preposition to from the use of to as an infinitive marker, and distinguish the preposition vs. verb uses of like.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexical disambiguation rules",
                "sec_num": "3.1.2."
            },
            {
                "text": "Ambiguous abbreviations are items such as no, which may signify either number or the negative particle. Since TDD texts lack punctuation, the only clue to usage in such cases is local context, e.g. the presence of the words the or phone before no are used as disambiguating context to identify no as number. User identify phrase boundary locations. These rules are strictly ordered. In general, they instruct the synthesizer to set off interjections (e.g. wow, oh ok, etc.), and to insert a phrase boundary before non-lexical coordinate conjunctions (e.g. and in I don't recall that and am not sure, see Bachenko and Fitzpatrick (1990:163) ), before sentences, and before subordinate conjunctions (after, during, etc.) . Boundaries are also inserted at noun-verb junctures unless the noun is short, and at prepositional phrase boundaries unless the prepositional phrase is short. A short noun is a single word noun phrase such as a pronoun or demonstrative (this, that); a short prepositional phrase is one with a pronominal object (with me, about it, etc.). Hence the noun-verb rule will produce the phrasings below, where double bars mark phrase boundaries (this and the prepositional phrase rule are adaptations of the verb and length rules, respectively, given in Bachenko and Fitzpatrick (1990) ).",
                "cite_spans": [
                    {
                        "start": 604,
                        "end": 639,
                        "text": "Bachenko and Fitzpatrick (1990:163)",
                        "ref_id": null
                    },
                    {
                        "start": 697,
                        "end": 718,
                        "text": "(after, during, etc.)",
                        "ref_id": null
                    },
                    {
                        "start": 1268,
                        "end": 1299,
                        "text": "Bachenko and Fitzpatrick (1990)",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexical disambiguation rules",
                "sec_num": "3.1.2."
            },
            {
                "text": "Our formulation of the phrasing rules assumes that, in the absence of syntactic structure, the subclass membership, part-of-speech and string position can provide sufficient information to infer structure in many cases. For example, we are assuming that the subclass 'nominative_pronoun', which includes he, she, we, etc., acts consistently as the leading edge of a sentence, so that the parser can compensate somewhat for the lack of punctuation by identifying and setting off some top-level sentential constituents.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "MY CAR [[ IS HAVING A TRANSMISSION PROBLEM IT IS HAVING I[ A TRANSMISSION PROBLEM",
                "sec_num": null
            },
            {
                "text": "Similarly, prepositions are assumed to act consistently as the leading edge of a prepositional phrase; the parser guesses about prepositional phrase length by checking the word class of the element following the preposition to see if the object is pronominal.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "MY CAR [[ IS HAVING A TRANSMISSION PROBLEM IT IS HAVING I[ A TRANSMISSION PROBLEM",
                "sec_num": null
            },
            {
                "text": "The phrase rules thus attempt to seek out major syntactic constituents. If there is evidence of constituency, the parser may look for a short constituent or it will simply insert a prosodic boundary at a presumed syntactic boundary (e.g. a verb phrase, sentence or subordinate conjunction).",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "MY CAR [[ IS HAVING A TRANSMISSION PROBLEM IT IS HAVING I[ A TRANSMISSION PROBLEM",
                "sec_num": null
            },
            {
                "text": "The quickest way to incorporate a TDD parser into a ser vice using text-to-speech (TrS) synthesis is to implemen the parser in a separate front-end module to the text-to speech system. The parser filters the input stream from TDD modem and sends the processed text to the text-to speech system where it is synthesized for the voice tele phone user, as shown in the block diagram in figure 2 This architecture minimizes the need to modify an~ existing equipment or system. Also, it allows us u maintain and change the parser module without introduc ing substantial, or unpredictable, changes elsewhere ii the system.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "SYSTEM ARCHITECTURE",
                "sec_num": "3.2.1."
            },
            {
                "text": "Integrating the TDD parser into a near real time systen architecture is a difficult task. To achieve it, the parse must (a) filter the TDD input stream in real-time in orde to identify tokens, i.e. words, abbreviations, and expres sions, that are suitable for processing by parser rules, an~ (b) group these tokens into natural sounding phrases tha can be sent to the text-to-speech system as soon as the,. are formed.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "IMPLEMENTATION",
                "sec_num": "3.2.2."
            },
            {
                "text": "In an ideal situation, it is desirable to parse the entir TDD input before sending the processed text to the text to-speech synthesizer.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "IMPLEMENTATION",
                "sec_num": "3.2.2."
            },
            {
                "text": "But the practical situatio~ demands that the voice user hear TDD text synthesizeq as soon as it is reasonably possible so that long period of silence can be prevented. Figure 3 below shows th, basic architecture chosen to implement the parse described in this paper.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 168,
                        "end": 176,
                        "text": "Figure 3",
                        "ref_id": "FIGREF1"
                    }
                ],
                "eq_spans": [],
                "section": "IMPLEMENTATION",
                "sec_num": "3.2.2."
            },
            {
                "text": "The canonical input filter process has to de.~ with the TDD input characters as they are being typed The output of the canonical filters consists of TDD wor, tokens i.e. groups of characters separated by whit spaces. Input characters arrive at irregular speeds wit nondeterministic periods of pauses due to uneven typin by the TDD user. Also incidences of spelling erro~ typographical mistakes, and attempts to amend previousl typed text occur at very irregular rates. Even the TDI modem can contribute text to the input stream that i seen by the canonical input filter. For instance, the TDD modem might periodically insert a carriage-return character to prevent text wraparounds on tim special operator's terminal. Unfommately, these carriage-return characters could split words typed by tim TDD user into incoherent parts, e.g., advantage might become adva<CR>ntage.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "3.2.2.1.",
                "sec_num": null
            },
            {
                "text": "Since the voice telephone user needs to hear TDD text synthesized after some, hopefully short, interval of time, the input filter cannot wait indefinitely for TDD characters that are delayed in arriving, as might occur when the TDD user pauses to consider what to type next. Hence, the filter includes an input character timeout mechanism. The timeout interval is set to an appropriately short duration to ensure the timely synthesis of available TDD text, but still long enough to prevent the exclusion of forthcoming input characters.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "3.2.2.1.",
                "sec_num": null
            },
            {
                "text": "word tokens to identify contiguous words that should be grouped together as individual units. The multi-word expressions include contractions (e.g. \"it .... s\" which becomes \"it's'3 and commonly used short phrases that can be viewed as sIngle lexical units (e.g.\"ray goodness\", \"as long as\", and \"mother in law\"). A simple stacking mechanism is used to save tokens that are identified as potential elements of multi-word expressions. The tokens are stacked until the longest potential multi-word expression has been identified, with three words being the maximum. After which the stack is popped and the corresponding structures (described below) are constructed.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexigraphical analysis examines the TDD",
                "sec_num": "3.22.2."
            },
            {
                "text": "3.223. The lexical lookup process builds a tddterm structure (record) from these tokenized words and multi-word expressions in preparation for invoking the phrasal segmentation rules. Fields in the structure include the tokenized input text (the original orthographic representation), the output orthography, lexical category (Noun, Verb, Adverb, NIL, etc.), word subclass, and other fields used internally by the phrasal segmentation process. At this point in the processing only the input text field has any non-nnll information. The output orthography, lexical category, and word subclass fields are filled via lexical lookup.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexigraphical analysis examines the TDD",
                "sec_num": "3.22.2."
            },
            {
                "text": "The lexicon is organized into the four fields mentioned above. The tdd-term input text field is compared with the corresponding fieM in the lexicon until a match is found and the three remaining fields in the matched entry am then copied into the tdd-term structure. If no match is found, then the input text field is copied into the output text field and the other two lexicon fields are set to NIL.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexigraphical analysis examines the TDD",
                "sec_num": "3.22.2."
            },
            {
                "text": "As an illustration, ff the single letter u is identified as our TDD token, the lexical lookup process might return with a tdd-term stmcnne that looks like: input text:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexigraphical analysis examines the TDD",
                "sec_num": "3.22.2."
            },
            {
                "text": "\"u\" output text:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexigraphical analysis examines the TDD",
                "sec_num": "3.22.2."
            },
            {
                "text": "\"you\" lexical category: NOUN subclasses:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexigraphical analysis examines the TDD",
                "sec_num": "3.22.2."
            },
            {
                "text": "(DESTRESS_PRONOUN SHORT_SUBJECT) other fields:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexigraphical analysis examines the TDD",
                "sec_num": "3.22.2."
            },
            {
                "text": "NIL.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexigraphical analysis examines the TDD",
                "sec_num": "3.22.2."
            },
            {
                "text": "For tim input text oic, the structure might look like: input text: \"oic\" output text:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexigraphical analysis examines the TDD",
                "sec_num": "3.22.2."
            },
            {
                "text": "\"oh, I see\" lexical category: INTJ subclasses:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexigraphical analysis examines the TDD",
                "sec_num": "3.22.2."
            },
            {
                "text": "INTERJECTION 1 other fields: NIL.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexigraphical analysis examines the TDD",
                "sec_num": "3.22.2."
            },
            {
                "text": "32.2.4. The phrasal segmentation process applies a modest set of disambiguation and phrasing roles to a sliding window containing three contiguous tdd-term structures. In the start condition the sliding window wiLl not have any tdd-term structures within it. Each new tdd-term structure generated by lexical lookup enters the first term position in tim window, bumping existing terms forward one position with the last (third term) discarded after its output orthography is copied into a text buffer awaiting transmission to the text-to-speech synthesizer. The various rules described in Section 3.1 above are then applied to the available tdd-term structures. After a pronounceable phrase is identified, the output orthography of all active tdd-terms is then copied to the TTS text buffer which is subsequently sent to the synthesizer for playback to the voice telephone user. Also, the invocation of a timeout alarm due to tardy TDD input text causes flushing of the sliding window and text buffer into tim synthesizer. The sliding window and Trs text buffer axe cleared and the roles restarted anew.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Lexigraphical analysis examines the TDD",
                "sec_num": "3.22.2."
            },
            {
                "text": "Lexica! Lookup ",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Canonical Input Filter Lexigraphical Analysis",
                "sec_num": null
            },
            {
                "text": "ABOUT IT GA T/S: I don't think so I will think about it. Go ahead. ",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "TDD: I DONT THINK SO I WILL THINK",
                "sec_num": null
            },
            {
                "text": "Evaluation of the parser has involved two quite different forms of testing: a field trial and laboratory evaluation. First, the parser was implemented as a component in a version of the Bell Labs text-to-speech synthesizer (Olive and Liberman 1985) . The synthesizer forms the core of a telecommunications system that ran for three months as a feature of TRS in California. Several thousand TDD texts were processed by the system. Although restrictions on confidentiality prevented us from collecting actual TDD text data, results of the field trial far surpassed expectations: disconnect rates for text-to-speech calls averaged less than 20% and follow-up surveys indicated a high degree of interest in and acceptance of the technology.",
                "cite_spans": [
                    {
                        "start": 223,
                        "end": 248,
                        "text": "(Olive and Liberman 1985)",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "EVALUATION OF PERFORMANCE",
                "sec_num": "4."
            },
            {
                "text": "A second type of testing that has enabled us to focus on the parser involves the collection of data from a questionnaire given to TDD users. Phrasing for these data was assigned manually by a linguist unfamiliar with the rules of the parser to allow for comparison with the parser's output.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "EVALUATION OF PERFORMANCE",
                "sec_num": "4."
            },
            {
                "text": "Several issues arise in the comparison of human judgements of phrasing with those of a phrase parser's output. One of the more ubiquitous is that of phrasal balancing. Apparently acting under rhythmic coastraint.~ speakers tend to aim for equivalent numbers of stresse~ syllables on either side of a break. However, the incot potation of rhythm into phrasing varies from speaker t, speaker, as well as being partially dependent on semanti intent. For example, the sentence so I feel there shoul, be a better system to say bye, taken from our data, coul, be phrased either as (a), (b), or (c):",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "EVALUATION OF PERFORMANCE",
                "sec_num": "4."
            },
            {
                "text": "(a) so I feel there should be [I a better system to say bye (b) so I feel I] there should be II a better system to say bye (c) so I feel II there should be a better system I I to say bye If the parser assigns, for example, the phrasing in (~ while the human judge assigns (b) it must be counted a qualitatively different from the parser's assignment of misleading boundary, where the bearer's understandin of the import of the utterance is altered because of th erroneous boundary placement. An example of mislead ing boundary placement as assigned by the parser i given below, where the bearer is incorrectly led to intel pret well as a modification of see, rather than as discourse comment.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "EVALUATION OF PERFORMANCE",
                "sec_num": "4."
            },
            {
                "text": "In a similar vein, giving equal weight in an evalu~ tion to the locations where pauses do and do not occur i misleading. The absence of a phrasal boundary betwee two words is much more common than the presence of boundary, so that predicting the absence of a boundary i always safer and leads to inflated evaluation scores th~ make comparison of systems difficult. For example, i the (a) sentence above there are 12 potential prosodi events, one after each word. If a given system assigr no breaks in this sentence, and if non-events are give equal weight with events, then the system will get score for this sentence of 91.6 percent since it gets l 1 c the 12 judgments right. Also, if a system assigns o~ break in this utterance, but puts it in a clearly inappropr ate place, say before the word bye, it will get a score c 83 percent since it gets 10 of the 12 judgements righ While 83 percent sounds like a decent score for a systex that must capture some subjective performance, tit method of evaluation has completely failed to capture tt fact that assigning an inappropriate prosodic break in th instance has completely misled the listener. Therefor we need to evaluate a phrasing system on the basis c positive occurrences of phrase boundaries only.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "oh i see well I [ so i call my boss",
                "sec_num": null
            },
            {
                "text": "Assigning phrases to TDD output is not a clear-c1 task. The output is not intended to be spoken anq because of the device, it has telegraphic characteristic In addition, many TDD users do not have standard sp, ken English at their command. Nevertheless, an effort was made to approximate the performance of TRS operators who speak the TDD output to voice users. Care was also taken to mark as parser errors those prosodic events that would mislead the listener. This is a task that, in itself, is problematic because of the human error involved in the judgments. We regard the field trial as the appropriate method of evaluation here, and we use the results of the laboratory test to help us characterize the parser's failures rather than to evaluate the parser.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "oh i see well I [ so i call my boss",
                "sec_num": null
            },
            {
                "text": "After the phrasing was assigned manually, the TDD data were run through the parser, and the parser's phrasing was compared with the human judgments. Approximately 20% of the corpus had been used to extrapolate rules for the parser, while the remainder of the corpus was saved for testing only; there was no appreciable performance difference between the two data subsets. The corpus contained 8788 words and, according to the human judgement, 2922 phrases. Since punctuation in these data is sparse, very few of the actual phrase boundaries come \"for free.\" Even so, the parser performed well: in the 2922 phrases it produced 573 misleading errors, rendering 80.4% of the phrases acceptably. (There were 896 sites where the parser produced a phrasing different from that of the human judge, but which we judged to be not misleading.)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "oh i see well I [ so i call my boss",
                "sec_num": null
            },
            {
                "text": "The parser's error rate reflects the constraints of its construction as a real-time system, in particular, its three term buffer size, its lack of hierarchical structure building rules and its pared down lexicon. Figure 4 gives a characterization of the most frequently encountered parsing errors, along with their frequency of occurrence, and an example for each characterization. In the examples, '[l' represents a prosodic pause and \"*' indicates that the parser performed incorrectly at this site. Figure 4 would be resolvable if the parser were to incorporate non-local structural information. For example, the pronouns it and you function as both subject and object. In a three element buffer, then, the status of it in to it will is undecidable, since it can be the object of to or the subject of will. In the context of the sentence i have friends who i long to talk to it will be great, where an element corresponding to who functions as the object of to, the function of it as subject of will be great, and the concomitant prosodic break before it, are apparent, but only when the structure of the who relative clause is available.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 213,
                        "end": 221,
                        "text": "Figure 4",
                        "ref_id": "FIGREF2"
                    },
                    {
                        "start": 502,
                        "end": 510,
                        "text": "Figure 4",
                        "ref_id": "FIGREF2"
                    }
                ],
                "eq_spans": [],
                "section": "oh i see well I [ so i call my boss",
                "sec_num": null
            },
            {
                "text": "The errors involving non-standard syntax would require sublanguage roles that indicate the possibility of non-overt subjects (oh i see understand) and copulas (there a pause) among other things, but again, given the limitation to local information and the lack of punctuation, this is not straightforward. For example, oh i see understand could continue as that i don't speak well.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Most of the parsing errors given in",
                "sec_num": null
            },
            {
                "text": "A smaller set of errors is undecidable even given non-local structural information, and require further pragmatic knowledge of the discourse. For example, the decision as to which clause the adverb occasionally modifies in other than that the services is great occasionally some operators are pretty slow depends on knowing that one does not give expansive praise to something that happens only occasionally.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Most of the parsing errors given in",
                "sec_num": null
            },
            {
                "text": "In general, it appears that the parser's accuracy in phrasing the incoming messages cannot he improved without a severe loss in real time efficiency that the storage of hierarchical structure would involve. Given this situation, it is worthwhile to consider that. despite what is probably about a 20% error rate in the system, consumers used it successfully and willingly. It may be that the system did no worse than the trs operators who, unlike our laboratory linguist, do not have the luxury of stopping to consider the felicity of a particular phrasing. This may be compounded with the possibility that users may be able to compensate more easily for machine degradation of an utterance than for an operator's error, since their expectations of the latter's performance are greater.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Most of the parsing errors given in",
                "sec_num": null
            },
            {
                "text": "We have described a text-to-speech parser for conversational texts generated by users of TDD's. The parser's main ta-ck.~ are to provide some regularization of nonstandard items and to determine prosodic phrasing of the text. Phrasing allows processing to take place in near real time because the synthesizer can generate speech while the TDD message is being typed instead of waiting for the finished text. FinaUy, although incorporating a relatively small vocabulary and rule set, the parser has proven unexpectedly successful in both laboratory and field tests,",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "CONCLUSION",
                "sec_num": "5."
            }
        ],
        "back_matter": [],
        "bib_entries": {
            "BIBREF0": {
                "ref_id": "b0",
                "title": "A Taxonomy of Syntactic Variation in Written Language of the Deaf. Unpublished data",
                "authors": [
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Bachenko",
                        "suffix": ""
                    }
                ],
                "year": 1990,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Bachenko, J. A Taxonomy of Syntactic Variation in Written Language of the Deaf. Unpublished data, 1990.",
                "links": null
            },
            "BIBREF1": {
                "ref_id": "b1",
                "title": "A Computational Grammar of Discourse-Neutral Prosodic Phrasing in English Computational Linguistics",
                "authors": [
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Bacheako",
                        "suffix": ""
                    },
                    {
                        "first": "E",
                        "middle": [],
                        "last": "Fitzpatrick",
                        "suffix": ""
                    }
                ],
                "year": 1990,
                "venue": "",
                "volume": "16",
                "issue": "",
                "pages": "155--172",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Bacheako, J. and E. Fitzpatrick. A Computational Gram- mar of Discourse-Neutral Prosodic Phrasing in English Computational Linguistics, 16:155-17, 1990.",
                "links": null
            },
            "BIBREF2": {
                "ref_id": "b2",
                "title": "Institute for Mathematical Studies in the Social Sciences",
                "authors": [
                    {
                        "first": "V",
                        "middle": [
                            "Deaf"
                        ],
                        "last": "Charrow",
                        "suffix": ""
                    },
                    {
                        "first": "",
                        "middle": [],
                        "last": "English",
                        "suffix": ""
                    }
                ],
                "year": 1974,
                "venue": "",
                "volume": "236",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Charrow, V. Deaf English. Technical Report 236, Insti- tute for Mathematical Studies in the Social Sciences, Stanford University, 1974.",
                "links": null
            },
            "BIBREF3": {
                "ref_id": "b3",
                "title": "The Multivoc Textto-Speech System",
                "authors": [
                    {
                        "first": "O",
                        "middle": [
                            "M"
                        ],
                        "last": "Emorine",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [
                            "M"
                        ],
                        "last": "Martin",
                        "suffix": ""
                    }
                ],
                "year": 1988,
                "venue": "Proceedings of the Second Conference on Applied Natural Language Processing",
                "volume": "",
                "issue": "",
                "pages": "115--120",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Emorine, O. M. and P. M. Martin. The Multivoc Text- to-Speech System. Proceedings of the Second Confer- ence on Applied Natural Language Processing (ACL): 115-120, 1988.",
                "links": null
            },
            "BIBREF4": {
                "ref_id": "b4",
                "title": "User Manual for Fidditch, a Deterministic Parser. NRL Technical Memorandum#7590-142",
                "authors": [
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Hindle",
                        "suffix": ""
                    }
                ],
                "year": 1983,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Hindle, D. User Manual for Fidditch, a Deterministic Parser. NRL Technical Memorandum#7590-142. 1983.",
                "links": null
            },
            "BIBREF5": {
                "ref_id": "b5",
                "title": "Spelling Correction for the Telecommunications Network for the Deaf",
                "authors": [
                    {
                        "first": "K",
                        "middle": [],
                        "last": "Kukich",
                        "suffix": ""
                    }
                ],
                "year": 1992,
                "venue": "Communications of the ACM",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Kukich, K. Spelling Correction for the Telecommunica- tions Network for the Deaf. Communications of the ACM, 1992.",
                "links": null
            },
            "BIBREF6": {
                "ref_id": "b6",
                "title": "A Theory of Syntactic Recognition for Natural Language",
                "authors": [
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Marcus",
                        "suffix": ""
                    }
                ],
                "year": 1980,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Marcus, M. A Theory of Syntactic Recognition for Natural Language. Cambridge, MA: MIT Press, 1980.",
                "links": null
            },
            "BIBREF7": {
                "ref_id": "b7",
                "title": "Text-to-Speech--An Overview",
                "authors": [
                    {
                        "first": "J",
                        "middle": [
                            "P"
                        ],
                        "last": "Olive",
                        "suffix": ""
                    },
                    {
                        "first": "M",
                        "middle": [
                            "Y"
                        ],
                        "last": "Liberman",
                        "suffix": ""
                    }
                ],
                "year": 1985,
                "venue": "Journal of the Acoustic Society of America",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Olive, J. P. and Liberman, M. Y. Text-to-Speech--An Overview. Journal of the Acoustic Society of America, Supplement 1:78, $6, 1985.",
                "links": null
            },
            "BIBREF8": {
                "ref_id": "b8",
                "title": "Parsing with a Small Dictionary for Applications such as Text-to-Speech Computational Linguistics",
                "authors": [
                    {
                        "first": "D",
                        "middle": [
                            "D"
                        ],
                        "last": "O'shaughnessy",
                        "suffix": ""
                    }
                ],
                "year": 1988,
                "venue": "",
                "volume": "15",
                "issue": "",
                "pages": "97--108",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "O'Shaughnessy, D. D. Parsing with a Small Dictionary for Applications such as Text-to-Speech Computational Linguistics, 15:97-108, 1988.",
                "links": null
            },
            "BIBREF9": {
                "ref_id": "b9",
                "title": "Language Transfer: A Foundation for Correcting the Written English of ASL Signers",
                "authors": [
                    {
                        "first": "L",
                        "middle": [],
                        "last": "Suri",
                        "suffix": ""
                    }
                ],
                "year": 1991,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Suri, L. Language Transfer: A Foundation for Correcting the Written English of ASL Signers. University of Delaware Technical Report #91-19, 1991.",
                "links": null
            },
            "BIBREF10": {
                "ref_id": "b10",
                "title": "A Lexical Study of Sentences Typed by Hearing-Impaired TDD Users",
                "authors": [
                    {
                        "first": "Y.-C",
                        "middle": [],
                        "last": "Tsao",
                        "suffix": ""
                    }
                ],
                "year": 1990,
                "venue": "Proceedings of 13th International Symposium",
                "volume": "",
                "issue": "",
                "pages": "19--201",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Tsao, Y.-C. A Lexical Study of Sentences Typed by Hearing-Impaired TDD Users. Proceedings of 13th International Symposium, Human Factors in Telecom- munications, Torino, Italy, 19%201, 1990.",
                "links": null
            }
        },
        "ref_entries": {
            "FIGREF0": {
                "uris": null,
                "text": "3.1.3. Phrasing rules consider part-of-speech, word subclass and length (as measured by word count) to",
                "type_str": "figure",
                "num": null
            },
            "FIGREF1": {
                "uris": null,
                "text": "TDD Parser ArchitectureListed below are a few examples of TDD text processed by the parser.",
                "type_str": "figure",
                "num": null
            },
            "FIGREF2": {
                "uris": null,
                "text": "Distribution of TDD Production Errors",
                "type_str": "figure",
                "num": null
            }
        }
    }
}