Benjamin Aw
Add updated pkl file v3
6fa4bc9
raw
history blame contribute delete
No virus
33.5 kB
{
"paper_id": "W12-0108",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T04:12:56.392858Z"
},
"title": "PRESEMT: Pattern Recognition-based Statistically Enhanced MT",
"authors": [
{
"first": "George",
"middle": [],
"last": "Tambouratzis",
"suffix": "",
"affiliation": {},
"email": ""
},
{
"first": "Marina",
"middle": [],
"last": "Vassiliou",
"suffix": "",
"affiliation": {},
"email": ""
},
{
"first": "Sokratis",
"middle": [],
"last": "Sofianopoulos",
"suffix": "",
"affiliation": {},
"email": ""
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "This document contains a brief presentation of the PRESEMT project that aims in the development of a novel language-independent methodology for the creation of a flexible and adaptable MT system.",
"pdf_parse": {
"paper_id": "W12-0108",
"_pdf_hash": "",
"abstract": [
{
"text": "This document contains a brief presentation of the PRESEMT project that aims in the development of a novel language-independent methodology for the creation of a flexible and adaptable MT system.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "The PRESEMT project constitutes a novel approach to the machine translation task. This approach is characterised by (a) introducing crossdisciplinary techniques, mainly borrowed from the machine learning and computational intelligence domains, in the MT paradigm and (b) using relatively inexpensive language resources. The aim is to develop a language-independent methodology for the creation of a flexible and adaptable MT system, the features of which ensure easy portability to new language pairs or adaptability to particular user requirements and to specialised domains with minimal effort. PRESEMT falls within the Corpus-based MT (CBMT) paradigm, using a small bilingual parallel corpus and a large TL monolingual corpus. Both these resources are collected as far as possible over the web, to simplify the development of resources for new language pairs.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "The main aim of PRESEMT has been to alleviate the reliance on specialised resources. In comparison, Statistical MT requires large parallel corpora for the source and target languages. PRESEMT relaxes this requirement by using a small parallel corpus, augmented by a large TL monolingual corpus.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "The PRESEMT system is distinguished into three stages, as shown in Figure 1: 1. Pre-processing stage: This is the stage where the essential resources for the MT system are compiled. It consists of four discrete modules: (a) the Corpus creation & annotation module, being responsible for the compilation of monolingual and bilingual corpora over the web and their annotation; (b) the Phrase aligner module, which processes a bilingual corpus to perform phrasal level alignment within a language pair; (c) the Phrasing model generator that elicits an SL phrasing model on the basis of the aforementioned alignment and employs it as a parsing tool during the translation process; (d) the Corpus modelling module, which creates semanticsbased TL models used for disambiguation purposes during the translation process.",
"cite_spans": [],
"ref_spans": [
{
"start": 67,
"end": 76,
"text": "Figure 1:",
"ref_id": null
}
],
"eq_spans": [],
"section": "PRESEMT system structure",
"sec_num": "2."
},
{
"text": "The translation in PRESEMT is a top-down two-phase process, distinguished into the Structure selection module, where the constituent phrases of an SL sentence are reordered according to the TL, and the Translation equivalent selection module where translation disambiguation is resolved and word order within phrases is established. Closely integrated to the translation engine, but not part of the main translation process, is the Optimisation module, which is responsible for automatically improving the performance of the two translation phases by fine-tuning the values of the various system parameters.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Main translation engine:",
"sec_num": "2."
},
{
"text": "The third stage is useroriented and comprises (i) the Post-processing and (ii) the User Adaptation modules. The first module allows the user to modify the systemgenerated translations towards their requirements. The second module enables PRESEMT to adapt to this input so that it learns to generate translations closer to the users' requirements. The postprocessing stage represents work in progress to be reported in future publications, the present article focussing on the actual strategy for generating the translation.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Post-processing stage:",
"sec_num": "3."
},
{
"text": "The bilingual corpus contains literal translations, to allow the extrapolation of mapping information from SL to TL, though this may affect the translation quality. The Phrase aligner module (PAM) performs offline SL -TL word and phrase alignment within this corpus. PAM serves as a language-independent method for mapping corresponding terms within a language pair, by circumventing the problem of achieving compatibility between the outputs of two different parsers, one for the SL and one for the TL. PAM relies on a single parser for the one language and generates an appropriate phrasing model for the other language in an automated manner. The phrases are assumed to be flat and linguistically valid. As a parser, any available tool may be used (the TreeTagger (Schmid, 1994) is used in the present implementation for English). PAM processes a bilingual corpus of SL -TL sentence pairs, taking into account the parsing information in one language (in the current implementation the TL side) and making use of a bilingual lexicon and information on potential phrase heads; the output being the bilingual corpus aligned at word, phrase and clause level. Thus, at a phrasal level, the PAM output indicates how an SL structure is transformed into the TL. For instance, based on a sentence pair from the parallel corpus, the SL sentence with structure A-B-C-D is transformed into A'-C'-D'-B', where X is a phrase in SL and X' is a phrase in TL. Further PAM details are reported in Tambouratzis et al. (2011).",
"cite_spans": [
{
"start": 767,
"end": 781,
"text": "(Schmid, 1994)",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Processing of the bilingual corpus",
"sec_num": "3."
},
{
"text": "The PAM output in terms of SL phrases is then handed over to the Phrasing model generator (PMG), which is trained to determine the phrasal structure of an input sentence. PMG reads the SL phrasing as defined by PAM and generates an SL phrasing model using a probabilistic methodology. This phrasing model is then applied in segmenting any arbitrary SL text being input to the PRESEMT system for translation. PMG is based on the Conditional Random Fields model (Lafferty et al., 1999) which has been found to provide the highest accuracy. The SL text segmented into phrases by PMG is then input to the 1 st translation phase. For a new language pair, the PAM-PMG chain is implemented without any manual correction of outputs.",
"cite_spans": [
{
"start": 460,
"end": 483,
"text": "(Lafferty et al., 1999)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Processing of the bilingual corpus",
"sec_num": "3."
},
{
"text": "The language models created by the Corpus modelling module can only serve translation dis-ambiguation purposes; thus another form of interfacing with the monolingual corpus is essential for the word reordering task within each phrase. The size of the data accessed is very large. Typically, a monolingual corpus contains 3 billion words, 10 8 sentences and approximately 10 9 phrases. Since the models for the TL phrases need to be accessed in real-time to allow word reordering within each phrase, the module uses the phrase indexed representation of the monolingual corpus. This phrase index is created based on four criteria: (i) phrase type, (ii) phrase head lemma, (iii) phrase head PoS tag and (iv) number of tokens in the phrase.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Organising the monolingual corpus",
"sec_num": "4."
},
{
"text": "Indexing is performed by extracting all phrases from the monolingual corpus, each of which is transformed to the java object instance used within the PRESEMT system. The phrases are then organised in a hash map that allows multiple values for each key, using as a key the 4 aforementioned criteria. Statistical information about the number of occurrences of each phrase in the corpus is also included. Finally, each map is serialised and stored in the appropriate file in the PRESEMT path, with each file being given a suitable name for easy retrieval. For example, for the English monolingual corpus, all verb phrases with head lemma \"read\" (verb) and PoS tag \"VV\" containing 2 tokens in total are stored in the file \"Corpora\\EN\\Phrases\\VC\\read_VV\". If any of these criteria has a different value, then a separate file is created (for instance for verb phrases with head \"read\" that contain 3 tokens).",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Organising the monolingual corpus",
"sec_num": "4."
},
{
"text": "The PRESEMT translation process entails first the establishment of the sentence phrasal structure and then the resolution of the intra-phrasal arrangements, i.e. specifying the correct word order and deciding upon the appropriate candidate translation. Both phases involve searching for suitable matching patterns at two different levels of granularity, the first (coarse-grained) aiming at defining a TL-compatible ordering of phrases in the sentence and the second (finegrained) determining the internal structure of phrases. While the first phase utilises the small bilingual corpus, the second phase makes use of the large monolingual corpus. To reduce the translation time required, both corpora are processed in advance and the processed resources are stored in such a form as be retrieved as rapidly as possible during translation.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Main translation engine",
"sec_num": "5."
},
{
"text": "Each SL sentence input for translation is tagged and lemmatised and then it is segmented into phrases by the Phrasing model generator on the basis of the SL phrasing model previously created. For establishing the correct phrase order according to the TL, the parallel corpus needs to be pre-processed using the Phrase aligner module to identify word and phrase alignments between the equivalent SL and TL sentences. During structure selection, the SL sentence is aligned to each SL sentence of the parallel corpus, as processed by the PAM and assigned a similarity score using an algorithm from the dynamic programming paradigm. The similarity score is calculated by taking into account edit operations (replacement, insertion or removal) needed to be performed in the input sentence in order to transform it to the corpus SL sentence. Each of these operations has an associated cost, considered as a system parameter. The aligned corpus sentence that achieves the highest similarity score is the most similar one to the input source sentence. This comparison process relies on a set of similarity parameters (e.g. phrase type, phrase head etc.), the values of which are optimised by employing the optimisation module.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Translation Phase 1: Structure selection module",
"sec_num": "5.1"
},
{
"text": "The implementation is based on the Smith-Waterman algorithm (Smith and Waterman, 1981) , initially proposed for determining similar regions between two protein or DNA sequences. The algorithm is guaranteed to find the optimal local alignment between the two input sequences at clause level.",
"cite_spans": [
{
"start": 60,
"end": 86,
"text": "(Smith and Waterman, 1981)",
"ref_id": "BIBREF6"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Translation Phase 1: Structure selection module",
"sec_num": "5.1"
},
{
"text": "After establishing the order of phrases within each sentence, the second phase of the translation process is initiated, comprising two distinct tasks. The first task is to resolve the lexical ambiguity, by picking one lemma from each set of possible translations (as provided by a bilingual dictionary). In doing so, this module makes use of the semantic similarities between words which have been determined by the Corpus Modelling module through a co-occurrence analysis on the monolingual TL corpus. That way, the best combination of lemmas from the sets of candidate translations is determined for a given context. In the second task, the most similar phrases to the TL structure phrases are retrieved from the monolingual corpus to provide local structural information such as word-reordering. A matching algorithm selects the most similar from the set of the retrieved TL phrases through a comparison process, which is viewed as an assignment problem, using the Gale-Shapley algorithm (Gale and Shapley, 1962) .",
"cite_spans": [
{
"start": 991,
"end": 1015,
"text": "(Gale and Shapley, 1962)",
"ref_id": "BIBREF0"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Translation Phase 2: Translation equivalent selection module",
"sec_num": "5.2"
},
{
"text": "To date MT systems based on the PRESEMT methodology have been created for a total of 8 languages, indicating the flexibility of the proposed approach. Table 1 illustrates an indicative set of results obtained by running automatic evaluation metrics on test data translated by the 1 st PRESEMT prototype for a selection of language pairs, due to space restrictions.",
"cite_spans": [],
"ref_spans": [
{
"start": 151,
"end": 158,
"text": "Table 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Experiments & evaluation results",
"sec_num": "6."
},
{
"text": "In the case of the language pair English-to-German, these results are contrasted to the ones obtained when translating the same test set with Moses (Koehn et al., 2007) .It is observed that for the English-to-German language pair, PRESEMT achieved approximately 50% of the MOSES BLEU score and 80% of the MOSES with respect to the Meteor and TER scores. These are reasonably competitive results compared to an established system such as Moses. Furthermore, it should taken into consideration that (a) the PRESEMT results were obtained by the 1 st system prototype, (b) PRESEMT is still under development and (c) only one reference translation was used per sentence.",
"cite_spans": [
{
"start": 148,
"end": 168,
"text": "(Koehn et al., 2007)",
"ref_id": "BIBREF1"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Experiments & evaluation results",
"sec_num": "6."
},
{
"text": "Newer versions of the PRESEMT system, incorporating more advanced versions of the different modules are expected to result in substantially improved translation accuracies. In particular, the second translation phase will be further researched. In addition, experiments have indicated that the language modelling module can provide additional improvement in the performance. Finally, refinements in PAM and PMG may lead in increased translation accuracies.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experiments & evaluation results",
"sec_num": "6."
},
{
"text": "Find out more about the project on the PRE-SEMT website: www.presemt.eu. Also, the PRESEMT prototype may be tried at: ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Links",
"sec_num": "7."
}
],
"back_matter": [
{
"text": "The research leading to these results has received funding from the European Community's Seventh Framework Programme (FP7/2007-2013 under grant agreement n\u00b0 248307.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Acknowledgments",
"sec_num": null
}
],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "College Admissions and the Stability of Marriage",
"authors": [
{
"first": "D",
"middle": [],
"last": "Gale",
"suffix": ""
},
{
"first": "L",
"middle": [
"S"
],
"last": "Shapley",
"suffix": ""
}
],
"year": 1962,
"venue": "American Mathematical Monthly",
"volume": "69",
"issue": "",
"pages": "9--14",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Gale D. and L. S. Shapley. 1962. College Admissions and the Stability of Marriage. American Mathe- matical Monthly, Vol. 69, pp. 9-14.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Moses: Open Source Toolkit for Statistical Machine Translation",
"authors": [
{
"first": "P",
"middle": [],
"last": "Koehn",
"suffix": ""
},
{
"first": "H",
"middle": [],
"last": "Hoang",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Birch",
"suffix": ""
},
{
"first": "C",
"middle": [],
"last": "Callison-Burch",
"suffix": ""
},
{
"first": "M",
"middle": [],
"last": "Federico",
"suffix": ""
},
{
"first": "N",
"middle": [],
"last": "Bertoldi",
"suffix": ""
},
{
"first": "B",
"middle": [],
"last": "Cowan",
"suffix": ""
},
{
"first": "W",
"middle": [],
"last": "Shen",
"suffix": ""
},
{
"first": "C",
"middle": [],
"last": "Moran",
"suffix": ""
},
{
"first": "R",
"middle": [],
"last": "Zens",
"suffix": ""
},
{
"first": "C",
"middle": [],
"last": "Dyer",
"suffix": ""
},
{
"first": "O",
"middle": [],
"last": "Bojar",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Constantin",
"suffix": ""
},
{
"first": "E",
"middle": [],
"last": "Herbst",
"suffix": ""
}
],
"year": 2007,
"venue": "Proceedings of the ACL-2007 Demo and Poster Sessions",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Koehn P., H. Hoang, A. Birch, C. Callison-Burch, M. Federico, N. Bertoldi, B. Cowan, W. Shen, C. Moran, R. Zens, C. Dyer, O. Bojar, A. Constantin, and E. Herbst. 2007. Moses: Open Source Toolkit for Statistical Machine Translation. In Proceedings of the ACL-2007 Demo and Poster Sessions.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "The Hungarian method for the assignment problem",
"authors": [
{
"first": "H",
"middle": [
"W"
],
"last": "Kuhn",
"suffix": ""
}
],
"year": 1955,
"venue": "Naval Research Logistics Quarterly",
"volume": "2",
"issue": "",
"pages": "83--97",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Kuhn H. W. 1955. The Hungarian method for the as- signment problem. Naval Research Logistics Quar- terly, Vol. 2, pp.83-97.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Conditional Random Fields: Probabilistic Models for Segmenting and Labelling Sequence Data. Proceedings of ICML Conference",
"authors": [
{
"first": "J",
"middle": [],
"last": "Lafferty",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Mccallum",
"suffix": ""
},
{
"first": "F",
"middle": [],
"last": "Pereira",
"suffix": ""
}
],
"year": 2001,
"venue": "",
"volume": "",
"issue": "",
"pages": "282--289",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Lafferty J., A. McCallum, F. Pereira. 2001. Condi- tional Random Fields: Probabilistic Models for Segmenting and Labelling Sequence Data. Pro- ceedings of ICML Conference, pp.282-289.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "Algorithms for the assignment and transportation problems",
"authors": [
{
"first": "J",
"middle": [],
"last": "Munkres",
"suffix": ""
}
],
"year": 1957,
"venue": "Journal of the Society for Industrial and Applied Mathematics",
"volume": "5",
"issue": "",
"pages": "32--38",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Munkres J. 1957. Algorithms for the assignment and transportation problems. Journal of the Society for Industrial and Applied Mathematics, Vol. 5, pp.32- 38.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Probabilistic Part-of-Speech Tagging Using Decision Trees",
"authors": [
{
"first": "H",
"middle": [],
"last": "Schmid",
"suffix": ""
}
],
"year": 1994,
"venue": "Proceedings of International Conference on New Methods in Language Processing",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Schmid, H. 1994. Probabilistic Part-of-Speech Tag- ging Using Decision Trees, Proceedings of Interna- tional Conference on New Methods in Language Processing, Manchester, UK.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Identification of Common Molecular Subsequences",
"authors": [
{
"first": "T",
"middle": [
"F"
],
"last": "Smith",
"suffix": ""
},
{
"first": "M",
"middle": [
"S"
],
"last": "Waterman",
"suffix": ""
}
],
"year": 1981,
"venue": "Journal of Molecular Biology",
"volume": "147",
"issue": "",
"pages": "195--197",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Smith T. F. and M. S. Waterman. 1981. Identification of Common Molecular Subsequences. Journal of Molecular Biology, 147: 195-197.",
"links": null
}
},
"ref_entries": {
"TABREF0": {
"type_str": "table",
"html": null,
"text": "presemt.cslab.ece.ntua.gr:8080/presemt_interface_test PRESEMT Evaluation results for different language pairs.",
"num": null,
"content": "<table><tr><td/><td/><td/><td/><td colspan=\"4\">Tambouratzis G., F. Simistira, S. Sofianopoulos, N.</td></tr><tr><td/><td/><td/><td/><td colspan=\"4\">Tsimboukakis and M. Vassiliou 2011. A resource-</td></tr><tr><td/><td/><td/><td/><td colspan=\"4\">light phrase scheme for language-portable MT,</td></tr><tr><td/><td/><td/><td/><td colspan=\"4\">Proceedings of the 15 th International Conference of</td></tr><tr><td/><td/><td/><td/><td colspan=\"4\">the European Association for Machine Translation,</td></tr><tr><td/><td/><td/><td/><td colspan=\"4\">30-31 May 2011, Leuven, Belgium, pp. 185-192.</td></tr><tr><td colspan=\"2\">Table 1 -Language Pair</td><td colspan=\"2\">Sentence set</td><td/><td colspan=\"2\">Metrics</td></tr><tr><td>SL</td><td>TL</td><td colspan=\"2\">Number Source</td><td>BLEU</td><td>NIST</td><td>Meteor</td><td>TER</td></tr><tr><td>English</td><td>German</td><td>189</td><td>web</td><td colspan=\"2\">0.1052 3.8433</td><td>0.1939</td><td>83.233</td></tr><tr><td>German</td><td>English</td><td>195</td><td>web</td><td colspan=\"2\">0.1305 4.5401</td><td>0.2058</td><td>74.804</td></tr><tr><td>Greek</td><td>English</td><td>200</td><td>web</td><td colspan=\"2\">0.1011 4.5124</td><td>0.2442</td><td>79.750</td></tr><tr><td>English</td><td>German</td><td>189</td><td>web</td><td colspan=\"2\">0.2108 5.6517</td><td>0.2497</td><td>68.190 Moses</td></tr><tr><td/><td/><td colspan=\"4\">Figure 1 -PRESEMT system architecture.</td><td/></tr><tr><td/><td/><td/><td>WEB</td><td/><td/><td/></tr><tr><td/><td/><td>Offline corpora</td><td colspan=\"2\">Corpus creation</td><td/><td/></tr><tr><td/><td/><td>creation</td><td colspan=\"2\">&amp; annotation</td><td/><td/></tr><tr><td/><td/><td/><td>module</td><td/><td/><td/></tr><tr><td/><td/><td/><td/><td/><td colspan=\"2\">TL Monolingual</td></tr><tr><td/><td/><td/><td/><td/><td colspan=\"2\">Corpus &amp;</td></tr><tr><td/><td/><td/><td/><td/><td colspan=\"2\">Corpus model</td></tr><tr><td/><td/><td/><td colspan=\"2\">Corpus modelling</td><td/><td/></tr><tr><td/><td/><td/><td>module</td><td/><td/><td/></tr><tr><td/><td/><td/><td/><td/><td colspan=\"2\">Bilingual parallel</td></tr><tr><td/><td/><td/><td colspan=\"2\">Phrase aligner</td><td colspan=\"2\">corpus</td></tr><tr><td/><td>Bilingual</td><td/><td>module</td><td/><td/><td/><td>Post-processing</td></tr><tr><td/><td>Lexicon</td><td/><td/><td/><td/><td/><td>&amp; User</td></tr><tr><td/><td/><td/><td colspan=\"2\">Phrasing model generator</td><td colspan=\"2\">Bilingual aligned corpus</td><td>adaptation modules</td></tr><tr><td/><td/><td colspan=\"3\">Phase 1: Structure selection</td><td colspan=\"2\">Phase 2: Translation equivalent</td></tr><tr><td>S SL L t te ex xt t</td><td>Tagging-Lemmatising, Lexicon look-up &amp;</td><td colspan=\"3\">Phrase reordering of each sentence in the SL text</td><td colspan=\"2\">selection Word reordering &amp; disambiguation</td><td>F Fi in na al l tr ra an ns sl la at ti io on n t</td></tr><tr><td/><td>Token generation</td><td/><td/><td/><td/><td/></tr><tr><td/><td/><td/><td colspan=\"3\">Optimisation module</td><td/></tr></table>"
}
}
}
}