File size: 36,696 Bytes
6fa4bc9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 |
{
"paper_id": "O12-1016",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T08:02:46.464674Z"
},
"title": "English-to-Traditional Chinese Cross-lingual Link Discovery in Articles with Wikipedia Corpus",
"authors": [
{
"first": "Liang-Pu",
"middle": [],
"last": "Chen",
"suffix": "",
"affiliation": {},
"email": ""
},
{
"first": "Yu-Lun",
"middle": [],
"last": "Shih",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "National Taipei Univeristy of Technology",
"location": {
"country": "Taiwan"
}
},
"email": ""
},
{
"first": "Chien-Ting",
"middle": [],
"last": "Chen",
"suffix": "",
"affiliation": {},
"email": ""
},
{
"first": "Tsun",
"middle": [],
"last": "Ku",
"suffix": "",
"affiliation": {},
"email": ""
},
{
"first": "Wen-Tai",
"middle": [],
"last": "Hsieh",
"suffix": "",
"affiliation": {},
"email": ""
},
{
"first": "Hung-Sheng",
"middle": [],
"last": "Chiu",
"suffix": "",
"affiliation": {},
"email": ""
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "In this paper, we design a processing flow to produce linked data in articles, providing anchorbased term's additional information and related terms in different languages (English to Chinese). Wikipedia has been a very important corpus and knowledge bank. Although Wikipedia describes itself not a dictionary or encyclopedia, it is if high potential values in applications and data mining researches. Link discovery is a useful IR application, based on Data Mining and NLP algorithms and has been used in several fields. According to the results of our experiment, this method does make the result has improved.",
"pdf_parse": {
"paper_id": "O12-1016",
"_pdf_hash": "",
"abstract": [
{
"text": "In this paper, we design a processing flow to produce linked data in articles, providing anchorbased term's additional information and related terms in different languages (English to Chinese). Wikipedia has been a very important corpus and knowledge bank. Although Wikipedia describes itself not a dictionary or encyclopedia, it is if high potential values in applications and data mining researches. Link discovery is a useful IR application, based on Data Mining and NLP algorithms and has been used in several fields. According to the results of our experiment, this method does make the result has improved.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "For our goal, we have to conquer some issues to find every potential linked data on articles. This paper focuses on Cross-lingual link discovery. Cross-lingual link discovery contains a lot of important tasks of NLP(Natural Language Processing) such as WSD(Word Sense Disambiguation) [1] , NED(Named Entities Disambiguation) [2] or Machine Translation. The cross-lingual links in the Wikipedia 1 are established by the human contributors, and not all Wikipedia Pages have cross lingual links because no human editors established these links yet. Thus, when one visits English Wikipedia page which describes some special information, users cannot find any cross lingual link to visit the Wikipedia page whose language is the same as the user's mother tongue. This problem has been raised by many recent studies [3, 4] , and recovering these missing links between two languages is the main goal of the CLLD (Cross-Lingual Link Discovery). In this paper, we propose a system which can automatically help users to tag potential links in their articles, and automatically find out the cross language link of the tag based on Wikipedia cross language links. As for cross lingual link discovery, our system is able to find the missing links between two related Wikipedia pages in two different language systems by exploiting and extracting data from Wikipedia dump files in two languages. In addition, we use two additional translation mechanisms to help find out the corresponding cross lingual translation , one is the Pattern Translate , the other one is Google Translate 2 . We further integrate the Lucene 3 software package to deal with the ambiguous phases in the articles. In order to find out the missing links between two pages, and automatically tagged this cross language link in users' articles.",
"cite_spans": [
{
"start": 284,
"end": 287,
"text": "[1]",
"ref_id": "BIBREF0"
},
{
"start": 325,
"end": 328,
"text": "[2]",
"ref_id": "BIBREF1"
},
{
"start": 810,
"end": 813,
"text": "[3,",
"ref_id": "BIBREF2"
},
{
"start": 814,
"end": 816,
"text": "4]",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "The remainder of this paper is organized as follows: First, we described corresponding background of Wikipedia and cross-lingual link discovery in Section 2. In Section 3,The proposed WSD method and translation mechanism will be described in detail. Finally, the experiment and conclusion will be discussed in Section 4.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Wikipedia is a free, collaboratively edited, and multilingual Internet encyclopedia supported by the non-profit Wikimedia Foundation 4 . Recently, many researchers focus on developing data mining applications with Wikipedia's large-scale collaborative user data. Although Wikipedia describes itself not a dictionary, textbook or encyclopedia, exploiting its characteristics to develop new services is regarded as a promising method on auto text explanation.",
"cite_spans": [
{
"start": 133,
"end": 134,
"text": "4",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Background 2.1 Wikipedia",
"sec_num": "2"
},
{
"text": "One of the special feature of Wikipedia is that it contains many hypertext links to help users easily retrieve the information they need. These hypertext links might be embedded within the text content under the corresponding pages, and each of these links is linking to other pages related with different terms. Obviously, information flow is thus being traversed very easy and smoothing when the hypertext links are extensively tagged. Unfortunately, the hypertext links between different languages are mostly not being tagged because of the hypertext link is generated by human contributor, mostly monolingual ones. To solve this problem, we design a process flow trying to make it more completely.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Background 2.1 Wikipedia",
"sec_num": "2"
},
{
"text": "The goal of cross-lingual link discovery(CLLD) is trying to find the potential links that are missing between the two different languages. There are three main challenges for the system to overcome. First, the system providing solution on CLLD can proactively recommends a set of words which called anchors. The set of words have higher chances to have their corresponding cross lingual links than other words in the same article. For example, considering different cases as following:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Cross-lingual link discovery",
"sec_num": "2.2"
},
{
"text": "1. Let's go dutch. 2.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Cross-lingual link discovery",
"sec_num": "2.2"
},
{
"text": "A Dutch auction is a type of auction that starts with a high bid.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Cross-lingual link discovery",
"sec_num": "2.2"
},
{
"text": "The system must determine the boundaries between anchor and rest of words, considering the first case above, the word \"dutch\" is meaning to share the money on something instead of meaning some behavior or something related to the country \"Holland\". In other words, the word \"dutch\" should not be chosen as an anchor here and choosing the phase of \"go dutch\" is more significant. Considering the second case above, the word \"Dutch auction\" is an appropriate anchor rather than \"Dutch\".",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Cross-lingual link discovery",
"sec_num": "2.2"
},
{
"text": "After the system identifies these anchors, there must exist many highly ambiguous cases in these anchors and this is the second challenge of CLLD, for example, the anchor Apple can be refer to the link which is related with Apple(Computer Manufacturer), or the link which is related to Apple(Fruit). The system must be able to choosing the most related corresponding links and also ensure the correctness of link discovery.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Cross-lingual link discovery",
"sec_num": "2.2"
},
{
"text": "Once the system can return the most related links of each anchor, there is only one more problem need to solve. In the end of the CLLD flow, the system have to automatically discover the cross-lingual link based on the anchors which generated from previous steps. The system can just use simple parser or crawler to check the content of corresponding wikipedia page or combines some different mechanism to increase the accuracy of link discovery. In this paper, we implement these CLLD steps to help us find the corresponding cross-lingual links and we focus on anchor disambiguation and cross-lingual link discovery, which are both described in Section 3.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Cross-lingual link discovery",
"sec_num": "2.2"
},
{
"text": "In English-to-Chinese cross-lingual link discovery, the goal is to find every potential links in documents. At first, the system searches out potential terms as candidate terms. Overlapping problem happens in this stage, and adequate candidate term selection is required. We propose an similarity-scoring formula to calculate score of relevance. When candidate terms are selected, relevant pages in Chinese Wikipedia need to be linked with these terms. There are some cross-lingual articles in Wikipedia; however, many more links are still missed. (eg. \"Hundred Schools of Thought\" with \"\u8af8\u5b50\u767e\u5bb6\").",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Method and System Description",
"sec_num": "3"
},
{
"text": "To find cross-lingual links in a language, every potential term or phrase is to be listed in the beginning. Here we adopt n-gram tokenizer [5] and Maximum Matching algorithm [6] to segment. For example, assume a sentence \"Magic Johnson is one of the best basketball player in NBA\", in our method , our approach will take \"Magic Johnson\" as an anchor rather than \"Magic\" or \"John\". The system will first examine the longer term in the sentence and exploit the Wikipedia as a anchor look-up table to check whether this long term is meaningful or not.",
"cite_spans": [
{
"start": 139,
"end": 142,
"text": "[5]",
"ref_id": "BIBREF4"
},
{
"start": 174,
"end": 177,
"text": "[6]",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Candidates finding",
"sec_num": "3.1"
},
{
"text": "Many terms in Wikipedia have the same title but different meanings depending on their occurrences in contexts. To address this problem, Wikipedia has already define it as \"Disambiguation\". In our system, we use redirect page, providing disambiguation information and candidate terms, to analysis and select one from terms for users by this system. For instance, a term \"Virus\" is shown in \"A virus is a parasitic agent that is smaller than a bacterium and that can only reproduce after infecting a host cell.\" and \"Virus (clothing), an Israeli clothing brand\"...etc. It indicates users may look out the clothing brand but Wikipedia gives him a virus' definition in biology domain. ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Anchor decision",
"sec_num": "3.2"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "SimilarityScore(D i , D j ) = T ermRecog(D i ) T ermRecog(D j ) T ermRecog(D i ) T ermRecog(D j )",
"eq_num": "(1)"
}
],
"section": "Anchor decision",
"sec_num": "3.2"
},
{
"text": "EQUATION",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [
{
"start": 0,
"end": 8,
"text": "EQUATION",
"ref_id": "EQREF",
"raw_str": "Anchor = max(SimilarityScore(D current , D i )), \u2200i \u2208 candidates",
"eq_num": "(2)"
}
],
"section": "Anchor decision",
"sec_num": "3.2"
},
{
"text": "In our work, we design a content-aware approach to perform auto selection among disambiguation terms. Our design principle is to analyze the input article, especially the source of terms, and use full-featured text search engine with a prepared index file. If a term has the disambiguation property, the system will extract the features from article and search the existed index to decide which term is more likely to the source article.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Anchor decision",
"sec_num": "3.2"
},
{
"text": "In this section, we describe how we translate the anchor first and than how we find the cross lingual Wikipedia link after the translation. There are two main approaches of the translation mechanism, namely Cross-Lingual Link Dictionary and Google Translate. We first use a Cross-Lingual Link Dictionary as the translation scheme, once if Cross-Lingual Link Dictionary can not provide any corresponding translation, Google Translate is then used by the system to discover the corresponding translation from the online Machine Translation mechanism. Google Translate is a state-of-the-art online commercial machine translation scheme, and it is exploited by our system to trying find out some possible translation when there doesn't have any corresponding translation which can be provided by the Cross-Lingual Link Dictionary. With the support by the Google Translate, the system can provide higher translation coverage compared to using Cross-Lingual Link Dictionary only. We will describe the detail about the two translation mechanisms below and will also discuss the missing link recovery approach in the end of this section. ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "English-Chinese Link Discovery",
"sec_num": "3.3"
},
{
"text": "We first describe the Google Translate here because we are going to introduce the translation and missing link recovery within Cross-Lingual Dictionary in the end of this section together.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Google Translate",
"sec_num": "3.4.1"
},
{
"text": "Google Translate has been a famous automatic translation mechanism, one distinguishing feature of this online translator is that it enables users to choose different languages that users want to translate. As for whole sentence translations, the users have a chance to modify the translation sentence once they find the output translation inadequate. As Google collects enough user data of modifying the translation sentence, Google Translator gets higher translation accuracy.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Google Translate",
"sec_num": "3.4.1"
},
{
"text": "Although Google Translate has such special characteristic, it can not providing good accuracy at Anchor translation [7] . However, there is a special characteristic of Google Translate; that is, it can provide more possible translation candidates than previous methods such like Cross-Lingual Link Dictionary. The reason is that Google Translate is tends to adopt a best-effort approach, it aims to provide many translation candidates which enable users to understand what the untranslated sentence might be supposed to mean.",
"cite_spans": [
{
"start": 116,
"end": 119,
"text": "[7]",
"ref_id": "BIBREF6"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Google Translate",
"sec_num": "3.4.1"
},
{
"text": "As a result, we put the lowest translation priority in Google Translate, namely, once the previous method(Cross-Lingual Dictionary) can not find out any possible translation candidates, we will try to get some translation suggested from Google Translate. The main reason is just what we describe above, we want to take a chance to find out the corresponding translation when we do not have any other translation candidate, only to use some anchor translation from Google Translate to find out the corresponding cross-language links.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Google Translate",
"sec_num": "3.4.1"
},
{
"text": "For example, in our Cross-Lingual Link Dictionary, it does not contain the Chinese Translation of \"Freeway\". However, Google Translate can provide some useful Chinese translation like \"\u9ad8 \u9ad8 \u9ad8\u901f \u901f \u901f\u516c \u516c \u516c\u8def \u8def \u8def\", thus we can find the corresponding link of Chinese article page of Wikipedia page at \"http://zh.wikipedia.org/wiki/\".",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Google Translate",
"sec_num": "3.4.1"
},
{
"text": "Wikipedia provides a well formatted dump file for all languages. As a result, we can get the chinese translation from the english dump files and vise-versa. We exploit this property to construct both Chinese-English bilingual link dictionary and an English-Chinese bilingual link dictionary. Furthermore, once the translation in the dictionary has be found, there is a high probability that we can directly discover the link by adding the translated anchor after the specific wikipedia URL(e.g. http://en.wikipedia.org/wiki/Computer_ accessibility), both in English and Chinese. We refer these two dictionaries as the translation dictionaries, one is the English to Chinese (E-C) translation dictionary and the other one is Chinese to English (C-E) translation dictionary. Once we use these two bilingual dictionaries as translation dictionaries, in our case, English-to-Chinese vise versa,we can have a chance to retrieve the link informations bidirectional. The reason is that we have noticed that links for Chinese-to-English are more than English-to-Chinese, because many Chinese editors will add English link for annotation or reference.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Cross-Lingual Link Dictionary",
"sec_num": "3.4.2"
},
{
"text": "On link discovery part, we find out that some links may be missing in one translation dictionary, such as the term \"Flag of Republic of China\" is not able to found any corresponding Chinese translation in E-C translation dictionary. However, we can find the corresponding english translation of chinese term \"\u8af8\u5b50\u767e\u5bb6\" in the C-E translation dictionary, which is the \"Hundred Schools of Thought\".",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Cross-Lingual Link Dictionary",
"sec_num": "3.4.2"
},
{
"text": "There is an additional problem about the English-Chinese dictionary with the Wikipedia disambiguation page. If the anchor which exist in the English-Chinese dictionary is a title of the Wikipedia disambiguation page, then we can not directly get the Chinese translation from the page content of disambiguation page. The reason is that a Wikipedia disambiguation page only contains the possible candidates that are referring to this title.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Cross-Lingual Link Dictionary",
"sec_num": "3.4.2"
},
{
"text": "Fortunately, Wikipedia have a complete dump file format and it provide the redirect information of the disambiguation page. Therefore, we can using the redirect link information to find out the corresponding Chinese translation. The problem may also occur at Chinese Wikipedia disambiguation page, and it can be also solved by redirection information.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Cross-Lingual Link Dictionary",
"sec_num": "3.4.2"
},
{
"text": "We use four articles as evaluation to see the performance of cross-lingual discovery, we first randomly choose four Bilingual news article from Yahoo! News, all terms in the Chinese articles are tagged by two human experts to generate correct answers. We apply two methods, the first method is tagging the English articles with English Wikipedia entries by means of longterm-first algorithm. Those tagged terms are then directly transformed into Chinese Wikipedia entries by original anchored links; the second method is to implement our proposed method, we then compare these two methods to see the coverage rates. As Figure 4 shows, the experiment result shows that our proposed method has 8% coverage rates higher than the that of direct anchor transformation method. ",
"cite_spans": [],
"ref_spans": [
{
"start": 619,
"end": 627,
"text": "Figure 4",
"ref_id": null
}
],
"eq_spans": [],
"section": "Results and Discussion",
"sec_num": "4"
},
{
"text": "In conclusion, we present a system to find potential cross-lingual linked data on articles, trying to discover miss cross-lingual links. The main contribution of our proposed method includes finding anchor and discovering missing cross-lingual links. We have successfully designed a practical system to perform tagging task on real-world articles. and proved that maximum match algorithm has a better performance than the original Wikipedia anchor links transformation. However, there are still issued to be improved for future work. First, the precision of WSD is still low, and second, we can apply machine learning approaches in our method, in which we are convinced that our proposed method might have higher performance in the future.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion",
"sec_num": "5"
},
{
"text": "http://wikipedia.org Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing (ROCLING 2012)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "http://translate.google.com 3 http://lucene.apache.org 4 http://en.wikipedia.org/wiki/Wikipedia Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing (ROCLING 2012)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing(ROCLING 2012)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "Wikify! linking documents to encyclopedic knowledge",
"authors": [
{
"first": "Csomai",
"middle": [],
"last": "Mihalcea",
"suffix": ""
}
],
"year": 2007,
"venue": "sixteenth ACM conference",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Mihalcea and Csomai, \"Wikify! linking documents to encyclopedic knowledge\", in six- teenth ACM conference, 2007.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "Using encyclopedic knowledge for named entity disambiguation",
"authors": [
{
"first": "Pasca",
"middle": [],
"last": "Bunescu",
"suffix": ""
}
],
"year": 2006,
"venue": "EACL",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Bunescu and Pasca, \"Using encyclopedic knowledge for named entity disambiguation\", in EACL, 2006.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "Ukp at crosslink: Anchor text translation for cross-lingual link discovery",
"authors": [
{
"first": "J",
"middle": [],
"last": "Kim",
"suffix": ""
},
{
"first": "I",
"middle": [],
"last": "Gurevych",
"suffix": ""
}
],
"year": 2011,
"venue": "NTCIR-9",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "J. Kim and I. Gurevych, \"Ukp at crosslink: Anchor text translation for cross-lingual link discovery\", in NTCIR-9, 2011.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Hits'graph-based system at the ntcir-9 cross-lingual link discovery task",
"authors": [
{
"first": "A",
"middle": [
"F V"
],
"last": "Nastase",
"suffix": ""
},
{
"first": "M",
"middle": [],
"last": "Strube",
"suffix": ""
}
],
"year": 2011,
"venue": "NTCIR-9",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "A.F.V. Nastase and M. Strube, \"Hits'graph-based system at the ntcir-9 cross-lingual link discovery task\", in NTCIR-9, 2011.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "N-gram based text categorization",
"authors": [
{
"first": "W",
"middle": [
"B"
],
"last": "Cavnar",
"suffix": ""
},
{
"first": "J",
"middle": [
"M"
],
"last": "Trenkle",
"suffix": ""
}
],
"year": 1994,
"venue": "Proceeding of the Symposium on Document Analysis and Information Retrieval",
"volume": "",
"issue": "",
"pages": "161--175",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "W. B. Cavnar and J. M. Trenkle., \"N-gram based text categorization\", in Proceeding of the Symposium on Document Analysis and Information Retrieval. University of Nevada, Las Vegas, 1994, pp. 161-175.",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "An improved parallel algorithm for maximal matching",
"authors": [
{
"first": "Y. Shiloach Amos",
"middle": [],
"last": "Israeli",
"suffix": ""
}
],
"year": 1986,
"venue": "Information Processing Letters",
"volume": "22",
"issue": "",
"pages": "57--60",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Y. Shiloach Amos Israeli, \"An improved parallel algorithm for maximal matching\", in Information Processing Letters, 1986, vol. 22, pp. 57-60.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Exploit wikipedia and name-entity pattern as translation method on chinene-korean cross language multimedia information retrival",
"authors": [
{
"first": "Yu-Chun Wang Richard Tzong-Han Tsai Liang-Pu",
"middle": [],
"last": "Chen",
"suffix": ""
},
{
"first": "Chen-Ting",
"middle": [],
"last": "Chen",
"suffix": ""
}
],
"year": 2009,
"venue": "International Conference on Digital Contents",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Yu-Chun Wang Richard Tzong-Han Tsai Liang-Pu Chen, Chen-Ting Chen, \"Exploit wikipedia and name-entity pattern as translation method on chinene-korean cross language multimedia information retrival\", in International Conference on Digital Contents, 2009.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"text": "Processing flow of our system.",
"num": null,
"uris": null,
"type_str": "figure"
},
"FIGREF1": {
"text": "Flow of our English-Chinese Link Discovery system.",
"num": null,
"uris": null,
"type_str": "figure"
},
"FIGREF2": {
"text": "Results of English to Chinese link discovery.",
"num": null,
"uris": null,
"type_str": "figure"
}
}
}
} |