{ "paper_id": "2020", "header": { "generated_with": "S2ORC 1.0.0", "date_generated": "2023-01-19T07:12:37.163209Z" }, "title": "If You Build Your Own NER Scorer, Non-replicable Results Will Come", "authors": [ { "first": "Constantine", "middle": [], "last": "Lignos", "suffix": "", "affiliation": { "laboratory": "", "institution": "Brandeis University", "location": { "addrLine": "415 South St. Waltham", "postCode": "02453", "region": "MA", "country": "USA" } }, "email": "lignos@brandeis.edu" }, { "first": "Marjan", "middle": [], "last": "Kamyab", "suffix": "", "affiliation": { "laboratory": "", "institution": "Brandeis University", "location": { "addrLine": "415 South St. Waltham", "postCode": "02453", "region": "MA", "country": "USA" } }, "email": "marjankamyab@brandeis.edu" } ], "year": "", "venue": null, "identifiers": {}, "abstract": "We attempt to replicate a named entity recognition (NER) model implemented in a popular toolkit and discover that a critical barrier to doing so is the inconsistent evaluation of improper label sequences. We define these sequences and examine how two scorers differ in their handling of them, finding that one approach produces F1 scores approximately 0.5 points higher on the CoNLL 2003 English development and test sets. We propose best practices to increase the replicability of NER evaluations by increasing transparency regarding the handling of improper label sequences.", "pdf_parse": { "paper_id": "2020", "_pdf_hash": "", "abstract": [ { "text": "We attempt to replicate a named entity recognition (NER) model implemented in a popular toolkit and discover that a critical barrier to doing so is the inconsistent evaluation of improper label sequences. We define these sequences and examine how two scorers differ in their handling of them, finding that one approach produces F1 scores approximately 0.5 points higher on the CoNLL 2003 English development and test sets. We propose best practices to increase the replicability of NER evaluations by increasing transparency regarding the handling of improper label sequences.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Abstract", "sec_num": null } ], "body_text": [ { "text": "The goal of this paper is to demonstrate an issue that complicates the comparison and replication of named entity recognition (NER) systems. Standard F1-based evaluation of NER models in the manner made popular by the CoNLL 2002-3 shared tasks (Tjong Kim Sang, 2002; Tjong Kim Sang and De Meulder, 2003) requires decoding a sequence of per-token labels into entity mentions and computing precision and recall by evaluating the types and spans of the mentions. While there are popular scorer implementations and popular NER toolkits, there is little transparency regarding the exact process that scorers use to decode label sequences.", "cite_spans": [ { "start": 244, "end": 266, "text": "(Tjong Kim Sang, 2002;", "ref_id": "BIBREF4" }, { "start": 267, "end": 303, "text": "Tjong Kim Sang and De Meulder, 2003)", "ref_id": "BIBREF5" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "In the case where all label sequences are properly formed, this lack of transparency should have no impact; all correct decoding processes should produce the same result. However, many systems can produce what we call improper label sequences (see Section 3.2), and different approaches to decoding improper label sequences for evaluation produce different scores. As the processes for label decoding have not been standardized and are often undocumented, comparisons between reported scores may not be fair, and replication of those scores can prove difficult.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "We did not intend to begin a project on NER evaluation reproducibility. The discovery process for our negative result began with three separate research projects encountering the same issue: higher-thanexpected F1 scores for a certain class of models implemented in a popular NER toolkit, NCRF++ .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Our negative result", "sec_num": "2" }, { "text": "One of the authors of this paper attempted to reimplement NCRF++'s models as a learning exercise. The reimplementation yielded similar F1 scores to NCRF++ when a CRF output layer was used but produced lower F1 scores when using softmax output. After the other author found the same result in an independent reimplementation, we turned our attention to the only commonality between our implementations: an open-source external scorer. We consulted with researchers in our lab using NCRF++ for two other projects, and a consistent story began to emerge: when scoring softmax models, NCRF++'s internal scorer produced higher scores than other NER scorers.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Our negative result", "sec_num": "2" }, { "text": "Our negative result was a failure to replicate the performance of NCRF++, and the cause of this failure was that we did not understand its approach to evaluation. NCRF++'s scorer differs from others in how it treats improper label sequences, and its approach consistently produces higher F1 scores for softmax-output models. The handling of these sequences is effectively undefined behavior for an NER scorer; there is no single correct strategy. In this paper, we quantify the impact of those strategies and propose an evaluation approach that would improve reproducibility by requiring explicit, transparent handling of improper label sequences.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Our negative result", "sec_num": "2" }, { "text": "Supplemental material and the resources needed to replicate this study are available at https:// lignos.org/repro-ner.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Our negative result", "sec_num": "2" }, { "text": "3 Entity encoding and decoding", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Our negative result", "sec_num": "2" }, { "text": "As the focus of this venue is on insights from negative results and not the finer points of NER systems, we will first review the process of entity encoding and decoding. Consider the following sentence fragment from the CoNLL 2003 English NER data:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Proper label sequences", "sec_num": "3.1" }, { "text": "[Australian]MISC [Davis Cup]MISC cap- tain [John Newcombe]PER.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Proper label sequences", "sec_num": "3.1" }, { "text": "In this fragment, three entity mentions are annotated. Table 1 shows three well-known approaches to encoding these tokens as a label sequence. Note that in BIO (Begin, Inside, Outside) encoding, every mention begins with a B label; in IOB encoding, mentions begin with I except where necessary to differentiate from the continuation of a preceding same-type mention by using a B label (e.g., I-PER B-PER for two adjacent singletoken names). In the BIOES encoding, singletoken entities use a single S label, and multi-token entities begin with B, end with E, and use I for everything but the first and last tokens. 1", "cite_spans": [], "ref_spans": [ { "start": 55, "end": 62, "text": "Table 1", "ref_id": "TABREF0" } ], "eq_spans": [], "section": "Proper label sequences", "sec_num": "3.1" }, { "text": "This paper is concerned with the implications for evaluation in NER when unexpected label sequences are produced. While there does not appear to be any standard term for this phenomenon, we define an improper label sequence as one where the label sequence does not conform to a sequence of labels allowed by the encoding. Consider the improper label sequences given in Table 2 .", "cite_spans": [], "ref_spans": [ { "start": 369, "end": 376, "text": "Table 2", "ref_id": "TABREF1" } ], "eq_spans": [], "section": "Improper label sequences", "sec_num": "3.2" }, { "text": "In all three cases, it is possible to infer the likely \"intent\" of the system that produced these 1 Historically, it has not been possible to identify with certainty the entity encoding an NER study uses based on the acronym used. In early work there was confusion between the IOB (IOB1, Ramshaw and Marcus, 1995; Tjong Kim Sang and Veenstra, 1999) and BIO (IOB2) encodings, and later similar confusion between BIOES and IOBES. Readers may disagree regarding precisely what entity encodings are signified by these acronyms. We do not discuss BMES and BILOU in this paper and consider them isomorphic to BIOES.", "cite_spans": [ { "start": 288, "end": 313, "text": "Ramshaw and Marcus, 1995;", "ref_id": "BIBREF2" }, { "start": 314, "end": 348, "text": "Tjong Kim Sang and Veenstra, 1999)", "ref_id": "BIBREF6" } ], "ref_spans": [], "eq_spans": [], "section": "Improper label sequences", "sec_num": "3.2" }, { "text": "Labels label sequences. For IOB and BIO encodings, the conlleval scorer (Tjong Kim Sang, 2004) used for the CoNLL 2002-3 NER evaluations (Tjong Kim Sang, 2002; Tjong Kim Sang and De Meulder, 2003) would effectively \"repair\" 2 these label sequences, such that they would produce the same entity mentions as the proper encodings in Table 1 . We refer to this approach to interpreting improper label sequences as CoNLL-style in this paper.", "cite_spans": [ { "start": 72, "end": 94, "text": "(Tjong Kim Sang, 2004)", "ref_id": null }, { "start": 137, "end": 159, "text": "(Tjong Kim Sang, 2002;", "ref_id": "BIBREF4" }, { "start": 160, "end": 196, "text": "Tjong Kim Sang and De Meulder, 2003)", "ref_id": "BIBREF5" } ], "ref_spans": [ { "start": 330, "end": 337, "text": "Table 1", "ref_id": "TABREF0" } ], "eq_spans": [], "section": "Encoding", "sec_num": null }, { "text": "IOB I-MISC B-MISC I-MISC O B-PER I-PER BIO B-MISC B-MISC I-MISC O I-PER I-PER BIOES S-MISC B-MISC I-MISC O B-PER E-PER", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Encoding", "sec_num": null }, { "text": "An alternative would be to interpret the label sequence more strictly. An obvious approach for BIO encoding would be to only begin a mention when a B label was encountered. In this example, only two mentions would be created, as the final two tokens do not have a proper beginning label.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Encoding", "sec_num": null }, { "text": "The conlleval scorer does not support the BIOES/IOBES encodings, but seqeval (chakki, 2019), a Python library that replicates conlleval's BIO decoding, supports them and uses similar CoNLL-style repair logic for improper BIOES/IOBES sequences. We return to the issue of BIOES/IOBES decoding in Section 5.1.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Encoding", "sec_num": null }, { "text": "Since at least HMM-based NER systems of the late 90s, sequence NER models have identified the most likely label sequence by taking into account the relationship between a label and preceding labels. The most common modern approach to modeling this relationship is to use a conditional random field (CRF), which can learn to avoid producing improper label sequences or be forced to, either through manual manipulation of its weights or forbidding improper transitions entirely.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Softmax output", "sec_num": "3.3" }, { "text": "With the advent of neural models capable of capturing substantial contextual information before the output layer, it is now feasible to create relatively good NER models without a CRF by using softmax to select the highest-scoring label for each token independently, which is significantly faster. While the model will still indirectly learn to prefer proper label sequences, there is no explicit representation of the sequential relationship between labels. When sufficiently trained, a model with CRF output rarely produces improper label sequences, but softmax-output models do so more frequently.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Softmax output", "sec_num": "3.3" }, { "text": "The goal of our experiments is to estimate the increase in F1 scores that can be attributed to use of NCRF++'s internal decoding compared to CoNLLstyle decoding when evaluating softmax-output models. We trained models using NCRF++ after modifying it to additionally use an external CoNLLstyle scorer, seqeval. 3 We report scores from the internal and external scorer. We use a bidirectional LSTM architecture at the word level and test multiple character-level architectures: a bi-LSTM, CNN, and no character-level representation. We use the same hyperparameters, pretrained word embedding (GLoVe 100d), and data (CoNLL 2003 English) used by . Each configuration was run ten times using different random initializations (seeds 0-9). The test set was evaluated using the model from the epoch that attained the highest development set F1 (as scored internally by NCRF++) during training. Table 3 gives F1 scores for entity mentions for BIOES and BIO encodings across all characterlevel architectures, including values previously reported by . When using NCRF++'s internal scorer, our results are close to the those previously reported. However, evaluating the same output with an external scorer leads to lower scores. Table 4 reports the distribution of the increase in F1 scores (NCRF++'s internal score minus the external score) for each system output across encodings and evaluation sets. We computed how much NCRF++'s scoring procedure increases F1 for each run's output and then average across all runs (character-level architectures and random initializations). To demonstrate the statistical reliability of this increase, we performed a Wilcoxon signed-rank test-a non-parametric version of the paired t-test-for each combination of encoding 3 We selected seqeval after reviewing its label decoding procedure and confirming that produces the same scores for BIO as conlleval. We chose it because it supports both BIO and BIOES encoding and provides greater numerical precision than conlleval. and evaluation set. All four p-values were below 0.0001, and the 95% confidence intervals of the differences were .49-.58 (BIOES) and .38-.43 (BIO) for development, .54-63 (BIOES) and .48-.56 (BIO) for test. In summary, for softmax models, the NCRF++ internal scorer produces scores approximately half a point of F1 higher.", "cite_spans": [ { "start": 310, "end": 311, "text": "3", "ref_id": null }, { "start": 1748, "end": 1749, "text": "3", "ref_id": null } ], "ref_spans": [ { "start": 886, "end": 893, "text": "Table 3", "ref_id": null }, { "start": 1217, "end": 1224, "text": "Table 4", "ref_id": null } ], "eq_spans": [], "section": "Results", "sec_num": "4" }, { "text": "All scores reported so far have been from converged models. It is also of interest to explore what the increase in scores looks like during training. Figure 1 gives the increase in development set F1 scores across all training epochs for all configurations we ran, displaying 3,000 points per encoding. Early in training, NCRF++'s internal scorer can produce F1 scores several points higher than a CoNLL-style scorer, presumably due to producing a high number of improper label sequences preconvergence. Crucially, in all 6,000 development set epochs we evaluated, NCRF++'s internal scorer always reported a higher F1 than seqeval in our evaluation of softmax-output models.", "cite_spans": [], "ref_spans": [ { "start": 150, "end": 158, "text": "Figure 1", "ref_id": "FIGREF1" } ], "eq_spans": [], "section": "Results", "sec_num": "4" }, { "text": "Before we discuss the insights gained from our study, it is important contextualize our findings. Our goals are to motivate the establishment of standard evaluation practices for NER, explain the impact of improper label sequence decoding, and encourage authors to be transparent about their approach to evaluation.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Analysis", "sec_num": "5.1" }, { "text": "We must emphasize that we are not claiming that the way that NCRF++'s scorer decodes improper label sequences is incorrect; it is one of many possible ways of doing so, and we do not wish to single out NCRF++ or its authors specifically for criticism. NCRF++'s approach to decoding improper label sequences is, however, different than the popular approach defined by the widely-used conlleval scorer for the CoNLL 2002-3 NER shared tasks (later faithfully reimplemented and extended by seqeval), and thus one must exercise caution when comparing scores it generates to those of other scorers.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Analysis", "sec_num": "5.1" }, { "text": "While we find the NCRF++ entity mention decoder difficult to fully understand, our inspection of the code leads us to believe that it will only begin a mention when the proper tag-B for BIO, B or S for BIOES-is supplied. It is effectively removing some improper label sequences by treating some labels as if they were O. Table 3 : Means and standard deviation of test set F1 for each tested configuration and from previously reported results (Yang et al., 2018, Table 4 ). Empty cells indicate configurations without previously reported scores. N gives the number of runs used to compute each value in the row (e.g., each mean was computed over N values). Table 4 : Means and standard deviations for softmax models of the increase in F1 due to using NCRF++'s internal scorer.", "cite_spans": [ { "start": 442, "end": 469, "text": "(Yang et al., 2018, Table 4", "ref_id": null } ], "ref_spans": [ { "start": 321, "end": 328, "text": "Table 3", "ref_id": null }, { "start": 656, "end": 663, "text": "Table 4", "ref_id": null } ], "eq_spans": [], "section": "Analysis", "sec_num": "5.1" }, { "text": "Why would it be advantageous for a system to decode in this way? Unlike in Table 2 , where repairing the improper sequences would result in correct answers, many improper sequences represent generalization errors. For example, consider a system using a BIO encoding that observed Pat Jones as B-PER I-PER frequently in training and must predict labels for Charity Jones, annotated as a person in the gold standard. If it has never seen the token Charity with the label B-PER and has seen Jones frequently with the label I-PER, it might predict the tag sequence O I-PER. A lenient CoNLL-style decoder would decode this as a mention of type PER for the token Jones, which would result in a false alarm (precision error) and a miss (recall error).", "cite_spans": [], "ref_spans": [ { "start": 75, "end": 82, "text": "Table 2", "ref_id": "TABREF1" } ], "eq_spans": [], "section": "Analysis", "sec_num": "5.1" }, { "text": "However, a stricter BIO decoder like NCRF++'s would predict no mentions, resulting in a miss but no false alarm. For softmax output, it appears to be universally advantageous to decode this way; for the 6,000 epochs we scored using the development set, NCRF++'s scorer always gave a higher F1. Our examples have focused on decoding BIO sequences, partly because conlleval set a standard approach to doing so almost two decades ago.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Analysis", "sec_num": "5.1" }, { "text": "Repairing improper BIOES label sequences is much more complex; for example, in the sequences B-PER E-PER E-PER and S-PER I-PER E-PER, how many entity mentions should be created? There is not a single answer, but seqeval implements an approach in the spirit of what conlleval does for BIO, and NCRF++ appears to implement a stricter decoder.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Analysis", "sec_num": "5.1" }, { "text": "While we were not able to find any published or unpublished papers specifically discussing the effect of decoding improper label sequences on F1, as we prepared this paper for submission we discovered a closed GitHub issue 4 opened by Mike Kroutikov pointing out that the way NCRF++ decodes label sequences does not match other NER systems. Reviewing this issue and other issues opened against NCRF++, it is clear that we are not the first to identify its departure from CoNLLstyle label decoding. In addition to opening an issue, Kroutikov (2019) blogged about the many potential ways to decode BIOES and examined the approaches taken by other NER implementations.", "cite_spans": [ { "start": 531, "end": 547, "text": "Kroutikov (2019)", "ref_id": "BIBREF0" } ], "ref_spans": [], "eq_spans": [], "section": "Analysis", "sec_num": "5.1" }, { "text": "Our findings raise the question of whether NCRF++'s internal scorer also differs from the external scorer when evaluating models that use CRF output. We performed a post-hoc study where we repeated the same set of experiments we performed for softmax-output models with ones that used a CRF. While NCRF++'s scorer does not give the exact same scores as the external scorer, the increase in F1 attributable to NCRF++'s scoring procedure is quite small when measured on the development set (BIOES 0.0032 \u00b1 0.0040; BIO 0.0072 \u00b1 0.0073), and test set (BIOES \u22120.015 \u00b1 0.020; BIO 0.0040 \u00b1 0.0051). The experiments using BIOES encoding and evaluating on the test set are the only ones we performed in which on average NCRF++'s internal scorer produced lower scores than the external one.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Analysis", "sec_num": "5.1" }, { "text": "What insights have we gained from our negative result? First, we have clarified how well softmaxoutput NER models perform when evaluated using CoNLL-style label decoding, which gives lower scores than NCRF++'s internal scorer. The issue of improper label decoding is unlikely to significantly affect any state of the art results, which do not generally use softmax output, but may affect decision-making for NER system designers exploring whether adding a CRF to a system is worth the performance penalty.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Insights", "sec_num": "5.2" }, { "text": "Second, we believe our study will help users of a popular toolkit-which received the COL-ING 2018 \"Most reproducible\" best paper awardunderstand how it computes scores and why attempts at replication will fail if they do not also replicate NCRF++'s approach to improper label sequence decoding.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Insights", "sec_num": "5.2" }, { "text": "More broadly, the insights from this negative result highlight for all NLP researchers the importance of using a standard evaluation procedure. When possible, using widely-used, welldocumented scorers enables fair comparisons of scores across systems.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Insights", "sec_num": "5.2" }, { "text": "Our study leads us to propose a vision for NER evaluation as follows: 1. We should have a well-tested, well-documented, open-source scorer which has been developed independently of any particular model. This scorer should only accept properly-formed label sequences, avoiding the question of the \"right\" way to decode improper label sequences. 2. The scorer should be accompanied by implementations of standard processes for converting improper label sequences into proper ones. These approaches should include the CoNLL-style approach and a more strict one, like NCRF++'s. 3. Work which wants to use an alternative approach to converting improper label sequences to proper ones should contain a documented, replicable process for doing so.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A vision for NER evaluation", "sec_num": "6" }, { "text": "This procedure separates the process of scoring from the process of interpreting improper label sequences. It also suggests a new research avenue of designing methods for optimally converting improper label sequences into proper ones.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A vision for NER evaluation", "sec_num": "6" }, { "text": "Until the tools required for our vision-perhaps something like SacreBLEU (Post, 2018)-are developed, we recommend seqeval as the best solution for NER scoring. It is an easy-to-inspect, faithful reimplementation of conlleval. Unlike conlleval, it can be called directly from Python and does not truncate scores, avoiding systematic downward bias when aggregating them.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A vision for NER evaluation", "sec_num": "6" }, { "text": "Regarding extensions to this study, while we evaluated on the CoNLL 2003 English data to compare with , evaluating in more languages is essential. Looking beyond the other CoNLL 2002-3 languages, testing against smaller annotated data sets and in lower-resourced languages may reveal more complexity to the problem of decoding improper sequences.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A vision for NER evaluation", "sec_num": "6" }, { "text": "It is not clear whether it was an explicit design goal to repair these sequences or they are simply the consequence of implementing a universal decoder for IOB and BIO encodings. Readers interested in examining the decoding logic should inspect the startOfChunk and endOfChunk functions.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "https://github.com/jiesutd/NCRFpp/ issues/87", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null } ], "back_matter": [ { "text": "Thanks to Elena\u00c1lvarez Mellado and Jingxuan Tu, whose usage of NCRF++ for other projects helped us identify the negative result featured here. Thank you to Chester Palen-Michel and three anonymous reviewers for providing feedback on this paper.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Acknowledgments", "sec_num": null } ], "bib_entries": { "BIBREF0": { "ref_id": "b0", "title": "7776 ways to compute F1 for an NER task", "authors": [ { "first": "Mike", "middle": [], "last": "Kroutikov", "suffix": "" } ], "year": 2019, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Mike Kroutikov. 2019. 7776 ways to compute F1 for an NER task. http://blog.innodatalabs. com/7776_ways_to_compute_f1_for_ner_ task/.", "links": null }, "BIBREF1": { "ref_id": "b1", "title": "A call for clarity in reporting BLEU scores", "authors": [ { "first": "Matt", "middle": [], "last": "Post", "suffix": "" } ], "year": 2018, "venue": "Proceedings of the Third Conference on Machine Translation: Research Papers", "volume": "", "issue": "", "pages": "186--191", "other_ids": { "DOI": [ "10.18653/v1/W18-6319" ] }, "num": null, "urls": [], "raw_text": "Matt Post. 2018. A call for clarity in reporting BLEU scores. In Proceedings of the Third Conference on Machine Translation: Research Papers, pages 186- 191, Brussels, Belgium. Association for Computa- tional Linguistics.", "links": null }, "BIBREF2": { "ref_id": "b2", "title": "Text chunking using transformation-based learning", "authors": [ { "first": "Lance", "middle": [], "last": "Ramshaw", "suffix": "" }, { "first": "Mitch", "middle": [], "last": "Marcus", "suffix": "" } ], "year": 1995, "venue": "Third Workshop on Very Large Corpora", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Lance Ramshaw and Mitch Marcus. 1995. Text chunk- ing using transformation-based learning. In Third Workshop on Very Large Corpora.", "links": null }, "BIBREF4": { "ref_id": "b4", "title": "Introduction to the CoNLL-2002 shared task: Language-independent named entity recognition", "authors": [ { "first": "Erik", "middle": [ "F" ], "last": "", "suffix": "" }, { "first": "Tjong Kim", "middle": [], "last": "Sang", "suffix": "" } ], "year": 2002, "venue": "COLING-02: The 6th Conference on Natural Language Learning", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Erik F. Tjong Kim Sang. 2002. Introduction to the CoNLL-2002 shared task: Language-independent named entity recognition. In COLING-02: The 6th Conference on Natural Language Learning 2002 (CoNLL-2002).", "links": null }, "BIBREF5": { "ref_id": "b5", "title": "Introduction to the CoNLL-2003 shared task: Language-independent named entity recognition", "authors": [ { "first": "Erik", "middle": [ "F" ], "last": "Tjong", "suffix": "" }, { "first": "Kim", "middle": [], "last": "Sang", "suffix": "" }, { "first": "Fien", "middle": [], "last": "De Meulder", "suffix": "" } ], "year": 2003, "venue": "Proceedings of the Seventh Conference on Natural Language Learning at HLT-NAACL 2003", "volume": "", "issue": "", "pages": "142--147", "other_ids": {}, "num": null, "urls": [], "raw_text": "Erik F. Tjong Kim Sang and Fien De Meulder. 2003. Introduction to the CoNLL-2003 shared task: Language-independent named entity recognition. In Proceedings of the Seventh Conference on Natu- ral Language Learning at HLT-NAACL 2003, pages 142-147.", "links": null }, "BIBREF6": { "ref_id": "b6", "title": "Representing text chunks", "authors": [ { "first": "Erik", "middle": [ "F" ], "last": "Tjong", "suffix": "" }, { "first": "Kim", "middle": [], "last": "Sang", "suffix": "" }, { "first": "Jorn", "middle": [], "last": "Veenstra", "suffix": "" } ], "year": 1999, "venue": "Ninth Conference of the European Chapter of the Association for Computational Linguistics", "volume": "", "issue": "", "pages": "173--179", "other_ids": {}, "num": null, "urls": [], "raw_text": "Erik F. Tjong Kim Sang and Jorn Veenstra. 1999. Rep- resenting text chunks. In Ninth Conference of the European Chapter of the Association for Computa- tional Linguistics, pages 173-179, Bergen, Norway. Association for Computational Linguistics.", "links": null }, "BIBREF7": { "ref_id": "b7", "title": "Design challenges and misconceptions in neural sequence labeling", "authors": [ { "first": "Jie", "middle": [], "last": "Yang", "suffix": "" }, { "first": "Shuailong", "middle": [], "last": "Liang", "suffix": "" }, { "first": "Yue", "middle": [], "last": "Zhang", "suffix": "" } ], "year": 2018, "venue": "Proceedings of the 27th International Conference on Computational Linguistics", "volume": "", "issue": "", "pages": "3879--3889", "other_ids": {}, "num": null, "urls": [], "raw_text": "Jie Yang, Shuailong Liang, and Yue Zhang. 2018. De- sign challenges and misconceptions in neural se- quence labeling. In Proceedings of the 27th Inter- national Conference on Computational Linguistics, pages 3879-3889, Santa Fe, New Mexico, USA. As- sociation for Computational Linguistics.", "links": null }, "BIBREF8": { "ref_id": "b8", "title": "NCRF++: An opensource neural sequence labeling toolkit", "authors": [ { "first": "Jie", "middle": [], "last": "Yang", "suffix": "" }, { "first": "Yue", "middle": [], "last": "Zhang", "suffix": "" } ], "year": 2018, "venue": "Proceedings of ACL 2018, System Demonstrations", "volume": "", "issue": "", "pages": "74--79", "other_ids": { "DOI": [ "10.18653/v1/P18-4013" ] }, "num": null, "urls": [], "raw_text": "Jie Yang and Yue Zhang. 2018. NCRF++: An open- source neural sequence labeling toolkit. In Proceed- ings of ACL 2018, System Demonstrations, pages 74-79, Melbourne, Australia. Association for Com- putational Linguistics.", "links": null } }, "ref_entries": { "FIGREF0": { "type_str": "figure", "uris": null, "text": "B-MISC I-MISC O I-PER I-PER BIO B-MISC B-MISC I-MISC O B-PER I-PER BIOES S-MISC B-MISC E-MISC O B-PER E-PER", "num": null }, "FIGREF1": { "type_str": "figure", "uris": null, "text": "Increase in development set F1 due to NCRF++'s internal scorer for softmax output models across all training epochs, with a local regression (LOESS) fit line.", "num": null }, "TABREF0": { "num": null, "content": "", "type_str": "table", "html": null, "text": "Proper entity encodings for the tokens of the string Australian Davis Cup captain John Newcombe." }, "TABREF1": { "num": null, "content": "
", "type_str": "table", "html": null, "text": "Improper entity encodings for the tokens of the string Australian Davis Cup captain John Newcombe, with improper labels identified using bold." }, "TABREF2": { "num": null, "content": "
BIOESBIO
SourceN ScorerNo Char.
", "type_str": "table", "html": null, "text": "Char. LSTM Char. CNN No Char. Char. LSTM Char. CNN Reported 5 NCRF++ 88.49 \u00b1 .17 90.77 \u00b1 .06 90.60 \u00b1 .11 ---Reproduction 10 NCRF++ 88.78 \u00b1 .28 90.76 \u00b1 .12 90.62 \u00b1 .15 88.41 \u00b1 .21 90.49 \u00b1 .15 90.46 \u00b1 .58 Reproduction 10 External 88.21 \u00b1 .27 90.20 \u00b1 .09 89.99 \u00b1 .19 87.98 \u00b1 .21 89.93 \u00b1 .19 89.88 \u00b1 .21" } } } }