gem_id
stringlengths 37
41
| paper_id
stringlengths 3
4
| paper_title
stringlengths 19
183
| paper_abstract
stringlengths 168
1.38k
| paper_content
sequence | paper_headers
sequence | slide_id
stringlengths 37
41
| slide_title
stringlengths 2
85
| slide_content_text
stringlengths 11
2.55k
| target
stringlengths 11
2.55k
| references
list |
---|---|---|---|---|---|---|---|---|---|---|
GEM-SciDuet-train-1#paper-954#slide-0 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-0 | Syntax in Statistical Machine Translation | Translation Model vs Language Model
Syntactic LM Decoder Integration Results Questions? | Translation Model vs Language Model
Syntactic LM Decoder Integration Results Questions? | [] |
GEM-SciDuet-train-1#paper-954#slide-1 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-1 | Syntax in the Language Model | Translation Model vs Language Model
Syntactic LM Decoder Integration Results Questions?
An incremental syntactic language model uses an incremental statistical parser to define a probability model over the dependency or phrase structure of target language strings.
Phrase-based decoder produces translation in the target language incrementally from left-to-right
Phrase-based syntactic LM parser should parse target language hypotheses incrementally from left-to-right
Galley & Manning (2009) obtained 1-best dependency parse using a greedy dependency parser
We use a standard HHMM parser (Schuler et al., 2010)
Engineering simple model, equivalent to PPDA
Algorithmic elegant fit into phrase-based decoder
Cognitive nice psycholinguistic properties | Translation Model vs Language Model
Syntactic LM Decoder Integration Results Questions?
An incremental syntactic language model uses an incremental statistical parser to define a probability model over the dependency or phrase structure of target language strings.
Phrase-based decoder produces translation in the target language incrementally from left-to-right
Phrase-based syntactic LM parser should parse target language hypotheses incrementally from left-to-right
Galley & Manning (2009) obtained 1-best dependency parse using a greedy dependency parser
We use a standard HHMM parser (Schuler et al., 2010)
Engineering simple model, equivalent to PPDA
Algorithmic elegant fit into phrase-based decoder
Cognitive nice psycholinguistic properties | [] |
GEM-SciDuet-train-1#paper-954#slide-2 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-2 | Incremental Parsing | DT NN VP PP
The president VB NP IN NP
meets DT NN on Friday NP/NN NN VP/NP DT board
Motivation Decoder Integration Results Questions?
the president VB NP VP/NN
Transform right-expanding sequences of constituents into left-expanding sequences of incomplete constituents
NP VP S/NP NP
the board DT president VB the
Incomplete constituents can be processed incrementally using a
Hierarchical Hidden Markov Model parser. (Murphy & Paskin, 2001; Schuler et al. | DT NN VP PP
The president VB NP IN NP
meets DT NN on Friday NP/NN NN VP/NP DT board
Motivation Decoder Integration Results Questions?
the president VB NP VP/NN
Transform right-expanding sequences of constituents into left-expanding sequences of incomplete constituents
NP VP S/NP NP
the board DT president VB the
Incomplete constituents can be processed incrementally using a
Hierarchical Hidden Markov Model parser. (Murphy & Paskin, 2001; Schuler et al. | [] |
GEM-SciDuet-train-1#paper-954#slide-3 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-3 | Incremental Parsing using HHMM Schuler et al 2010 | Hierarchical Hidden Markov Model
Circles denote hidden random variables
Edges denote conditional dependencies
NP/NN NN VP/NP DT board
Isomorphic Tree Path DT president VB the
Shaded circles denote observed values
Motivation Decoder Integration Results Questions?
Analogous to Maximally Incremental
e1 =The e2 =president e3 =meets e4 =the e5 =board e =on e7 =Friday
Push-Down Automata NP VP/NN NN | Hierarchical Hidden Markov Model
Circles denote hidden random variables
Edges denote conditional dependencies
NP/NN NN VP/NP DT board
Isomorphic Tree Path DT president VB the
Shaded circles denote observed values
Motivation Decoder Integration Results Questions?
Analogous to Maximally Incremental
e1 =The e2 =president e3 =meets e4 =the e5 =board e =on e7 =Friday
Push-Down Automata NP VP/NN NN | [] |
GEM-SciDuet-train-1#paper-954#slide-4 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-4 | Phrase Based Translation | Der Prasident trifft am Freitag den Vorstand
The president meets the board on Friday
s president president Friday
s that that president Obama met
AAAAAA EAAAAA EEAAAA EEIAAA
s s the the president president meets
Stack Stack Stack Stack
Motivation Syntactic LM Results Questions? | Der Prasident trifft am Freitag den Vorstand
The president meets the board on Friday
s president president Friday
s that that president Obama met
AAAAAA EAAAAA EEAAAA EEIAAA
s s the the president president meets
Stack Stack Stack Stack
Motivation Syntactic LM Results Questions? | [] |
GEM-SciDuet-train-1#paper-954#slide-5 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-5 | Phrase Based Translation with Syntactic LM | represents parses of the partial translation at node h in stack t
s president president Friday
s that that president Obama met
AAAAAA EAAAAA EEAAAA EEIAAA
s s the the president president meets
Stack Stack Stack Stack
Motivation Syntactic LM Results Questions? | represents parses of the partial translation at node h in stack t
s president president Friday
s that that president Obama met
AAAAAA EAAAAA EEAAAA EEIAAA
s s the the president president meets
Stack Stack Stack Stack
Motivation Syntactic LM Results Questions? | [] |
GEM-SciDuet-train-1#paper-954#slide-6 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-6 | Integrate Parser into Phrase based Decoder | EAAAAA EEAAAA EEIAAA EEIIAA
s the the president president meets meets the
Motivation Syntactic LM Results Questions?
president meets the board | EAAAAA EEAAAA EEIAAA EEIIAA
s the the president president meets meets the
Motivation Syntactic LM Results Questions?
president meets the board | [] |
GEM-SciDuet-train-1#paper-954#slide-7 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-7 | Direct Maximum Entropy Model of Translation | e argmax exp jhj(e,f)
h Distortion model n-gram LM
Set of j feature weights
Syntactic LM P( th)
AAAAAA EAAAAA EEAAAA EEIAAA
s s the the president president meets
Stack Stack Stack Stack
Motivation Syntactic LM Results Questions? | e argmax exp jhj(e,f)
h Distortion model n-gram LM
Set of j feature weights
Syntactic LM P( th)
AAAAAA EAAAAA EEAAAA EEIAAA
s s the the president president meets
Stack Stack Stack Stack
Motivation Syntactic LM Results Questions? | [] |
GEM-SciDuet-train-1#paper-954#slide-8 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-8 | Does an Incremental Syntactic LM Help Translation | but will it make my BLEU score go up?
Motivation Syntactic LM Decoder Integration Questions?
Moses with LM(s) BLEU
Using n-gram LM only
Using n-gram LM + Syntactic LM
NIST OpenMT 2008 Urdu-English data set
Moses with standard phrase-based translation model
Tuning and testing restricted to sentences 20 words long
Results reported on devtest set
n-gram LM is WSJ 5-gram LM | but will it make my BLEU score go up?
Motivation Syntactic LM Decoder Integration Questions?
Moses with LM(s) BLEU
Using n-gram LM only
Using n-gram LM + Syntactic LM
NIST OpenMT 2008 Urdu-English data set
Moses with standard phrase-based translation model
Tuning and testing restricted to sentences 20 words long
Results reported on devtest set
n-gram LM is WSJ 5-gram LM | [] |
GEM-SciDuet-train-1#paper-954#slide-9 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-9 | Perplexity Results | Language models trained on WSJ Treebank corpus
Motivation Syntactic LM Decoder Integration Questions?
WSJ 5-gram + WSJ SynLM
...and n-gram model for larger English Gigaword corpus.
Gigaword 5-gram + WSJ SynLM | Language models trained on WSJ Treebank corpus
Motivation Syntactic LM Decoder Integration Questions?
WSJ 5-gram + WSJ SynLM
...and n-gram model for larger English Gigaword corpus.
Gigaword 5-gram + WSJ SynLM | [] |
GEM-SciDuet-train-1#paper-954#slide-10 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-10 | Summary | Straightforward general framework for incorporating any
Incremental Syntactic LM into Phrase-based Translation
We used an Incremental HHMM Parser as Syntactic LM
Syntactic LM shows substantial decrease in perplexity on out-of-domain data over n-gram LM when trained on same data
Syntactic LM interpolated with n-gram LM shows even greater decrease in perplexity on both in-domain and out-of-domain data, even when n-gram LM is trained on substantially larger corpus
+1 BLEU on Urdu-English task with Syntactic LM
All code is open source and integrated into Moses
Motivation Syntactic LM Decoder Integration Results | Straightforward general framework for incorporating any
Incremental Syntactic LM into Phrase-based Translation
We used an Incremental HHMM Parser as Syntactic LM
Syntactic LM shows substantial decrease in perplexity on out-of-domain data over n-gram LM when trained on same data
Syntactic LM interpolated with n-gram LM shows even greater decrease in perplexity on both in-domain and out-of-domain data, even when n-gram LM is trained on substantially larger corpus
+1 BLEU on Urdu-English task with Syntactic LM
All code is open source and integrated into Moses
Motivation Syntactic LM Decoder Integration Results | [] |
GEM-SciDuet-train-1#paper-954#slide-11 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-11 | This looks a lot like CCG | Our parser performs some CCG-style operations:
Type raising in conjunction with forward function composition
Motivation Syntactic LM Decoder Integration Results | Our parser performs some CCG-style operations:
Type raising in conjunction with forward function composition
Motivation Syntactic LM Decoder Integration Results | [] |
GEM-SciDuet-train-1#paper-954#slide-12 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-12 | Why not just use CCG | No probablistic version of incremental CCG
Our parser is constrained
(we dont have backward composition)
We do use those components of CCG (forward function application and forward function composition) which are useful for probabilistic incremental parsing
Motivation Syntactic LM Decoder Integration Results | No probablistic version of incremental CCG
Our parser is constrained
(we dont have backward composition)
We do use those components of CCG (forward function application and forward function composition) which are useful for probabilistic incremental parsing
Motivation Syntactic LM Decoder Integration Results | [] |
GEM-SciDuet-train-1#paper-954#slide-13 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-13 | Speed Results | Mean per-sentence decoding time
Parser beam sizes are indicated for the syntactic LM
Parser runs in linear time, but were parsing all paths through the Moses lattice as they are generated by the decoder
More informed pruning, but slower decoding
Motivation Syntactic LM Decoder Integration Results | Mean per-sentence decoding time
Parser beam sizes are indicated for the syntactic LM
Parser runs in linear time, but were parsing all paths through the Moses lattice as they are generated by the decoder
More informed pruning, but slower decoding
Motivation Syntactic LM Decoder Integration Results | [] |
GEM-SciDuet-train-1#paper-954#slide-14 | 954 | Incremental Syntactic Language Models for Phrase-based Translation | This paper describes a novel technique for incorporating syntactic knowledge into phrasebased machine translation through incremental syntactic parsing. Bottom-up and topdown parsers typically require a completed string as input. This requirement makes it difficult to incorporate them into phrase-based translation, which generates partial hypothesized translations from left-to-right. Incremental syntactic language models score sentences in a similar left-to-right fashion, and are therefore a good mechanism for incorporating syntax into phrase-based translation. We give a formal definition of one such lineartime syntactic language model, detail its relation to phrase-based decoding, and integrate the model with the Moses phrase-based translation system. We present empirical results on a constrained Urdu-English translation task that demonstrate a significant BLEU score improvement and a large decrease in perplexity. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158
],
"paper_content_text": [
"Introduction Early work in statistical machine translation viewed translation as a noisy channel process comprised of a translation model, which functioned to posit adequate translations of source language words, and a target language model, which guided the fluency of generated target language strings (Brown et al., This research was supported by NSF CAREER/PECASE award 0447685, NSF grant IIS-0713448, and the European Commission through the EuroMatrixPlus project.",
"Opinions, interpretations, conclusions, and recommendations are those of the authors and are not necessarily endorsed by the sponsors or the United States Air Force.",
"Cleared for public release (Case Number 88ABW-2010-6489) on 10 Dec 2010.",
"1990).",
"Drawing on earlier successes in speech recognition, research in statistical machine translation has effectively used n-gram word sequence models as language models.",
"Modern phrase-based translation using large scale n-gram language models generally performs well in terms of lexical choice, but still often produces ungrammatical output.",
"Syntactic parsing may help produce more grammatical output by better modeling structural relationships and long-distance dependencies.",
"Bottom-up and top-down parsers typically require a completed string as input; this requirement makes it difficult to incorporate these parsers into phrase-based translation, which generates hypothesized translations incrementally, from left-to-right.",
"1 As a workaround, parsers can rerank the translated output of translation systems (Och et al., 2004) .",
"On the other hand, incremental parsers (Roark, 2001; Henderson, 2004; Schuler et al., 2010; Huang and Sagae, 2010) process input in a straightforward left-to-right manner.",
"We observe that incremental parsers, used as structured language models, provide an appropriate algorithmic match to incremental phrase-based decoding.",
"We directly integrate incremental syntactic parsing into phrase-based translation.",
"This approach re-exerts the role of the language model as a mechanism for encouraging syntactically fluent translations.",
"The contributions of this work are as follows: • A novel method for integrating syntactic LMs into phrase-based translation ( §3) • A formal definition of an incremental parser for statistical MT that can run in linear-time ( §4) • Integration with Moses ( §5) along with empirical results for perplexity and significant translation score improvement on a constrained Urdu-English task ( §6) Related Work Neither phrase-based (Koehn et al., 2003) nor hierarchical phrase-based translation (Chiang, 2005) take explicit advantage of the syntactic structure of either source or target language.",
"The translation models in these techniques define phrases as contiguous word sequences (with gaps allowed in the case of hierarchical phrases) which may or may not correspond to any linguistic constituent.",
"Early work in statistical phrase-based translation considered whether restricting translation models to use only syntactically well-formed constituents might improve translation quality (Koehn et al., 2003) but found such restrictions failed to improve translation quality.",
"Significant research has examined the extent to which syntax can be usefully incorporated into statistical tree-based translation models: string-to-tree (Yamada and Knight, 2001; Gildea, 2003; Imamura et al., 2004; Galley et al., 2004; Graehl and Knight, 2004; Melamed, 2004; Galley et al., 2006; Huang et al., 2006; Shen et al., 2008) , tree-to-string (Liu et al., 2006; Liu et al., 2007; Huang and Mi, 2010) , tree-to-tree (Abeillé et al., 1990; Shieber and Schabes, 1990; Poutsma, 1998; Eisner, 2003; Shieber, 2004; Cowan et al., 2006; Nesson et al., 2006; Zhang et al., 2007; DeNeefe et al., 2007; DeNeefe and Knight, 2009; Liu et al., 2009; Chiang, 2010) , and treelet (Ding and Palmer, 2005; Quirk et al., 2005) techniques use syntactic information to inform the translation model.",
"Recent work has shown that parsing-based machine translation using syntax-augmented (Zollmann and Venugopal, 2006) hierarchical translation grammars with rich nonterminal sets can demonstrate substantial gains over hierarchical grammars for certain language pairs (Baker et al., 2009) .",
"In contrast to the above tree-based translation models, our approach maintains a standard (non-syntactic) phrase-based translation model.",
"Instead, we incorporate syntax into the language model.",
"Traditional approaches to language models in speech recognition and statistical machine translation focus on the use of n-grams, which provide a simple finite-state model approximation of the target language.",
"Chelba and Jelinek (1998) proposed that syntactic structure could be used as an alternative technique in language modeling.",
"This insight has been explored in the context of speech recognition (Chelba and Jelinek, 2000; Collins et al., 2005) .",
"Hassan et al.",
"(2007) and use supertag n-gram LMs.",
"Syntactic language models have also been explored with tree-based translation models.",
"Charniak et al.",
"(2003) use syntactic language models to rescore the output of a tree-based translation system.",
"Post and Gildea (2008) investigate the integration of parsers as syntactic language models during binary bracketing transduction translation (Wu, 1997) ; under these conditions, both syntactic phrase-structure and dependency parsing language models were found to improve oracle-best translations, but did not improve actual translation results.",
"Post and Gildea (2009) use tree substitution grammar parsing for language modeling, but do not use this language model in a translation system.",
"Our work, in contrast to the above approaches, explores the use of incremental syntactic language models in conjunction with phrase-based translation models.",
"Our syntactic language model fits into the family of linear-time dynamic programming parsers described in (Huang and Sagae, 2010) .",
"Like (Galley and Manning, 2009 ) our work implements an incremental syntactic language model; our approach differs by calculating syntactic LM scores over all available phrase-structure parses at each hypothesis instead of the 1-best dependency parse.",
"The syntax-driven reordering model of Ge (2010) uses syntax-driven features to influence word order within standard phrase-based translation.",
"The syntactic cohesion features of Cherry (2008) encourages the use of syntactically well-formed translation phrases.",
"These approaches are fully orthogonal to our proposed incremental syntactic language model, and could be applied in concert with our work.",
"Parser as Syntactic Language Model in Phrase-Based Translation Parsing is the task of selecting the representationτ (typically a tree) that best models the structure of s τ 0 s thẽ τ 1 1 s that τ 1 2 s president τ 1 3 .",
".",
".",
"the president τ 2 1 that president τ 2 2 president Fridaỹ τ 2 3 .",
".",
".",
"president meets τ 3 1 Obama met τ 3 2 .",
".",
".",
"Figure 1 : Partial decoding lattice for standard phrase-based decoding stack algorithm translating the German sentence Der Präsident trifft am Freitag den Vorstand.",
"Each node h in decoding stack t represents the application of a translation option, and includes the source sentence coverage vector, target language ngram state, and syntactic language model stateτ t h .",
"Hypothesis combination is also shown, indicating where lattice paths with identical n-gram histories converge.",
"We use the English translation The president meets the board on Friday as a running example throughout all Figures.",
"sentence e, out of all such possible representations τ .",
"This set of representations may be all phrase structure trees or all dependency trees allowed by the parsing model.",
"Typically, treeτ is taken to be: τ = argmax τ P(τ | e) (1) We define a syntactic language model P(e) based on the total probability mass over all possible trees for string e. This is shown in Equation 2 and decomposed in Equation 3.",
"P(e) = τ ∈τ P(τ, e) (2) P(e) = τ ∈τ P(e | τ )P(τ ) (3) Incremental syntactic language model An incremental parser processes each token of input sequentially from the beginning of a sentence to the end, rather than processing input in a top-down (Earley, 1968) or bottom-up (Cocke and Schwartz, 1970; Kasami, 1965; Younger, 1967) fashion.",
"After processing the tth token in string e, an incremental parser has some internal representation of possible hypothesized (incomplete) trees, τ t .",
"The syntactic language model probability of a partial sentence e 1 ...e t is defined: P(e 1 ...e t ) = τ ∈τt P(e 1 ...e t | τ )P(τ ) (4) In practice, a parser may constrain the set of trees under consideration toτ t , that subset of analyses or partial analyses that remains after any pruning is performed.",
"An incremental syntactic language model can then be defined by a probability mass function (Equation 5) and a transition function δ (Equation 6 ).",
"The role of δ is explained in §3.3 below.",
"Any parser which implements these two functions can serve as a syntactic language model.",
"P(e 1 ...e t ) ≈ P(τ t ) = τ ∈τ t P(e 1 ...e t | τ )P(τ ) (5) δ(e t ,τ t−1 ) →τ t (6) 3.2 Decoding in phrase-based translation Given a source language input sentence f , a trained source-to-target translation model, and a target language model, the task of translation is to find the maximally probable translationê using a linear combination of j feature functions h weighted according to tuned parameters λ (Och and Ney, 2002) .",
"e = argmax e exp( j λ j h j (e, f )) (7) Phrase-based translation constructs a set of translation options -hypothesized translations for contiguous portions of the source sentence -from a trained phrase table, then incrementally constructs a lattice of partial target translations (Koehn, 2010) .",
"To prune the search space, lattice nodes are organized into beam stacks (Jelinek, 1969) according to the number of source words translated.",
"An n-gram language model history is also maintained at each node in the translation lattice.",
"The search space is further trimmed with hypothesis recombination, which collapses lattice nodes that share a common coverage vector and n-gram state.",
"Incorporating a Syntactic Language Model Phrase-based translation produces target language words in an incremental left-to-right fashion, generating words at the beginning of a translation first and words at the end of a translation last.",
"Similarly, incremental parsers process sentences in an incremental fashion, analyzing words at the beginning of a sentence first and words at the end of a sentence last.",
"As such, an incremental parser with transition function δ can be incorporated into the phrase-based decoding process in a straightforward manner.",
"Each node in the translation lattice is augmented with a syntactic language model stateτ t .",
"The hypothesis at the root of the translation lattice is initialized withτ 0 , representing the internal state of the incremental parser before any input words are processed.",
"The phrase-based translation decoding process adds nodes to the lattice; each new node contains one or more target language words.",
"Each node contains a backpointer to its parent node, in whichτ t−1 is stored.",
"Given a new target language word e t andτ t−1 , the incremental parser's transition function δ calculatesτ t .",
"Figure 1 a sample phrase-based decoding lattice where each translation lattice node is augmented with syntactic language model stateτ t .",
"In phrase-based translation, many translation lattice nodes represent multi-word target language phrases.",
"For such translation lattice nodes, δ will be called once for each newly hypothesized target language word in the node.",
"Only the final syntactic language model state in such sequences need be stored in the translation lattice node.",
"Incremental Bounded-Memory Parsing with a Time Series Model Having defined the framework by which any incremental parser may be incorporated into phrasebased translation, we now formally define a specific incremental parser for use in our experiments.",
"The parser must process target language words incrementally as the phrase-based decoder adds hypotheses to the translation lattice.",
"To facilitate this incremental processing, ordinary phrase-structure trees can be transformed into right-corner recur- r 1 t−1 r 2 t−1 r 3 t−1 s 1 t−1 s 2 t−1 s 3 t−1 r 1 t r 2 t r 3 t s 1 t s 2 t s 3 t e t−1 e t .",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
".",
"Figure 4 : Graphical representation of the dependency structure in a standard Hierarchic Hidden Markov Model with D = 3 hidden levels that can be used to parse syntax.",
"Circles denote random variables, and edges denote conditional dependencies.",
"Shaded circles denote variables with observed values.",
"sive phrase structure trees using the tree transforms in Schuler et al.",
"(2010) .",
"Constituent nonterminals in right-corner transformed trees take the form of incomplete constituents c η /c ηι consisting of an 'active' constituent c η lacking an 'awaited' constituent c ηι yet to come, similar to non-constituent categories in a Combinatory Categorial Grammar (Ades and Steedman, 1982; Steedman, 2000) .",
"As an example, the parser might consider VP/NN as a possible category for input \"meets the\".",
"A sample phrase structure tree is shown before and after the right-corner transform in Figures 2 and 3.",
"Our parser operates over a right-corner transformed probabilistic context-free grammar (PCFG).",
"Parsing runs in linear time on the length of the input.",
"This model of incremental parsing is implemented as a Hierarchical Hidden Markov Model (HHMM) (Murphy and Paskin, 2001) , and is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The parser runs in O(n) time, where n is the number of words in the input.",
"This model is shown graphically in Figure 4 and formally defined in §4.1 below.",
"The incremental parser assigns a probability (Eq.",
"5) for a partial target language hypothesis, using a bounded store of incomplete constituents c η /c ηι .",
"The phrase-based decoder uses this probability value as the syntactic language model feature score.",
"Formal Parsing Model: Scoring Partial Translation Hypotheses This model is essentially an extension of an HHMM, which obtains a most likely sequence of hidden store states,ŝ 1..D 1..T , of some length T and some maximum depth D, given a sequence of observed tokens (e.g.",
"generated target language words), e 1..T , using HHMM state transition model θ A and observation symbol model θ B (Rabiner, 1990) : s 1..D 1..T def = argmax s 1..D 1..T T t=1 P θ A (s 1..D t | s 1..D t−1 )·P θ B (e t | s 1..D t ) (8) The HHMM parser is equivalent to a probabilistic pushdown automaton with a bounded pushdown store.",
"The model generates each successive store (using store model θ S ) only after considering whether each nested sequence of incomplete constituents has completed and reduced (using reduction model θ R ): P θ A (s 1..D t | s 1..D t−1 ) def = r 1 t ..r D t D d=1 P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) · P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) (9) Store elements are defined to contain only the active (c η ) and awaited (c ηι ) constituent categories necessary to compute an incomplete constituent probability: s d t def = c η , c ηι (10) Reduction states are defined to contain only the complete constituent category c r d t necessary to compute an inside likelihood probability, as well as a flag f r d t indicating whether a reduction has taken place (to end a sequence of incomplete constituents): r d t def = c r d t , f r d t (11) The model probabilities for these store elements and reduction states can then be defined (from Murphy and Paskin 2001) to expand a new incomplete constituent after a reduction has taken place (f r d t = 1; using depth-specific store state expansion model θ S-E,d ), transition along a sequence of store elements if no reduction has taken place (f r d t = 0; using depthspecific store state transition model θ S-T,d ): 2 P θ S (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if f r d+1 t = 1, f r d t = 1 : P θ S-E,d (s d t | s d−1 t ) if f r d+1 t = 1, f r d t = 0 : P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) if f r d+1 t = 0, f r d t = 0 : s d t = s d t−1 (12) and possibly reduce a store element (terminate a sequence) if the store state below it has reduced (f r d+1 t = 1; using depth-specific reduction model θ R,d ): P θ R (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if f r d+1 t = 0 : r d t = r ⊥ if f r d+1 t = 1 : P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) (13) where r ⊥ is a null state resulting from the failure of an incomplete constituent to complete, and constants are defined for the edge conditions of s 0 t and r D+1 t .",
"Figure 5 illustrates this model in action.",
"These pushdown automaton operations are then refined for right-corner parsing (Schuler, 2009) , distinguishing active transitions (model θ S-T-A,d , in which an incomplete constituent is completed, but not reduced, and then immediately expanded to a 2 An indicator function · is used to denote deterministic probabilities: φ = 1 if φ is true, 0 otherwise.",
"new incomplete constituent in the same store element) from awaited transitions (model θ S-T-W,d , which involve no completion): P θ S-T,d (s d t | r d+1 t r d t s d t−1 s d−1 t ) def = if r d t = r ⊥ : P θ S-T-A,d (s d t | s d−1 t r d t ) if r d t = r ⊥ : P θ S-T-W,d (s d t | s d t−1 r d+1 t ) (14) P θ R,d (r d t | r d+1 t s d t−1 s d−1 t−1 ) def = if c r d+1 t = x t : r d t = r ⊥ if c r d+1 t = x t : P θ R-R,d (r d t | s d t−1 s d−1 t−1 ) (15) These HHMM right-corner parsing operations are then defined in terms of branch-and depth-specific PCFG probabilities θ G-R,d and θ G-L,d : 3 3 Model probabilities are also defined in terms of leftprogeny probability distribution E θ G-RL * ,d which is itself defined in terms of PCFG probabilities: are calculated by transition function δ (Eq.",
"6, as defined by §4.1), but are not stored.",
"Observed random variables (e 3 ..e 5 ) are shown for clarity, but are not explicitly stored in any syntactic language model state.",
"E θ G-RL * ,d (cη 0 → cη0 ...) def = c η1 P θ G-R,d (cη → cη0 cη1) (16) E θ G-RL * ,d (cη k → c η0 k 0 ...) def = c η0 k E θ G-RL * ,d (cη k−1 → c η0 k ...) · c η0 k 1 P θ G-L,d (c η0 k → c η0 k 0 c η0 k 1 ) (17) E θ G-RL * ,d (cη * → cηι ...) def = ∞ k=0 E θ G-RL * ,d (cη k → cηι ...) (18) E θ G-RL * ,d (cη + → cηι ...) def = E θ G-RL * ,d (cη * → cηι ...) − E θ G-RL * ,d (cη 0 → cηι ...) (19) coder's hypothesis stacks.",
"Figure 1 illustrates an excerpt from a standard phrase-based translation lattice.",
"Within each decoder stack t, each hypothesis h is augmented with a syntactic language model stateτ t h .",
"Each syntactic language model state is a random variable store, containing a slice of random variables from the HHMM.",
"Specifically,τ t h contains those random variables s 1..D t that maintain distributions over syntactic elements.",
"By maintaining these syntactic random variable stores, each hypothesis has access to the current language model probability for the partial translation ending at that hypothesis, as calculated by an incremental syntactic language model defined by the HHMM.",
"Specifically, the random variable store at hypothesis h provides P(τ t h ) = P(e h 1..t , s 1..D 1..t ), where e h 1..t is the sequence of words in a partial hypothesis ending at h which contains t target words, and where there are D syntactic random variables in each random variable store (Eq.",
"5).",
"During stack decoding, the phrase-based decoder progressively constructs new hypotheses by extending existing hypotheses.",
"New hypotheses are placed in appropriate hypothesis stacks.",
"In the simplest case, a new hypothesis extends an existing hypothesis by exactly one target word.",
"As the new hypothesis is constructed by extending an existing stack element, the store and reduction state random variables are processed, along with the newly hypothesized word.",
"This results in a new store of syntactic random variables (Eq.",
"6) that are associated with the new stack element.",
"When a new hypothesis extends an existing hypothesis by more than one word, this process is first carried out for the first new word in the hypothesis.",
"It is then repeated for the remaining words in the hypothesis extension.",
"Once the final word in the hypothesis has been processed, the resulting random variable store is associated with that hypothesis.",
"The random variable stores created for the non-final words in the extending hypothesis are discarded, and need not be explicitly retained.",
"Figure 6 illustrates this process, showing how a syntactic language model stateτ 5 1 in a phrase-based decoding lattice is obtained from a previous syntactic language model stateτ 3 1 (from Figure 1) by parsing the target language words from a phrasebased translation option.",
"Our syntactic language model is integrated into the current version of Moses .",
"Results As an initial measure to compare language models, average per-word perplexity, ppl, reports how surprised a model is by test data.",
"Equation 25 calculates ppl using log base b for a test set of T tokens.",
"ppl = b −log b P(e 1 ...e T ) T (25) We trained the syntactic language model from §4 (HHMM) and an interpolated n-gram language model with modified Kneser-Ney smoothing (Chen and Goodman, 1998); models were trained on sections 2-21 of the Wall Street Journal (WSJ) treebank (Marcus et al., 1993 HHMM and n-gram LMs (Figure 7) .",
"To show the effects of training an LM on more data, we also report perplexity results on the 5-gram LM trained for the GALE Arabic-English task using the English Gigaword corpus.",
"In all cases, including the HHMM significantly reduces perplexity.",
"We trained a phrase-based translation model on the full NIST Open MT08 Urdu-English translation model using the full training data.",
"We trained the HHMM and n-gram LMs on the WSJ data in order to make them as similar as possible.",
"During tuning, Moses was first configured to use just the n-gram LM, then configured to use both the n-gram LM and the syntactic HHMM LM.",
"MERT consistently assigned positive weight to the syntactic LM feature, typically slightly less than the n-gram LM weight.",
"In our integration with Moses, incorporating a syntactic language model dramatically slows the decoding process.",
"Figure 8 illustrates a slowdown around three orders of magnitude.",
"Although speed remains roughly linear to the size of the source sentence (ruling out exponential behavior), it is with an extremely large constant time factor.",
"Due to this slowdown, we tuned the parameters using a constrained dev set (only sentences with 1-20 words), and tested using a constrained devtest set (only sentences with 1-20 words).",
"Figure 9 shows a statistically significant improvement to the BLEU score when using the HHMM and the n-gram LMs together on this reduced test set.",
"Discussion This paper argues that incremental syntactic languages models are a straightforward and appro-Moses LM(s) BLEU n-gram only 18.78 HHMM + n-gram 19.78 Figure 9 : Results for Ur-En devtest (only sentences with 1-20 words) with HHMM beam size of 2000 and Moses settings of distortion limit 10, stack size 200, and ttable limit 20. priate algorithmic fit for incorporating syntax into phrase-based statistical machine translation, since both process sentences in an incremental left-toright fashion.",
"This means incremental syntactic LM scores can be calculated during the decoding process, rather than waiting until a complete sentence is posited, which is typically necessary in top-down or bottom-up parsing.",
"We provided a rigorous formal definition of incremental syntactic languages models, and detailed what steps are necessary to incorporate such LMs into phrase-based decoding.",
"We integrated an incremental syntactic language model into Moses.",
"The translation quality significantly improved on a constrained task, and the perplexity improvements suggest that interpolating between n-gram and syntactic LMs may hold promise on larger data sets.",
"The use of very large n-gram language models is typically a key ingredient in the best-performing machine translation systems (Brants et al., 2007) .",
"Our n-gram model trained only on WSJ is admittedly small.",
"Our future work seeks to incorporate largescale n-gram language models in conjunction with incremental syntactic language models.",
"The added decoding time cost of our syntactic language model is very high.",
"By increasing the beam size and distortion limit of the baseline system, future work may examine whether a baseline system with comparable runtimes can achieve comparable translation quality.",
"A more efficient implementation of the HHMM parser would speed decoding and make more extensive and conclusive translation experiments possible.",
"Various additional improvements could include caching the HHMM LM calculations, and exploiting properties of the right-corner transform that limit the number of decisions between successive time steps."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.3",
"4",
"4.1",
"6",
"7"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Parser as Syntactic Language Model in",
"Incremental syntactic language model",
"Incorporating a Syntactic Language Model",
"Incremental Bounded-Memory Parsing with a Time Series Model",
"Formal Parsing Model: Scoring Partial Translation Hypotheses",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-1#paper-954#slide-14 | Phrase Based Translation w ntactic | e string of n target language words e1. . .en
et the first t words in e, where tn
t set of all incremental parses of et
def t subset of parses t that remain after parser pruning
e argmax P( e) t1 t
Motivation Syntactic LM Decoder Integration Results | e string of n target language words e1. . .en
et the first t words in e, where tn
t set of all incremental parses of et
def t subset of parses t that remain after parser pruning
e argmax P( e) t1 t
Motivation Syntactic LM Decoder Integration Results | [] |
GEM-SciDuet-train-2#paper-957#slide-0 | 957 | LINA: Identifying Comparable Documents from Wikipedia | This paper describes the LINA system for the BUCC 2015 shared track. Following (Enright and Kondrak, 2007), our system identify comparable documents by collecting counts of hapax words. We extend this method by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53
],
"paper_content_text": [
"Introduction Parallel corpora, that is, collections of documents that are mutual translations, are used in many natural language processing applications, particularly for statistical machine translation.",
"Building such resources is however exceedingly expensive, requiring highly skilled annotators or professional translators (Preiss, 2012) .",
"Comparable corpora, that are sets of texts in two or more languages without being translations of each other, are often considered as a solution for the lack of parallel corpora, and many techniques have been proposed to extract parallel sentences (Munteanu et al., 2004; Abdul-Rauf and Schwenk, 2009; Smith et al., 2010) , or mine word translations (Fung, 1995; Rapp, 1999; Chiao and Zweigenbaum, 2002; Morin et al., 2007; Vulić and Moens, 2012) .",
"Identifying comparable resources in a large amount of multilingual data remains a very challenging task.",
"The purpose of the Building and Using Comparable Corpora (BUCC) 2015 shared task 1 is to provide the first evaluation of existing approaches for identifying comparable resources.",
"More precisely, given a large collection of Wikipedia pages in several languages, the task is to identify the most similar pages across languages.",
"1 https://comparable.limsi.fr/bucc2015/ In this paper, we describe the system that we developed for the BUCC 2015 shared track and show that a language agnostic approach can achieve promising results.",
"Proposed Method The method we propose is based on (Enright and Kondrak, 2007) 's approach to parallel document identification.",
"Documents are treated as bags of words, in which only blank separated strings that are at least four characters long and that appear only once in the document (hapax words) are indexed.",
"Given a document in language A, the document in language B that share the largest number of these words is considered as parallel.",
"Although very simple, this approach was shown to perform very well in detecting parallel documents in Wikipedia (Patry and Langlais, 2011) .",
"The reason for this is that most hapax words are in practice proper nouns or numerical entities, which are often cognates.",
"An example of hapax words extracted from a document is given in Table 1 .",
"We purposely keep urls and special characters, as these are useful clues for identifying translated Wikipedia pages.",
"website major gaston links flutist marcel debost states sources college crunelle conservatoire principal rampal united currently recorded chastain competitions music http://www.oberlin.edu/faculty/mdebost/ under international flutists jean-pierre profile moyse french repertoire amazon lives external *http://www.amazon.com/micheldebost/dp/b000s9zsk0 known teaches conservatory school professor studied kathleen orchestre replaced michel Here, we experiment with this approach for detecting near-parallel (comparable) documents.",
"Following (Patry and Langlais, 2011) , we first search for the potential source-target document pairs.",
"To do so, we select for each document in the source language, the N = 20 documents in the target language that share the largest number of hapax words (hereafter baseline).",
"Scoring each pair of documents independently of other candidate pairs leads to several source documents being paired to a same target document.",
"As indicated in Table 2 , the percentage of English articles that are paired with multiple source documents is high (57.3% for French and 60.4% for German).",
"To address this problem, we remove potential multiple source documents by keeping the document pairs with the highest number of shared words (hereafter pigeonhole).",
"This strategy greatly reduces the number of multiply assigned source documents from roughly 60% to 10%.",
"This in turn removes needlessly paired documents and greatly improves the effectiveness of the method.",
"In an attempt to break the remaining score ties between document pairs, we further extend our model to exploit cross-lingual information.",
"When multiple source documents are paired to a given English document with the same score, we use the paired documents in a third language to order them (hereafter cross-lingual).",
"Here we make two assumptions that are valid for the BUCC 2015 shared Task: (1) we have access to comparable documents in a third language, and (2) source documents should be paired 1-to-1 with target documents.",
"Strategy An example of two French documents (doc fr 1 and doc fr 2) being paired to the same English document (doc en ) is given in Figure 1 .",
"We use the German document (doc de ) paired with doc en and select the French document that shares the largest number of hapax words, which for this example is doc fr 2.",
"This strategy further reduces the number of multiply assigned source documents from 10% to less than 4%.",
"Experiments Experimental settings The BUCC 2015 shared task consists in returning for each Wikipedia page in a source language, up to five ranked suggestions to its linked page in English.",
"Inter-language links, that is, links from a page in one language to an equivalent page in another language, are used to evaluate the effectiveness of the systems.",
"Here, we only focus on the French-English and German-English pairs.",
"Following the task guidelines, we use the following evaluation measures investigate the effectiveness of our method: • Mean Average Precision (MAP).",
"Average of precisions computed at the point of each correctly paired document in the ranked list of paired documents.",
"• Success (Succ.).",
"Precision computed on the first returned paired document.",
"• Precision at 5 (P@5).",
"Precision computed on the 5 topmost paired documents.",
"Results Results are presented in Table 3 .",
"Overall, we observe that the two strategies that filter out multiply assigned source documents improve the performance of the method.",
"The largest part of the improvement comes from using pigeonhole reasoning.",
"The use of cross-lingual information to Table 3 : Performance in terms of MAP, success (Succ.)",
"and precision at 5 (P@5) of our model.",
"break ties between the remaining multiply assigned source documents only gives a small improvement.",
"We assume that the limited number of potential source-target document pairs we use in our experiments (N = 20) is a reason for this.",
"Interestingly, results are consistent across languages and datasets (test and train).",
"Our best configuration, that is, with pigeonhole and crosslingual, achieves nearly 60% of success for the first returned pair.",
"Here we show that a simple and straightforward approach that requires no language-specific resources still yields some interesting results.",
"Discussion In this paper we described the LINA system for the BUCC 2015 shared track.",
"We proposed to extend (Enright and Kondrak, 2007) 's approach to parallel document identification by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information.",
"Experimental results show that our system identifies comparable documents with a precision of about 60%.",
"Scoring document pairs using the number of shared hapax words was first intended to be a baseline for comparison purposes.",
"We tried a finer grained scoring approach relying on bilingual dictionaries and information retrieval weighting schemes.",
"For reasonable computation time, we were unable to include low-frequency words in our system.",
"Partial results were very low and we are still in the process of investigating the reasons for this."
]
} | {
"paper_header_number": [
"1",
"2",
"3.1",
"3.2",
"4"
],
"paper_header_content": [
"Introduction",
"Proposed Method",
"Experimental settings",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-2#paper-957#slide-0 | Introduction | I How far can we go with a language agnostic model?
I We experiment with [Enright and Kondrak, 2007]s parallel document identification
I We adapt the method to the BUCC-2015 Shared task based on two assumptions:
Source documents should be paired 1-to-1 with target documents
We have access to comparable documents in several languages | I How far can we go with a language agnostic model?
I We experiment with [Enright and Kondrak, 2007]s parallel document identification
I We adapt the method to the BUCC-2015 Shared task based on two assumptions:
Source documents should be paired 1-to-1 with target documents
We have access to comparable documents in several languages | [] |
GEM-SciDuet-train-2#paper-957#slide-1 | 957 | LINA: Identifying Comparable Documents from Wikipedia | This paper describes the LINA system for the BUCC 2015 shared track. Following (Enright and Kondrak, 2007), our system identify comparable documents by collecting counts of hapax words. We extend this method by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53
],
"paper_content_text": [
"Introduction Parallel corpora, that is, collections of documents that are mutual translations, are used in many natural language processing applications, particularly for statistical machine translation.",
"Building such resources is however exceedingly expensive, requiring highly skilled annotators or professional translators (Preiss, 2012) .",
"Comparable corpora, that are sets of texts in two or more languages without being translations of each other, are often considered as a solution for the lack of parallel corpora, and many techniques have been proposed to extract parallel sentences (Munteanu et al., 2004; Abdul-Rauf and Schwenk, 2009; Smith et al., 2010) , or mine word translations (Fung, 1995; Rapp, 1999; Chiao and Zweigenbaum, 2002; Morin et al., 2007; Vulić and Moens, 2012) .",
"Identifying comparable resources in a large amount of multilingual data remains a very challenging task.",
"The purpose of the Building and Using Comparable Corpora (BUCC) 2015 shared task 1 is to provide the first evaluation of existing approaches for identifying comparable resources.",
"More precisely, given a large collection of Wikipedia pages in several languages, the task is to identify the most similar pages across languages.",
"1 https://comparable.limsi.fr/bucc2015/ In this paper, we describe the system that we developed for the BUCC 2015 shared track and show that a language agnostic approach can achieve promising results.",
"Proposed Method The method we propose is based on (Enright and Kondrak, 2007) 's approach to parallel document identification.",
"Documents are treated as bags of words, in which only blank separated strings that are at least four characters long and that appear only once in the document (hapax words) are indexed.",
"Given a document in language A, the document in language B that share the largest number of these words is considered as parallel.",
"Although very simple, this approach was shown to perform very well in detecting parallel documents in Wikipedia (Patry and Langlais, 2011) .",
"The reason for this is that most hapax words are in practice proper nouns or numerical entities, which are often cognates.",
"An example of hapax words extracted from a document is given in Table 1 .",
"We purposely keep urls and special characters, as these are useful clues for identifying translated Wikipedia pages.",
"website major gaston links flutist marcel debost states sources college crunelle conservatoire principal rampal united currently recorded chastain competitions music http://www.oberlin.edu/faculty/mdebost/ under international flutists jean-pierre profile moyse french repertoire amazon lives external *http://www.amazon.com/micheldebost/dp/b000s9zsk0 known teaches conservatory school professor studied kathleen orchestre replaced michel Here, we experiment with this approach for detecting near-parallel (comparable) documents.",
"Following (Patry and Langlais, 2011) , we first search for the potential source-target document pairs.",
"To do so, we select for each document in the source language, the N = 20 documents in the target language that share the largest number of hapax words (hereafter baseline).",
"Scoring each pair of documents independently of other candidate pairs leads to several source documents being paired to a same target document.",
"As indicated in Table 2 , the percentage of English articles that are paired with multiple source documents is high (57.3% for French and 60.4% for German).",
"To address this problem, we remove potential multiple source documents by keeping the document pairs with the highest number of shared words (hereafter pigeonhole).",
"This strategy greatly reduces the number of multiply assigned source documents from roughly 60% to 10%.",
"This in turn removes needlessly paired documents and greatly improves the effectiveness of the method.",
"In an attempt to break the remaining score ties between document pairs, we further extend our model to exploit cross-lingual information.",
"When multiple source documents are paired to a given English document with the same score, we use the paired documents in a third language to order them (hereafter cross-lingual).",
"Here we make two assumptions that are valid for the BUCC 2015 shared Task: (1) we have access to comparable documents in a third language, and (2) source documents should be paired 1-to-1 with target documents.",
"Strategy An example of two French documents (doc fr 1 and doc fr 2) being paired to the same English document (doc en ) is given in Figure 1 .",
"We use the German document (doc de ) paired with doc en and select the French document that shares the largest number of hapax words, which for this example is doc fr 2.",
"This strategy further reduces the number of multiply assigned source documents from 10% to less than 4%.",
"Experiments Experimental settings The BUCC 2015 shared task consists in returning for each Wikipedia page in a source language, up to five ranked suggestions to its linked page in English.",
"Inter-language links, that is, links from a page in one language to an equivalent page in another language, are used to evaluate the effectiveness of the systems.",
"Here, we only focus on the French-English and German-English pairs.",
"Following the task guidelines, we use the following evaluation measures investigate the effectiveness of our method: • Mean Average Precision (MAP).",
"Average of precisions computed at the point of each correctly paired document in the ranked list of paired documents.",
"• Success (Succ.).",
"Precision computed on the first returned paired document.",
"• Precision at 5 (P@5).",
"Precision computed on the 5 topmost paired documents.",
"Results Results are presented in Table 3 .",
"Overall, we observe that the two strategies that filter out multiply assigned source documents improve the performance of the method.",
"The largest part of the improvement comes from using pigeonhole reasoning.",
"The use of cross-lingual information to Table 3 : Performance in terms of MAP, success (Succ.)",
"and precision at 5 (P@5) of our model.",
"break ties between the remaining multiply assigned source documents only gives a small improvement.",
"We assume that the limited number of potential source-target document pairs we use in our experiments (N = 20) is a reason for this.",
"Interestingly, results are consistent across languages and datasets (test and train).",
"Our best configuration, that is, with pigeonhole and crosslingual, achieves nearly 60% of success for the first returned pair.",
"Here we show that a simple and straightforward approach that requires no language-specific resources still yields some interesting results.",
"Discussion In this paper we described the LINA system for the BUCC 2015 shared track.",
"We proposed to extend (Enright and Kondrak, 2007) 's approach to parallel document identification by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information.",
"Experimental results show that our system identifies comparable documents with a precision of about 60%.",
"Scoring document pairs using the number of shared hapax words was first intended to be a baseline for comparison purposes.",
"We tried a finer grained scoring approach relying on bilingual dictionaries and information retrieval weighting schemes.",
"For reasonable computation time, we were unable to include low-frequency words in our system.",
"Partial results were very low and we are still in the process of investigating the reasons for this."
]
} | {
"paper_header_number": [
"1",
"2",
"3.1",
"3.2",
"4"
],
"paper_header_content": [
"Introduction",
"Proposed Method",
"Experimental settings",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-2#paper-957#slide-1 | Method | I Fast parallel document identification [Enright and Kondrak, 2007]
I Documents = bags of hapax words
I Words = blank separated strings that are 4+ characters long
I Given a document in language A, the document in language B that shares the largest
number of words is considered as parallel
I Works very well for parallel documents
I 80% precision on Wikipedia [Patry and Langlais, 2011]
I We use this approach as baseline for detecting comparable documents | I Fast parallel document identification [Enright and Kondrak, 2007]
I Documents = bags of hapax words
I Words = blank separated strings that are 4+ characters long
I Given a document in language A, the document in language B that shares the largest
number of words is considered as parallel
I Works very well for parallel documents
I 80% precision on Wikipedia [Patry and Langlais, 2011]
I We use this approach as baseline for detecting comparable documents | [] |
GEM-SciDuet-train-2#paper-957#slide-2 | 957 | LINA: Identifying Comparable Documents from Wikipedia | This paper describes the LINA system for the BUCC 2015 shared track. Following (Enright and Kondrak, 2007), our system identify comparable documents by collecting counts of hapax words. We extend this method by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53
],
"paper_content_text": [
"Introduction Parallel corpora, that is, collections of documents that are mutual translations, are used in many natural language processing applications, particularly for statistical machine translation.",
"Building such resources is however exceedingly expensive, requiring highly skilled annotators or professional translators (Preiss, 2012) .",
"Comparable corpora, that are sets of texts in two or more languages without being translations of each other, are often considered as a solution for the lack of parallel corpora, and many techniques have been proposed to extract parallel sentences (Munteanu et al., 2004; Abdul-Rauf and Schwenk, 2009; Smith et al., 2010) , or mine word translations (Fung, 1995; Rapp, 1999; Chiao and Zweigenbaum, 2002; Morin et al., 2007; Vulić and Moens, 2012) .",
"Identifying comparable resources in a large amount of multilingual data remains a very challenging task.",
"The purpose of the Building and Using Comparable Corpora (BUCC) 2015 shared task 1 is to provide the first evaluation of existing approaches for identifying comparable resources.",
"More precisely, given a large collection of Wikipedia pages in several languages, the task is to identify the most similar pages across languages.",
"1 https://comparable.limsi.fr/bucc2015/ In this paper, we describe the system that we developed for the BUCC 2015 shared track and show that a language agnostic approach can achieve promising results.",
"Proposed Method The method we propose is based on (Enright and Kondrak, 2007) 's approach to parallel document identification.",
"Documents are treated as bags of words, in which only blank separated strings that are at least four characters long and that appear only once in the document (hapax words) are indexed.",
"Given a document in language A, the document in language B that share the largest number of these words is considered as parallel.",
"Although very simple, this approach was shown to perform very well in detecting parallel documents in Wikipedia (Patry and Langlais, 2011) .",
"The reason for this is that most hapax words are in practice proper nouns or numerical entities, which are often cognates.",
"An example of hapax words extracted from a document is given in Table 1 .",
"We purposely keep urls and special characters, as these are useful clues for identifying translated Wikipedia pages.",
"website major gaston links flutist marcel debost states sources college crunelle conservatoire principal rampal united currently recorded chastain competitions music http://www.oberlin.edu/faculty/mdebost/ under international flutists jean-pierre profile moyse french repertoire amazon lives external *http://www.amazon.com/micheldebost/dp/b000s9zsk0 known teaches conservatory school professor studied kathleen orchestre replaced michel Here, we experiment with this approach for detecting near-parallel (comparable) documents.",
"Following (Patry and Langlais, 2011) , we first search for the potential source-target document pairs.",
"To do so, we select for each document in the source language, the N = 20 documents in the target language that share the largest number of hapax words (hereafter baseline).",
"Scoring each pair of documents independently of other candidate pairs leads to several source documents being paired to a same target document.",
"As indicated in Table 2 , the percentage of English articles that are paired with multiple source documents is high (57.3% for French and 60.4% for German).",
"To address this problem, we remove potential multiple source documents by keeping the document pairs with the highest number of shared words (hereafter pigeonhole).",
"This strategy greatly reduces the number of multiply assigned source documents from roughly 60% to 10%.",
"This in turn removes needlessly paired documents and greatly improves the effectiveness of the method.",
"In an attempt to break the remaining score ties between document pairs, we further extend our model to exploit cross-lingual information.",
"When multiple source documents are paired to a given English document with the same score, we use the paired documents in a third language to order them (hereafter cross-lingual).",
"Here we make two assumptions that are valid for the BUCC 2015 shared Task: (1) we have access to comparable documents in a third language, and (2) source documents should be paired 1-to-1 with target documents.",
"Strategy An example of two French documents (doc fr 1 and doc fr 2) being paired to the same English document (doc en ) is given in Figure 1 .",
"We use the German document (doc de ) paired with doc en and select the French document that shares the largest number of hapax words, which for this example is doc fr 2.",
"This strategy further reduces the number of multiply assigned source documents from 10% to less than 4%.",
"Experiments Experimental settings The BUCC 2015 shared task consists in returning for each Wikipedia page in a source language, up to five ranked suggestions to its linked page in English.",
"Inter-language links, that is, links from a page in one language to an equivalent page in another language, are used to evaluate the effectiveness of the systems.",
"Here, we only focus on the French-English and German-English pairs.",
"Following the task guidelines, we use the following evaluation measures investigate the effectiveness of our method: • Mean Average Precision (MAP).",
"Average of precisions computed at the point of each correctly paired document in the ranked list of paired documents.",
"• Success (Succ.).",
"Precision computed on the first returned paired document.",
"• Precision at 5 (P@5).",
"Precision computed on the 5 topmost paired documents.",
"Results Results are presented in Table 3 .",
"Overall, we observe that the two strategies that filter out multiply assigned source documents improve the performance of the method.",
"The largest part of the improvement comes from using pigeonhole reasoning.",
"The use of cross-lingual information to Table 3 : Performance in terms of MAP, success (Succ.)",
"and precision at 5 (P@5) of our model.",
"break ties between the remaining multiply assigned source documents only gives a small improvement.",
"We assume that the limited number of potential source-target document pairs we use in our experiments (N = 20) is a reason for this.",
"Interestingly, results are consistent across languages and datasets (test and train).",
"Our best configuration, that is, with pigeonhole and crosslingual, achieves nearly 60% of success for the first returned pair.",
"Here we show that a simple and straightforward approach that requires no language-specific resources still yields some interesting results.",
"Discussion In this paper we described the LINA system for the BUCC 2015 shared track.",
"We proposed to extend (Enright and Kondrak, 2007) 's approach to parallel document identification by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information.",
"Experimental results show that our system identifies comparable documents with a precision of about 60%.",
"Scoring document pairs using the number of shared hapax words was first intended to be a baseline for comparison purposes.",
"We tried a finer grained scoring approach relying on bilingual dictionaries and information retrieval weighting schemes.",
"For reasonable computation time, we were unable to include low-frequency words in our system.",
"Partial results were very low and we are still in the process of investigating the reasons for this."
]
} | {
"paper_header_number": [
"1",
"2",
"3.1",
"3.2",
"4"
],
"paper_header_content": [
"Introduction",
"Proposed Method",
"Experimental settings",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-2#paper-957#slide-2 | Improvements using 1 to 1 alignments | I In baseline, document pairs are scored independently
I Multiple source documents are paired to a same target document
I 60% of English pages are paired with multiple pages in French or German
I We remove multiply assigned source documents using pigeonhole reasoning
I From 60% to 11% of multiply assigned source documents | I In baseline, document pairs are scored independently
I Multiple source documents are paired to a same target document
I 60% of English pages are paired with multiple pages in French or German
I We remove multiply assigned source documents using pigeonhole reasoning
I From 60% to 11% of multiply assigned source documents | [] |
GEM-SciDuet-train-2#paper-957#slide-3 | 957 | LINA: Identifying Comparable Documents from Wikipedia | This paper describes the LINA system for the BUCC 2015 shared track. Following (Enright and Kondrak, 2007), our system identify comparable documents by collecting counts of hapax words. We extend this method by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53
],
"paper_content_text": [
"Introduction Parallel corpora, that is, collections of documents that are mutual translations, are used in many natural language processing applications, particularly for statistical machine translation.",
"Building such resources is however exceedingly expensive, requiring highly skilled annotators or professional translators (Preiss, 2012) .",
"Comparable corpora, that are sets of texts in two or more languages without being translations of each other, are often considered as a solution for the lack of parallel corpora, and many techniques have been proposed to extract parallel sentences (Munteanu et al., 2004; Abdul-Rauf and Schwenk, 2009; Smith et al., 2010) , or mine word translations (Fung, 1995; Rapp, 1999; Chiao and Zweigenbaum, 2002; Morin et al., 2007; Vulić and Moens, 2012) .",
"Identifying comparable resources in a large amount of multilingual data remains a very challenging task.",
"The purpose of the Building and Using Comparable Corpora (BUCC) 2015 shared task 1 is to provide the first evaluation of existing approaches for identifying comparable resources.",
"More precisely, given a large collection of Wikipedia pages in several languages, the task is to identify the most similar pages across languages.",
"1 https://comparable.limsi.fr/bucc2015/ In this paper, we describe the system that we developed for the BUCC 2015 shared track and show that a language agnostic approach can achieve promising results.",
"Proposed Method The method we propose is based on (Enright and Kondrak, 2007) 's approach to parallel document identification.",
"Documents are treated as bags of words, in which only blank separated strings that are at least four characters long and that appear only once in the document (hapax words) are indexed.",
"Given a document in language A, the document in language B that share the largest number of these words is considered as parallel.",
"Although very simple, this approach was shown to perform very well in detecting parallel documents in Wikipedia (Patry and Langlais, 2011) .",
"The reason for this is that most hapax words are in practice proper nouns or numerical entities, which are often cognates.",
"An example of hapax words extracted from a document is given in Table 1 .",
"We purposely keep urls and special characters, as these are useful clues for identifying translated Wikipedia pages.",
"website major gaston links flutist marcel debost states sources college crunelle conservatoire principal rampal united currently recorded chastain competitions music http://www.oberlin.edu/faculty/mdebost/ under international flutists jean-pierre profile moyse french repertoire amazon lives external *http://www.amazon.com/micheldebost/dp/b000s9zsk0 known teaches conservatory school professor studied kathleen orchestre replaced michel Here, we experiment with this approach for detecting near-parallel (comparable) documents.",
"Following (Patry and Langlais, 2011) , we first search for the potential source-target document pairs.",
"To do so, we select for each document in the source language, the N = 20 documents in the target language that share the largest number of hapax words (hereafter baseline).",
"Scoring each pair of documents independently of other candidate pairs leads to several source documents being paired to a same target document.",
"As indicated in Table 2 , the percentage of English articles that are paired with multiple source documents is high (57.3% for French and 60.4% for German).",
"To address this problem, we remove potential multiple source documents by keeping the document pairs with the highest number of shared words (hereafter pigeonhole).",
"This strategy greatly reduces the number of multiply assigned source documents from roughly 60% to 10%.",
"This in turn removes needlessly paired documents and greatly improves the effectiveness of the method.",
"In an attempt to break the remaining score ties between document pairs, we further extend our model to exploit cross-lingual information.",
"When multiple source documents are paired to a given English document with the same score, we use the paired documents in a third language to order them (hereafter cross-lingual).",
"Here we make two assumptions that are valid for the BUCC 2015 shared Task: (1) we have access to comparable documents in a third language, and (2) source documents should be paired 1-to-1 with target documents.",
"Strategy An example of two French documents (doc fr 1 and doc fr 2) being paired to the same English document (doc en ) is given in Figure 1 .",
"We use the German document (doc de ) paired with doc en and select the French document that shares the largest number of hapax words, which for this example is doc fr 2.",
"This strategy further reduces the number of multiply assigned source documents from 10% to less than 4%.",
"Experiments Experimental settings The BUCC 2015 shared task consists in returning for each Wikipedia page in a source language, up to five ranked suggestions to its linked page in English.",
"Inter-language links, that is, links from a page in one language to an equivalent page in another language, are used to evaluate the effectiveness of the systems.",
"Here, we only focus on the French-English and German-English pairs.",
"Following the task guidelines, we use the following evaluation measures investigate the effectiveness of our method: • Mean Average Precision (MAP).",
"Average of precisions computed at the point of each correctly paired document in the ranked list of paired documents.",
"• Success (Succ.).",
"Precision computed on the first returned paired document.",
"• Precision at 5 (P@5).",
"Precision computed on the 5 topmost paired documents.",
"Results Results are presented in Table 3 .",
"Overall, we observe that the two strategies that filter out multiply assigned source documents improve the performance of the method.",
"The largest part of the improvement comes from using pigeonhole reasoning.",
"The use of cross-lingual information to Table 3 : Performance in terms of MAP, success (Succ.)",
"and precision at 5 (P@5) of our model.",
"break ties between the remaining multiply assigned source documents only gives a small improvement.",
"We assume that the limited number of potential source-target document pairs we use in our experiments (N = 20) is a reason for this.",
"Interestingly, results are consistent across languages and datasets (test and train).",
"Our best configuration, that is, with pigeonhole and crosslingual, achieves nearly 60% of success for the first returned pair.",
"Here we show that a simple and straightforward approach that requires no language-specific resources still yields some interesting results.",
"Discussion In this paper we described the LINA system for the BUCC 2015 shared track.",
"We proposed to extend (Enright and Kondrak, 2007) 's approach to parallel document identification by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information.",
"Experimental results show that our system identifies comparable documents with a precision of about 60%.",
"Scoring document pairs using the number of shared hapax words was first intended to be a baseline for comparison purposes.",
"We tried a finer grained scoring approach relying on bilingual dictionaries and information retrieval weighting schemes.",
"For reasonable computation time, we were unable to include low-frequency words in our system.",
"Partial results were very low and we are still in the process of investigating the reasons for this."
]
} | {
"paper_header_number": [
"1",
"2",
"3.1",
"3.2",
"4"
],
"paper_header_content": [
"Introduction",
"Proposed Method",
"Experimental settings",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-2#paper-957#slide-3 | Improvements using cross lingual information | I Simple document weighting function score ties
I We break the remaining score ties using a third language
I From 11% to less than 4% of multiply assigned source documents | I Simple document weighting function score ties
I We break the remaining score ties using a third language
I From 11% to less than 4% of multiply assigned source documents | [] |
GEM-SciDuet-train-2#paper-957#slide-4 | 957 | LINA: Identifying Comparable Documents from Wikipedia | This paper describes the LINA system for the BUCC 2015 shared track. Following (Enright and Kondrak, 2007), our system identify comparable documents by collecting counts of hapax words. We extend this method by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53
],
"paper_content_text": [
"Introduction Parallel corpora, that is, collections of documents that are mutual translations, are used in many natural language processing applications, particularly for statistical machine translation.",
"Building such resources is however exceedingly expensive, requiring highly skilled annotators or professional translators (Preiss, 2012) .",
"Comparable corpora, that are sets of texts in two or more languages without being translations of each other, are often considered as a solution for the lack of parallel corpora, and many techniques have been proposed to extract parallel sentences (Munteanu et al., 2004; Abdul-Rauf and Schwenk, 2009; Smith et al., 2010) , or mine word translations (Fung, 1995; Rapp, 1999; Chiao and Zweigenbaum, 2002; Morin et al., 2007; Vulić and Moens, 2012) .",
"Identifying comparable resources in a large amount of multilingual data remains a very challenging task.",
"The purpose of the Building and Using Comparable Corpora (BUCC) 2015 shared task 1 is to provide the first evaluation of existing approaches for identifying comparable resources.",
"More precisely, given a large collection of Wikipedia pages in several languages, the task is to identify the most similar pages across languages.",
"1 https://comparable.limsi.fr/bucc2015/ In this paper, we describe the system that we developed for the BUCC 2015 shared track and show that a language agnostic approach can achieve promising results.",
"Proposed Method The method we propose is based on (Enright and Kondrak, 2007) 's approach to parallel document identification.",
"Documents are treated as bags of words, in which only blank separated strings that are at least four characters long and that appear only once in the document (hapax words) are indexed.",
"Given a document in language A, the document in language B that share the largest number of these words is considered as parallel.",
"Although very simple, this approach was shown to perform very well in detecting parallel documents in Wikipedia (Patry and Langlais, 2011) .",
"The reason for this is that most hapax words are in practice proper nouns or numerical entities, which are often cognates.",
"An example of hapax words extracted from a document is given in Table 1 .",
"We purposely keep urls and special characters, as these are useful clues for identifying translated Wikipedia pages.",
"website major gaston links flutist marcel debost states sources college crunelle conservatoire principal rampal united currently recorded chastain competitions music http://www.oberlin.edu/faculty/mdebost/ under international flutists jean-pierre profile moyse french repertoire amazon lives external *http://www.amazon.com/micheldebost/dp/b000s9zsk0 known teaches conservatory school professor studied kathleen orchestre replaced michel Here, we experiment with this approach for detecting near-parallel (comparable) documents.",
"Following (Patry and Langlais, 2011) , we first search for the potential source-target document pairs.",
"To do so, we select for each document in the source language, the N = 20 documents in the target language that share the largest number of hapax words (hereafter baseline).",
"Scoring each pair of documents independently of other candidate pairs leads to several source documents being paired to a same target document.",
"As indicated in Table 2 , the percentage of English articles that are paired with multiple source documents is high (57.3% for French and 60.4% for German).",
"To address this problem, we remove potential multiple source documents by keeping the document pairs with the highest number of shared words (hereafter pigeonhole).",
"This strategy greatly reduces the number of multiply assigned source documents from roughly 60% to 10%.",
"This in turn removes needlessly paired documents and greatly improves the effectiveness of the method.",
"In an attempt to break the remaining score ties between document pairs, we further extend our model to exploit cross-lingual information.",
"When multiple source documents are paired to a given English document with the same score, we use the paired documents in a third language to order them (hereafter cross-lingual).",
"Here we make two assumptions that are valid for the BUCC 2015 shared Task: (1) we have access to comparable documents in a third language, and (2) source documents should be paired 1-to-1 with target documents.",
"Strategy An example of two French documents (doc fr 1 and doc fr 2) being paired to the same English document (doc en ) is given in Figure 1 .",
"We use the German document (doc de ) paired with doc en and select the French document that shares the largest number of hapax words, which for this example is doc fr 2.",
"This strategy further reduces the number of multiply assigned source documents from 10% to less than 4%.",
"Experiments Experimental settings The BUCC 2015 shared task consists in returning for each Wikipedia page in a source language, up to five ranked suggestions to its linked page in English.",
"Inter-language links, that is, links from a page in one language to an equivalent page in another language, are used to evaluate the effectiveness of the systems.",
"Here, we only focus on the French-English and German-English pairs.",
"Following the task guidelines, we use the following evaluation measures investigate the effectiveness of our method: • Mean Average Precision (MAP).",
"Average of precisions computed at the point of each correctly paired document in the ranked list of paired documents.",
"• Success (Succ.).",
"Precision computed on the first returned paired document.",
"• Precision at 5 (P@5).",
"Precision computed on the 5 topmost paired documents.",
"Results Results are presented in Table 3 .",
"Overall, we observe that the two strategies that filter out multiply assigned source documents improve the performance of the method.",
"The largest part of the improvement comes from using pigeonhole reasoning.",
"The use of cross-lingual information to Table 3 : Performance in terms of MAP, success (Succ.)",
"and precision at 5 (P@5) of our model.",
"break ties between the remaining multiply assigned source documents only gives a small improvement.",
"We assume that the limited number of potential source-target document pairs we use in our experiments (N = 20) is a reason for this.",
"Interestingly, results are consistent across languages and datasets (test and train).",
"Our best configuration, that is, with pigeonhole and crosslingual, achieves nearly 60% of success for the first returned pair.",
"Here we show that a simple and straightforward approach that requires no language-specific resources still yields some interesting results.",
"Discussion In this paper we described the LINA system for the BUCC 2015 shared track.",
"We proposed to extend (Enright and Kondrak, 2007) 's approach to parallel document identification by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information.",
"Experimental results show that our system identifies comparable documents with a precision of about 60%.",
"Scoring document pairs using the number of shared hapax words was first intended to be a baseline for comparison purposes.",
"We tried a finer grained scoring approach relying on bilingual dictionaries and information retrieval weighting schemes.",
"For reasonable computation time, we were unable to include low-frequency words in our system.",
"Partial results were very low and we are still in the process of investigating the reasons for this."
]
} | {
"paper_header_number": [
"1",
"2",
"3.1",
"3.2",
"4"
],
"paper_header_content": [
"Introduction",
"Proposed Method",
"Experimental settings",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-2#paper-957#slide-4 | Experimental settings | I We focus on the French-English and German-English pairs
I The following measures are considered relevant
I Mean Average Precision (MAP) | I We focus on the French-English and German-English pairs
I The following measures are considered relevant
I Mean Average Precision (MAP) | [] |
GEM-SciDuet-train-2#paper-957#slide-5 | 957 | LINA: Identifying Comparable Documents from Wikipedia | This paper describes the LINA system for the BUCC 2015 shared track. Following (Enright and Kondrak, 2007), our system identify comparable documents by collecting counts of hapax words. We extend this method by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53
],
"paper_content_text": [
"Introduction Parallel corpora, that is, collections of documents that are mutual translations, are used in many natural language processing applications, particularly for statistical machine translation.",
"Building such resources is however exceedingly expensive, requiring highly skilled annotators or professional translators (Preiss, 2012) .",
"Comparable corpora, that are sets of texts in two or more languages without being translations of each other, are often considered as a solution for the lack of parallel corpora, and many techniques have been proposed to extract parallel sentences (Munteanu et al., 2004; Abdul-Rauf and Schwenk, 2009; Smith et al., 2010) , or mine word translations (Fung, 1995; Rapp, 1999; Chiao and Zweigenbaum, 2002; Morin et al., 2007; Vulić and Moens, 2012) .",
"Identifying comparable resources in a large amount of multilingual data remains a very challenging task.",
"The purpose of the Building and Using Comparable Corpora (BUCC) 2015 shared task 1 is to provide the first evaluation of existing approaches for identifying comparable resources.",
"More precisely, given a large collection of Wikipedia pages in several languages, the task is to identify the most similar pages across languages.",
"1 https://comparable.limsi.fr/bucc2015/ In this paper, we describe the system that we developed for the BUCC 2015 shared track and show that a language agnostic approach can achieve promising results.",
"Proposed Method The method we propose is based on (Enright and Kondrak, 2007) 's approach to parallel document identification.",
"Documents are treated as bags of words, in which only blank separated strings that are at least four characters long and that appear only once in the document (hapax words) are indexed.",
"Given a document in language A, the document in language B that share the largest number of these words is considered as parallel.",
"Although very simple, this approach was shown to perform very well in detecting parallel documents in Wikipedia (Patry and Langlais, 2011) .",
"The reason for this is that most hapax words are in practice proper nouns or numerical entities, which are often cognates.",
"An example of hapax words extracted from a document is given in Table 1 .",
"We purposely keep urls and special characters, as these are useful clues for identifying translated Wikipedia pages.",
"website major gaston links flutist marcel debost states sources college crunelle conservatoire principal rampal united currently recorded chastain competitions music http://www.oberlin.edu/faculty/mdebost/ under international flutists jean-pierre profile moyse french repertoire amazon lives external *http://www.amazon.com/micheldebost/dp/b000s9zsk0 known teaches conservatory school professor studied kathleen orchestre replaced michel Here, we experiment with this approach for detecting near-parallel (comparable) documents.",
"Following (Patry and Langlais, 2011) , we first search for the potential source-target document pairs.",
"To do so, we select for each document in the source language, the N = 20 documents in the target language that share the largest number of hapax words (hereafter baseline).",
"Scoring each pair of documents independently of other candidate pairs leads to several source documents being paired to a same target document.",
"As indicated in Table 2 , the percentage of English articles that are paired with multiple source documents is high (57.3% for French and 60.4% for German).",
"To address this problem, we remove potential multiple source documents by keeping the document pairs with the highest number of shared words (hereafter pigeonhole).",
"This strategy greatly reduces the number of multiply assigned source documents from roughly 60% to 10%.",
"This in turn removes needlessly paired documents and greatly improves the effectiveness of the method.",
"In an attempt to break the remaining score ties between document pairs, we further extend our model to exploit cross-lingual information.",
"When multiple source documents are paired to a given English document with the same score, we use the paired documents in a third language to order them (hereafter cross-lingual).",
"Here we make two assumptions that are valid for the BUCC 2015 shared Task: (1) we have access to comparable documents in a third language, and (2) source documents should be paired 1-to-1 with target documents.",
"Strategy An example of two French documents (doc fr 1 and doc fr 2) being paired to the same English document (doc en ) is given in Figure 1 .",
"We use the German document (doc de ) paired with doc en and select the French document that shares the largest number of hapax words, which for this example is doc fr 2.",
"This strategy further reduces the number of multiply assigned source documents from 10% to less than 4%.",
"Experiments Experimental settings The BUCC 2015 shared task consists in returning for each Wikipedia page in a source language, up to five ranked suggestions to its linked page in English.",
"Inter-language links, that is, links from a page in one language to an equivalent page in another language, are used to evaluate the effectiveness of the systems.",
"Here, we only focus on the French-English and German-English pairs.",
"Following the task guidelines, we use the following evaluation measures investigate the effectiveness of our method: • Mean Average Precision (MAP).",
"Average of precisions computed at the point of each correctly paired document in the ranked list of paired documents.",
"• Success (Succ.).",
"Precision computed on the first returned paired document.",
"• Precision at 5 (P@5).",
"Precision computed on the 5 topmost paired documents.",
"Results Results are presented in Table 3 .",
"Overall, we observe that the two strategies that filter out multiply assigned source documents improve the performance of the method.",
"The largest part of the improvement comes from using pigeonhole reasoning.",
"The use of cross-lingual information to Table 3 : Performance in terms of MAP, success (Succ.)",
"and precision at 5 (P@5) of our model.",
"break ties between the remaining multiply assigned source documents only gives a small improvement.",
"We assume that the limited number of potential source-target document pairs we use in our experiments (N = 20) is a reason for this.",
"Interestingly, results are consistent across languages and datasets (test and train).",
"Our best configuration, that is, with pigeonhole and crosslingual, achieves nearly 60% of success for the first returned pair.",
"Here we show that a simple and straightforward approach that requires no language-specific resources still yields some interesting results.",
"Discussion In this paper we described the LINA system for the BUCC 2015 shared track.",
"We proposed to extend (Enright and Kondrak, 2007) 's approach to parallel document identification by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information.",
"Experimental results show that our system identifies comparable documents with a precision of about 60%.",
"Scoring document pairs using the number of shared hapax words was first intended to be a baseline for comparison purposes.",
"We tried a finer grained scoring approach relying on bilingual dictionaries and information retrieval weighting schemes.",
"For reasonable computation time, we were unable to include low-frequency words in our system.",
"Partial results were very low and we are still in the process of investigating the reasons for this."
]
} | {
"paper_header_number": [
"1",
"2",
"3.1",
"3.2",
"4"
],
"paper_header_content": [
"Introduction",
"Proposed Method",
"Experimental settings",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-2#paper-957#slide-5 | Results FR EN | Strategy MAP Succ. P@5 MAP Succ. P@5 | Strategy MAP Succ. P@5 MAP Succ. P@5 | [] |
GEM-SciDuet-train-2#paper-957#slide-6 | 957 | LINA: Identifying Comparable Documents from Wikipedia | This paper describes the LINA system for the BUCC 2015 shared track. Following (Enright and Kondrak, 2007), our system identify comparable documents by collecting counts of hapax words. We extend this method by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53
],
"paper_content_text": [
"Introduction Parallel corpora, that is, collections of documents that are mutual translations, are used in many natural language processing applications, particularly for statistical machine translation.",
"Building such resources is however exceedingly expensive, requiring highly skilled annotators or professional translators (Preiss, 2012) .",
"Comparable corpora, that are sets of texts in two or more languages without being translations of each other, are often considered as a solution for the lack of parallel corpora, and many techniques have been proposed to extract parallel sentences (Munteanu et al., 2004; Abdul-Rauf and Schwenk, 2009; Smith et al., 2010) , or mine word translations (Fung, 1995; Rapp, 1999; Chiao and Zweigenbaum, 2002; Morin et al., 2007; Vulić and Moens, 2012) .",
"Identifying comparable resources in a large amount of multilingual data remains a very challenging task.",
"The purpose of the Building and Using Comparable Corpora (BUCC) 2015 shared task 1 is to provide the first evaluation of existing approaches for identifying comparable resources.",
"More precisely, given a large collection of Wikipedia pages in several languages, the task is to identify the most similar pages across languages.",
"1 https://comparable.limsi.fr/bucc2015/ In this paper, we describe the system that we developed for the BUCC 2015 shared track and show that a language agnostic approach can achieve promising results.",
"Proposed Method The method we propose is based on (Enright and Kondrak, 2007) 's approach to parallel document identification.",
"Documents are treated as bags of words, in which only blank separated strings that are at least four characters long and that appear only once in the document (hapax words) are indexed.",
"Given a document in language A, the document in language B that share the largest number of these words is considered as parallel.",
"Although very simple, this approach was shown to perform very well in detecting parallel documents in Wikipedia (Patry and Langlais, 2011) .",
"The reason for this is that most hapax words are in practice proper nouns or numerical entities, which are often cognates.",
"An example of hapax words extracted from a document is given in Table 1 .",
"We purposely keep urls and special characters, as these are useful clues for identifying translated Wikipedia pages.",
"website major gaston links flutist marcel debost states sources college crunelle conservatoire principal rampal united currently recorded chastain competitions music http://www.oberlin.edu/faculty/mdebost/ under international flutists jean-pierre profile moyse french repertoire amazon lives external *http://www.amazon.com/micheldebost/dp/b000s9zsk0 known teaches conservatory school professor studied kathleen orchestre replaced michel Here, we experiment with this approach for detecting near-parallel (comparable) documents.",
"Following (Patry and Langlais, 2011) , we first search for the potential source-target document pairs.",
"To do so, we select for each document in the source language, the N = 20 documents in the target language that share the largest number of hapax words (hereafter baseline).",
"Scoring each pair of documents independently of other candidate pairs leads to several source documents being paired to a same target document.",
"As indicated in Table 2 , the percentage of English articles that are paired with multiple source documents is high (57.3% for French and 60.4% for German).",
"To address this problem, we remove potential multiple source documents by keeping the document pairs with the highest number of shared words (hereafter pigeonhole).",
"This strategy greatly reduces the number of multiply assigned source documents from roughly 60% to 10%.",
"This in turn removes needlessly paired documents and greatly improves the effectiveness of the method.",
"In an attempt to break the remaining score ties between document pairs, we further extend our model to exploit cross-lingual information.",
"When multiple source documents are paired to a given English document with the same score, we use the paired documents in a third language to order them (hereafter cross-lingual).",
"Here we make two assumptions that are valid for the BUCC 2015 shared Task: (1) we have access to comparable documents in a third language, and (2) source documents should be paired 1-to-1 with target documents.",
"Strategy An example of two French documents (doc fr 1 and doc fr 2) being paired to the same English document (doc en ) is given in Figure 1 .",
"We use the German document (doc de ) paired with doc en and select the French document that shares the largest number of hapax words, which for this example is doc fr 2.",
"This strategy further reduces the number of multiply assigned source documents from 10% to less than 4%.",
"Experiments Experimental settings The BUCC 2015 shared task consists in returning for each Wikipedia page in a source language, up to five ranked suggestions to its linked page in English.",
"Inter-language links, that is, links from a page in one language to an equivalent page in another language, are used to evaluate the effectiveness of the systems.",
"Here, we only focus on the French-English and German-English pairs.",
"Following the task guidelines, we use the following evaluation measures investigate the effectiveness of our method: • Mean Average Precision (MAP).",
"Average of precisions computed at the point of each correctly paired document in the ranked list of paired documents.",
"• Success (Succ.).",
"Precision computed on the first returned paired document.",
"• Precision at 5 (P@5).",
"Precision computed on the 5 topmost paired documents.",
"Results Results are presented in Table 3 .",
"Overall, we observe that the two strategies that filter out multiply assigned source documents improve the performance of the method.",
"The largest part of the improvement comes from using pigeonhole reasoning.",
"The use of cross-lingual information to Table 3 : Performance in terms of MAP, success (Succ.)",
"and precision at 5 (P@5) of our model.",
"break ties between the remaining multiply assigned source documents only gives a small improvement.",
"We assume that the limited number of potential source-target document pairs we use in our experiments (N = 20) is a reason for this.",
"Interestingly, results are consistent across languages and datasets (test and train).",
"Our best configuration, that is, with pigeonhole and crosslingual, achieves nearly 60% of success for the first returned pair.",
"Here we show that a simple and straightforward approach that requires no language-specific resources still yields some interesting results.",
"Discussion In this paper we described the LINA system for the BUCC 2015 shared track.",
"We proposed to extend (Enright and Kondrak, 2007) 's approach to parallel document identification by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information.",
"Experimental results show that our system identifies comparable documents with a precision of about 60%.",
"Scoring document pairs using the number of shared hapax words was first intended to be a baseline for comparison purposes.",
"We tried a finer grained scoring approach relying on bilingual dictionaries and information retrieval weighting schemes.",
"For reasonable computation time, we were unable to include low-frequency words in our system.",
"Partial results were very low and we are still in the process of investigating the reasons for this."
]
} | {
"paper_header_number": [
"1",
"2",
"3.1",
"3.2",
"4"
],
"paper_header_content": [
"Introduction",
"Proposed Method",
"Experimental settings",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-2#paper-957#slide-6 | Results DE EN | Strategy MAP Succ. P@5 MAP Succ. P@5 | Strategy MAP Succ. P@5 MAP Succ. P@5 | [] |
GEM-SciDuet-train-2#paper-957#slide-7 | 957 | LINA: Identifying Comparable Documents from Wikipedia | This paper describes the LINA system for the BUCC 2015 shared track. Following (Enright and Kondrak, 2007), our system identify comparable documents by collecting counts of hapax words. We extend this method by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53
],
"paper_content_text": [
"Introduction Parallel corpora, that is, collections of documents that are mutual translations, are used in many natural language processing applications, particularly for statistical machine translation.",
"Building such resources is however exceedingly expensive, requiring highly skilled annotators or professional translators (Preiss, 2012) .",
"Comparable corpora, that are sets of texts in two or more languages without being translations of each other, are often considered as a solution for the lack of parallel corpora, and many techniques have been proposed to extract parallel sentences (Munteanu et al., 2004; Abdul-Rauf and Schwenk, 2009; Smith et al., 2010) , or mine word translations (Fung, 1995; Rapp, 1999; Chiao and Zweigenbaum, 2002; Morin et al., 2007; Vulić and Moens, 2012) .",
"Identifying comparable resources in a large amount of multilingual data remains a very challenging task.",
"The purpose of the Building and Using Comparable Corpora (BUCC) 2015 shared task 1 is to provide the first evaluation of existing approaches for identifying comparable resources.",
"More precisely, given a large collection of Wikipedia pages in several languages, the task is to identify the most similar pages across languages.",
"1 https://comparable.limsi.fr/bucc2015/ In this paper, we describe the system that we developed for the BUCC 2015 shared track and show that a language agnostic approach can achieve promising results.",
"Proposed Method The method we propose is based on (Enright and Kondrak, 2007) 's approach to parallel document identification.",
"Documents are treated as bags of words, in which only blank separated strings that are at least four characters long and that appear only once in the document (hapax words) are indexed.",
"Given a document in language A, the document in language B that share the largest number of these words is considered as parallel.",
"Although very simple, this approach was shown to perform very well in detecting parallel documents in Wikipedia (Patry and Langlais, 2011) .",
"The reason for this is that most hapax words are in practice proper nouns or numerical entities, which are often cognates.",
"An example of hapax words extracted from a document is given in Table 1 .",
"We purposely keep urls and special characters, as these are useful clues for identifying translated Wikipedia pages.",
"website major gaston links flutist marcel debost states sources college crunelle conservatoire principal rampal united currently recorded chastain competitions music http://www.oberlin.edu/faculty/mdebost/ under international flutists jean-pierre profile moyse french repertoire amazon lives external *http://www.amazon.com/micheldebost/dp/b000s9zsk0 known teaches conservatory school professor studied kathleen orchestre replaced michel Here, we experiment with this approach for detecting near-parallel (comparable) documents.",
"Following (Patry and Langlais, 2011) , we first search for the potential source-target document pairs.",
"To do so, we select for each document in the source language, the N = 20 documents in the target language that share the largest number of hapax words (hereafter baseline).",
"Scoring each pair of documents independently of other candidate pairs leads to several source documents being paired to a same target document.",
"As indicated in Table 2 , the percentage of English articles that are paired with multiple source documents is high (57.3% for French and 60.4% for German).",
"To address this problem, we remove potential multiple source documents by keeping the document pairs with the highest number of shared words (hereafter pigeonhole).",
"This strategy greatly reduces the number of multiply assigned source documents from roughly 60% to 10%.",
"This in turn removes needlessly paired documents and greatly improves the effectiveness of the method.",
"In an attempt to break the remaining score ties between document pairs, we further extend our model to exploit cross-lingual information.",
"When multiple source documents are paired to a given English document with the same score, we use the paired documents in a third language to order them (hereafter cross-lingual).",
"Here we make two assumptions that are valid for the BUCC 2015 shared Task: (1) we have access to comparable documents in a third language, and (2) source documents should be paired 1-to-1 with target documents.",
"Strategy An example of two French documents (doc fr 1 and doc fr 2) being paired to the same English document (doc en ) is given in Figure 1 .",
"We use the German document (doc de ) paired with doc en and select the French document that shares the largest number of hapax words, which for this example is doc fr 2.",
"This strategy further reduces the number of multiply assigned source documents from 10% to less than 4%.",
"Experiments Experimental settings The BUCC 2015 shared task consists in returning for each Wikipedia page in a source language, up to five ranked suggestions to its linked page in English.",
"Inter-language links, that is, links from a page in one language to an equivalent page in another language, are used to evaluate the effectiveness of the systems.",
"Here, we only focus on the French-English and German-English pairs.",
"Following the task guidelines, we use the following evaluation measures investigate the effectiveness of our method: • Mean Average Precision (MAP).",
"Average of precisions computed at the point of each correctly paired document in the ranked list of paired documents.",
"• Success (Succ.).",
"Precision computed on the first returned paired document.",
"• Precision at 5 (P@5).",
"Precision computed on the 5 topmost paired documents.",
"Results Results are presented in Table 3 .",
"Overall, we observe that the two strategies that filter out multiply assigned source documents improve the performance of the method.",
"The largest part of the improvement comes from using pigeonhole reasoning.",
"The use of cross-lingual information to Table 3 : Performance in terms of MAP, success (Succ.)",
"and precision at 5 (P@5) of our model.",
"break ties between the remaining multiply assigned source documents only gives a small improvement.",
"We assume that the limited number of potential source-target document pairs we use in our experiments (N = 20) is a reason for this.",
"Interestingly, results are consistent across languages and datasets (test and train).",
"Our best configuration, that is, with pigeonhole and crosslingual, achieves nearly 60% of success for the first returned pair.",
"Here we show that a simple and straightforward approach that requires no language-specific resources still yields some interesting results.",
"Discussion In this paper we described the LINA system for the BUCC 2015 shared track.",
"We proposed to extend (Enright and Kondrak, 2007) 's approach to parallel document identification by filtering out document pairs sharing target documents using pigeonhole reasoning and cross-lingual information.",
"Experimental results show that our system identifies comparable documents with a precision of about 60%.",
"Scoring document pairs using the number of shared hapax words was first intended to be a baseline for comparison purposes.",
"We tried a finer grained scoring approach relying on bilingual dictionaries and information retrieval weighting schemes.",
"For reasonable computation time, we were unable to include low-frequency words in our system.",
"Partial results were very low and we are still in the process of investigating the reasons for this."
]
} | {
"paper_header_number": [
"1",
"2",
"3.1",
"3.2",
"4"
],
"paper_header_content": [
"Introduction",
"Proposed Method",
"Experimental settings",
"Results",
"Discussion"
]
} | GEM-SciDuet-train-2#paper-957#slide-7 | Summary | I Unsupervised, hapax words-based method
I Promising results, about 60% of success using pigeonhole reasoning
I Using a third language slightly improves the performance
I Finding the optimal alignment across the all languages
I Relaxing the hapax-words constraint | I Unsupervised, hapax words-based method
I Promising results, about 60% of success using pigeonhole reasoning
I Using a third language slightly improves the performance
I Finding the optimal alignment across the all languages
I Relaxing the hapax-words constraint | [] |
GEM-SciDuet-train-3#paper-964#slide-0 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-0 | Sentence Representation in Conversations | Traditional System: hand-crafted semantic frame
Not scalable to complex domains
Neural dialog models: continuous hidden vectors
Directly output system responses in words
Hard to interpret & control
[Ritter et al 2011, Vinyals et al | Traditional System: hand-crafted semantic frame
Not scalable to complex domains
Neural dialog models: continuous hidden vectors
Directly output system responses in words
Hard to interpret & control
[Ritter et al 2011, Vinyals et al | [] |
GEM-SciDuet-train-3#paper-964#slide-1 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-1 | Why discrete sentence representation | 1. Inrepteablity & controbility & multimodal distribution
2. Semi-supervised Learning [Kingma et al 2014 NIPS, Zhou et al 2017 ACL]
3. Reinforcement Learning [Wen et al 2017]
X = What time do you want to travel?
Model Z1Z2Z3 Encoder Decoder | 1. Inrepteablity & controbility & multimodal distribution
2. Semi-supervised Learning [Kingma et al 2014 NIPS, Zhou et al 2017 ACL]
3. Reinforcement Learning [Wen et al 2017]
X = What time do you want to travel?
Model Z1Z2Z3 Encoder Decoder | [] |
GEM-SciDuet-train-3#paper-964#slide-2 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-2 | Baseline Discrete Variational Autoencoder VAE | M discrete K-way latent variables z with RNN recognition & generation network.
Reparametrization using Gumbel-Softmax [Jang et al., 2016; Maddison et al., 2016]
M discrete K-way latent variables z with GRU encoder & decoder.
FAIL to learn meaningful z because of posterior collapse (z is constant regardless of x)
MANY prior solution on continuous VAE, e.g. (not exhaustive), yet still open-ended question
KL-annealing, decoder word dropout [Bowman et a2015] Bag-of-word loss [Zhao et al 2017] Dilated CNN decoder | M discrete K-way latent variables z with RNN recognition & generation network.
Reparametrization using Gumbel-Softmax [Jang et al., 2016; Maddison et al., 2016]
M discrete K-way latent variables z with GRU encoder & decoder.
FAIL to learn meaningful z because of posterior collapse (z is constant regardless of x)
MANY prior solution on continuous VAE, e.g. (not exhaustive), yet still open-ended question
KL-annealing, decoder word dropout [Bowman et a2015] Bag-of-word loss [Zhao et al 2017] Dilated CNN decoder | [] |
GEM-SciDuet-train-3#paper-964#slide-3 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-3 | Anti Info Nature in Evidence Lower Bound ELBO | Write ELBO as an expectation over the whole dataset
Expand the KL term, and plug back in:
Minimize I(Z, X) to 0
Posterior collapse with powerful decoder. | Write ELBO as an expectation over the whole dataset
Expand the KL term, and plug back in:
Minimize I(Z, X) to 0
Posterior collapse with powerful decoder. | [] |
GEM-SciDuet-train-3#paper-964#slide-4 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-4 | Discrete Information VAE DI VAE | A natural solution is to maximize both data log likelihood & mutual information.
Match prior result for continuous VAE. [Mazhazni et al 2015, Kim et al 2017]
Propose Batch Prior Regularization (BPR) to minimize KL [q(z)||p(z)] for discrete latent
Fundamentally different from KL-annealing, since | A natural solution is to maximize both data log likelihood & mutual information.
Match prior result for continuous VAE. [Mazhazni et al 2015, Kim et al 2017]
Propose Batch Prior Regularization (BPR) to minimize KL [q(z)||p(z)] for discrete latent
Fundamentally different from KL-annealing, since | [] |
GEM-SciDuet-train-3#paper-964#slide-5 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-5 | Learning from Context Predicting DI VST | Skip-Thought (ST) is well-known distributional sentence representation [Hill et al 2016]
The meaning of sentences in dialogs is highly contextual, e.g. dialog acts.
We extend DI-VAE to Discrete Information Variational Skip Thought (DI-VST). | Skip-Thought (ST) is well-known distributional sentence representation [Hill et al 2016]
The meaning of sentences in dialogs is highly contextual, e.g. dialog acts.
We extend DI-VAE to Discrete Information Variational Skip Thought (DI-VST). | [] |
GEM-SciDuet-train-3#paper-964#slide-6 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-6 | Integration with Encoder Decoders | Policy Network z P(z|c)
Recognition Network z Generator
Optional: penalize decoder if generated x not exhibiting z [Hu et al 2017] | Policy Network z P(z|c)
Recognition Network z Generator
Optional: penalize decoder if generated x not exhibiting z [Hu et al 2017] | [] |
GEM-SciDuet-train-3#paper-964#slide-7 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-7 | Evaluation Datasets | a. Past evaluation dataset for text VAE [Bowman et al 2015]
Stanford Multi-domain Dialog Dataset (SMD) [Eric and Manning 2017]
a. 3,031 Human-Woz dialog dataset from 3 domains: weather, navigation & scheduling.
Switchboard (SW) [Jurafsky et al 1997]
a. 2,400 human-human telephone non-task-oriented dialogues about a given topic.
a. 13,188 human-human non-task-oriented dialogs from chat room. | a. Past evaluation dataset for text VAE [Bowman et al 2015]
Stanford Multi-domain Dialog Dataset (SMD) [Eric and Manning 2017]
a. 3,031 Human-Woz dialog dataset from 3 domains: weather, navigation & scheduling.
Switchboard (SW) [Jurafsky et al 1997]
a. 2,400 human-human telephone non-task-oriented dialogues about a given topic.
a. 13,188 human-human non-task-oriented dialogs from chat room. | [] |
GEM-SciDuet-train-3#paper-964#slide-8 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-8 | The Effectiveness of Batch Prior Regularization BPR | DAE: Autoencoder + Gumbel Softmax
DVAE: Discrete VAE with ELBO loss
DI-VAE: Discrete VAE + BPR
DST: Skip thought + Gumbel Softmax
DI-VST: Variational Skip Thought + BPR Table 1: Results for various discrete sentence representations. | DAE: Autoencoder + Gumbel Softmax
DVAE: Discrete VAE with ELBO loss
DI-VAE: Discrete VAE + BPR
DST: Skip thought + Gumbel Softmax
DI-VST: Variational Skip Thought + BPR Table 1: Results for various discrete sentence representations. | [] |
GEM-SciDuet-train-3#paper-964#slide-9 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-9 | How large should the batch size be | When batch size N = 0
A large batch size leads to
more meaningful latent action z
I(x,z) is not the final goal | When batch size N = 0
A large batch size leads to
more meaningful latent action z
I(x,z) is not the final goal | [] |
GEM-SciDuet-train-3#paper-964#slide-11 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-11 | Differences between DI VAE DI VST | DI-VAE cluster utterances based on the
More error-prone since harder to predict
Utterance used in the similar context
Easier to get agreement. | DI-VAE cluster utterances based on the
More error-prone since harder to predict
Utterance used in the similar context
Easier to get agreement. | [] |
GEM-SciDuet-train-3#paper-964#slide-12 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-12 | Interpreting Latent Actions | M=3, K=5. The trained R will map any utterance into a1 -a2 -a3 . E.g. How are you?
Automatic Evaluation on SW & DD
Compare latent actions with
The higher the more correlated
Human Evaluation on SMD
Expert look at 5 examples and give a
name to the latent actions
5 workers look at the expert name and
Select the ones that match the expert | M=3, K=5. The trained R will map any utterance into a1 -a2 -a3 . E.g. How are you?
Automatic Evaluation on SW & DD
Compare latent actions with
The higher the more correlated
Human Evaluation on SMD
Expert look at 5 examples and give a
name to the latent actions
5 workers look at the expert name and
Select the ones that match the expert | [] |
GEM-SciDuet-train-3#paper-964#slide-13 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-13 | Predict Latent Action by the Policy Network | Provide useful measure about the
complexity of the domain.
Usr > Sys & Chat > Task
Predict latent actions from DI-VAE is harder
than the ones from DI-VST
Two types of latent actions has their own
pros & cons. Which one is better is | Provide useful measure about the
complexity of the domain.
Usr > Sys & Chat > Task
Predict latent actions from DI-VAE is harder
than the ones from DI-VST
Two types of latent actions has their own
pros & cons. Which one is better is | [] |
GEM-SciDuet-train-3#paper-964#slide-14 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-14 | Interpretable Response Generation | Examples of interpretable dialog
First time, a neural dialog system | Examples of interpretable dialog
First time, a neural dialog system | [] |
GEM-SciDuet-train-3#paper-964#slide-15 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-15 | Conclusions and Future Work | An analysis of ELBO that explains the posterior collapse issue for sentence VAE.
DI-VAE and DI-VST for learning rich sentence latent representation and integration
Learn better context-based latent actions
Encode human knowledge into the learning process.
Learn structured latent action space for complex domains.
Evaluate dialog generation performance in human-study. | An analysis of ELBO that explains the posterior collapse issue for sentence VAE.
DI-VAE and DI-VST for learning rich sentence latent representation and integration
Learn better context-based latent actions
Encode human knowledge into the learning process.
Learn structured latent action space for complex domains.
Evaluate dialog generation performance in human-study. | [] |
GEM-SciDuet-train-3#paper-964#slide-16 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-16 | Semantic Consistency of the Generation | Use the recognition network as a classifier to
predict the latent action z based on the
Report accuracy by comparing z and z.
DI-VAE has higher consistency than DI-VST
L helps more in complex domain attr
L helps DI-VST more than DI-VAE attr
DI-VST is not directly helping generating x
ST-ED doesnt work well on SW due to complex
Spoken language and turn taking | Use the recognition network as a classifier to
predict the latent action z based on the
Report accuracy by comparing z and z.
DI-VAE has higher consistency than DI-VST
L helps more in complex domain attr
L helps DI-VST more than DI-VAE attr
DI-VST is not directly helping generating x
ST-ED doesnt work well on SW due to complex
Spoken language and turn taking | [] |
GEM-SciDuet-train-3#paper-964#slide-17 | 964 | Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation | The encoder-decoder dialog model is one of the most prominent methods used to build dialog systems in complex domains. Yet it is limited because it cannot output interpretable actions as in traditional systems, which hinders humans from understanding its generation process. We present an unsupervised discrete sentence representation learning method that can integrate with any existing encoderdecoder dialog models for interpretable response generation. Building upon variational autoencoders (VAEs), we present two novel models, DI-VAE and DI-VST that improve VAEs and can discover interpretable semantics via either auto encoding or context predicting. Our methods have been validated on real-world dialog datasets to discover semantic representations and enhance encoder-decoder models with interpretable generation. 1 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167,
168,
169,
170,
171,
172,
173,
174,
175,
176,
177,
178,
179,
180,
181,
182,
183,
184,
185,
186,
187,
188,
189,
190,
191,
192,
193,
194,
195,
196,
197,
198,
199,
200,
201,
202,
203,
204,
205,
206,
207,
208,
209,
210,
211,
212,
213,
214,
215,
216,
217,
218,
219,
220,
221,
222,
223,
224,
225,
226,
227,
228,
229,
230
],
"paper_content_text": [
"Introduction Classic dialog systems rely on developing a meaning representation to represent the utterances from both the machine and human users (Larsson and Traum, 2000; Bohus et al., 2007) .",
"The dialog manager of a conventional dialog system outputs the system's next action in a semantic frame that usually contains hand-crafted dialog acts and slot values (Williams and Young, 2007) .",
"Then a natural language generation module is used to generate the system's output in natural language based on the given semantic frame.",
"This approach suffers from generalization to more complex domains because it soon become intractable to man-ually design a frame representation that covers all of the fine-grained system actions.",
"The recently developed neural dialog system is one of the most prominent frameworks for developing dialog agents in complex domains.",
"The basic model is based on encoder-decoder networks and can learn to generate system responses without the need for hand-crafted meaning representations and other annotations.",
"Although generative dialog models have advanced rapidly (Serban et al., 2016; Li et al., 2016; , they cannot provide interpretable system actions as in the conventional dialog systems.",
"This inability limits the effectiveness of generative dialog models in several ways.",
"First, having interpretable system actions enables human to understand the behavior of a dialog system and better interpret the system intentions.",
"Also, modeling the high-level decision-making policy in dialogs enables useful generalization and dataefficient domain adaptation (Gašić et al., 2010) .",
"Therefore, the motivation of this paper is to develop an unsupervised neural recognition model that can discover interpretable meaning representations of utterances (denoted as latent actions) as a set of discrete latent variables from a large unlabelled corpus as shown in Figure 1 .",
"The discovered meaning representations will then be integrated with encoder decoder networks to achieve interpretable dialog generation while preserving all the merit of neural dialog systems.",
"We focus on learning discrete latent representations instead of dense continuous ones because discrete variables are easier to interpret (van den Oord et al., 2017) and can naturally correspond to categories in natural languages, e.g.",
"topics, dialog acts and etc.",
"Despite the difficulty of learning discrete latent variables in neural networks, the recently proposed Gumbel-Softmax offers a reliable way to back-propagate through discrete variables (Maddison et al., 2016; Jang et al., 2016) .",
"However, we found a simple combination of sentence variational autoencoders (VAEs) (Bowman et al., 2015) and Gumbel-Softmax fails to learn meaningful discrete representations.",
"We then highlight the anti-information limitation of the evidence lowerbound objective (ELBO) in VAEs and improve it by proposing Discrete Information VAE (DI-VAE) that maximizes the mutual information between data and latent actions.",
"We further enrich the learning signals beyond auto encoding by extending Skip Thought (Kiros et al., 2015) to Discrete Information Variational Skip Thought (DI-VST) that learns sentence-level distributional semantics.",
"Finally, an integration mechanism is presented that combines the learned latent actions with encoder decoder models.",
"The proposed systems are tested on several realworld dialog datasets.",
"Experiments show that the proposed methods significantly outperform the standard VAEs and can discover meaningful latent actions from these datasets.",
"Also, experiments confirm the effectiveness of the proposed integration mechanism and show that the learned latent actions can control the sentence-level attributes of the generated responses and provide humaninterpretable meaning representations.",
"Related Work Our work is closely related to research in latent variable dialog models.",
"The majority of models are based on Conditional Variational Autoencoders (CVAEs) (Serban et al., 2016; Cao and Clark, 2017) with continuous latent variables to better model the response distribution and encourage diverse responses.",
"further introduced dialog acts to guide the learning of the CVAEs.",
"Discrete latent variables have also been used for task-oriented dialog systems (Wen et al., 2017) , where the latent space is used to represent intention.",
"The second line of related work is enriching the dialog context encoder with more fine-grained information than the dialog history.",
"Li et al., (2016) captured speakers' characteristics by encoding background information and speaking style into the distributed embeddings.",
"Xing et al., (2016) maintain topic encoding based on Latent Dirichlet Allocation (LDA) (Blei et al., 2003) of the conversation to encourage the model to output more topic coherent responses.",
"The proposed method also relates to sentence representation learning using neural networks.",
"Most work learns continuous distributed representations of sentences from various learning signals (Hill et al., 2016) , e.g.",
"the Skip Thought learns representations by predicting the previous and next sentences (Kiros et al., 2015) .",
"Another area of work focused on learning regularized continuous sentence representation, which enables sentence generation by sampling the latent space (Bowman et al., 2015; Kim et al., 2017) .",
"There is less work on discrete sentence representations due to the difficulty of passing gradients through discrete outputs.",
"The recently developed Gumbel Softmax (Jang et al., 2016; Maddison et al., 2016) and vector quantization (van den Oord et al., 2017) enable us to train discrete variables.",
"Notably, discrete variable models have been proposed to discover document topics (Miao et al., 2016) and semi-supervised sequence transaction (Zhou and Neubig, 2017) Our work differs from these as follows: (1) we focus on learning interpretable variables; in prior research the semantics of latent variables are mostly ignored in the dialog generation setting.",
"(2) we improve the learning objective for discrete VAEs and overcome the well-known posterior collapsing issue (Bowman et al., 2015; Chen et al., 2016) .",
"(3) we focus on unsupervised learning of salient features in dialog responses instead of hand-crafted features.",
"Proposed Methods Our formulation contains three random variables: the dialog context c, the response x and the latent action z.",
"The context often contains the discourse history in the format of a list of utterances.",
"The response is an utterance that contains a list of word tokens.",
"The latent action is a set of discrete variables that define high-level attributes of x.",
"Before introducing the proposed framework, we first identify two key properties that are essential in or-der for z to be interpretable: 1. z should capture salient sentence-level features about the response x.",
"2.",
"The meaning of latent symbols z should be independent of the context c. The first property is self-evident.",
"The second can be explained: assume z contains a single discrete variable with K classes.",
"Since the context c can be any dialog history, if the meaning of each class changes given a different context, then it is difficult to extract an intuitive interpretation by only looking at all responses with class k ∈ [1, K].",
"Therefore, the second property looks for latent actions that have context-independent semantics so that each assignment of z conveys the same meaning in all dialog contexts.",
"With the above definition of interpretable latent actions, we first introduce a recognition network R : q R (z|x) and a generation network G. The role of R is to map an sentence to the latent variable z and the generator G defines the learning signals that will be used to train z's representation.",
"Notably, our recognition network R does not depend on the context c as has been the case in prior work (Serban et al., 2016) .",
"The motivation of this design is to encourage z to capture context-independent semantics, which are further elaborated in Section 3.4.",
"With the z learned by R and G, we then introduce an encoder decoder network F : p F (x|z, c) and and a policy network π : p π (z|c).",
"At test time, given a context c, the policy network and encoder decoder will work together to generate the next response viã x = p F (x|z ∼ p π (z|c), c).",
"In short, R, G, F and π are the four components that comprise our proposed framework.",
"The next section will first focus on developing R and G for learning interpretable z and then will move on to integrating R with F and π in Section 3.3.",
"Learning Sentence Representations from Auto-Encoding Our baseline model is a sentence VAE with discrete latent space.",
"We use an RNN as the recognition network to encode the response x.",
"Its last hidden state h R |x| is used to represent x.",
"We define z to be a set of K-way categorical variables z = {z 1 ...z m ...z M }, where M is the number of variables.",
"For each z m , its posterior distribution is defined as q R (z m |x) = Softmax(W q h R |x| + b q ).",
"During training, we use the Gumbel-Softmax trick to sample from this distribution and obtain lowvariance gradients.",
"To map the latent samples to the initial state of the decoder RNN, we define {e 1 ...e m ...e M } where e m ∈ R K×D and D is the generator cell size.",
"Thus the initial state of the generator is: h G 0 = M m=1 e m (z m ).",
"Finally, the generator RNN is used to reconstruct the response given h G 0 .",
"VAEs is trained to maxmimize the evidence lowerbound objective (ELBO) (Kingma and Welling, 2013) .",
"For simplicity, later discussion drops the subscript m in z m and assumes a single latent z.",
"Since each z m is independent, we can easily extend the results below to multiple variables.",
"Anti-Information Limitation of ELBO It is well-known that sentence VAEs are hard to train because of the posterior collapse issue.",
"Many empirical solutions have been proposed: weakening the decoder, adding auxiliary loss etc.",
"(Bowman et al., 2015; Chen et al., 2016; .",
"We argue that the posterior collapse issue lies in ELBO and we offer a novel decomposition to understand its behavior.",
"First, instead of writing ELBO for a single data point, we write it as an expectation over a dataset: L VAE = E x [E q R (z|x) [log p G (x|z)] − KL(q R (z|x) p(z))] (1) We can expand the KL term as Eq.",
"2 (derivations in Appendix A.1) and rewrite ELBO as: E x [KL(q R (z|x) p(z))] = (2) I(Z, X)+KL(q(z) p(z)) L VAE = E q(z|x)p(x) [log p(x|z)] − I(Z, X) − KL(q(z) p(z)) (3) where q(z) = E x [q R (z|x)] and I(Z, X) is the mutual information between Z and X.",
"This expansion shows that the KL term in ELBO is trying to reduce the mutual information between latent variables and the input data, which explains why VAEs often ignore the latent variable, especially when equipped with powerful decoders.",
"VAE with Information Maximization and Batch Prior Regularization A natural solution to correct the anti-information issue in Eq.",
"3 is to maximize both the data likeli-hood lowerbound and the mutual information between z and the input data: L VAE + I(Z, X) = E q R (z|x)p(x) [log p G (x|z)] − KL(q(z) p(z)) (4) Therefore, jointly optimizing ELBO and mutual information simply cancels out the informationdiscouraging term.",
"Also, we can still sample from the prior distribution for generation because of KL(q(z) p(z)).",
"Eq.",
"4 is similar to the objectives used in adversarial autoencoders (Makhzani et al., 2015; Kim et al., 2017) .",
"Our derivation provides a theoretical justification to their superior performance.",
"Notably, Eq.",
"4 arrives at the same loss function proposed in infoVAE (Zhao S et al., 2017) .",
"However, our derivation is different, offering a new way to understand ELBO behavior.",
"The remaining challenge is how to minimize KL(q(z) p(z)), since q(z) is an expectation over q(z|x).",
"When z is continuous, prior work has used adversarial training (Makhzani et al., 2015; Kim et al., 2017) or Maximum Mean Discrepancy (MMD) (Zhao S et al., 2017) to regularize q(z).",
"It turns out that minimizing KL(q(z) p(z)) for discrete z is much simpler than its continuous counterparts.",
"Let x n be a sample from a batch of N data points.",
"Then we have: q(z) ≈ 1 N N n=1 q(z|x n ) = q (z) (5) where q (z) is a mixture of softmax from the posteriors q(z|x n ) of each x n .",
"We can approximate KL(q(z) p(z)) by: KL(q (z) p(z)) = K k=1 q (z = k) log q (z = k) p(z = k) (6) We refer to Eq.",
"6 as Batch Prior Regularization (BPR).",
"When N approaches infinity, q (z) approaches the true marginal distribution of q(z).",
"In practice, we only need to use the data from each mini-batch assuming that the mini batches are randomized.",
"Last, BPR is fundamentally different from multiplying a coefficient < 1 to anneal the KL term in VAE (Bowman et al., 2015) .",
"This is because BPR is a non-linear operation log sum exp.",
"For later discussion, we denote our discrete infoVAE with BPR as DI-VAE.",
"Learning Sentence Representations from the Context DI-VAE infers sentence representations by reconstruction of the input sentence.",
"Past research in distributional semantics has suggested the meaning of language can be inferred from the adjacent context (Harris, 1954; Hill et al., 2016) .",
"The distributional hypothesis is especially applicable to dialog since the utterance meaning is highly contextual.",
"For example, the dialog act is a wellknown utterance feature and depends on dialog state (Austin, 1975; Stolcke et al., 2000) .",
"Thus, we introduce a second type of latent action based on sentence-level distributional semantics.",
"Skip thought (ST) is a powerful sentence representation that captures contextual information (Kiros et al., 2015) .",
"ST uses an RNN to encode a sentence, and then uses the resulting sentence representation to predict the previous and next sentences.",
"Inspired by ST's robust performance across multiple tasks (Hill et al., 2016) , we adapt our DI-VAE to Discrete Information Variational Skip Thought (DI-VST) to learn discrete latent actions that model distributional semantics of sentences.",
"We use the same recognition network from DI-VAE to output z's posterior distribution q R (z|x).",
"Given the samples from q R (z|x), two RNN generators are used to predict the previous sentence x p and the next sentences x n .",
"Finally, the learning objective is to maximize: L DI-VST = E q R (z|x)p(x)) [log(p n G (x n |z)p p G (x p |z))] − KL(q(z) p(z)) (7) Integration with Encoder Decoders We now describe how to integrate a given q R (z|x) with an encoder decoder and a policy network.",
"Let the dialog context c be a sequence of utterances.",
"Then a dialog context encoder network can encode the dialog context into a distributed representation h e = F e (c).",
"The decoder F d can generate the responsesx = F d (h e , z) using samples from q R (z|x).",
"Meanwhile, we train π to predict the aggregated posterior E p(x|c) [q R (z|x)] from c via maximum likelihood training.",
"This model is referred as Latent Action Encoder Decoder (LAED) with the following objective.",
"L LAED (θ F , θ π ) = E q R (z|x)p(x,c) [logp π (z|c) + log p F (x|z, c)] (8) Also simply augmenting the inputs of the decoders with latent action does not guarantee that the generated response exhibits the attributes of the give action.",
"Thus we use the controllable text generation framework (Hu et al., 2017) by introducing L Attr , which reuses the same recognition network q R (z|x) as a fixed discriminator to penalize the decoder if its generated responses do not reflect the attributes in z. L Attr (θ F ) = E q R (z|x)p(c,x) [log q R (z|F(c, z))] (9) Since it is not possible to propagate gradients through the discrete outputs at F d at each word step, we use a deterministic continuous relaxation (Hu et al., 2017) by replacing output of F d with the probability of each word.",
"Let o t be the normalized probability at step t ∈ [1, |x|], the inputs to q R at time t are then the sum of word embeddings weighted by o t , i.e.",
"h R t = RNN(h R t−1 , Eo t ) and E is the word embedding matrix.",
"Finally this loss is combined with L LAED and a hyperparameter λ to have Attribute Forcing LAED.",
"L attrLAED = L LAED + λL Attr (10) Relationship with Conditional VAEs It is not hard to see L LAED is closely related to the objective of CVAEs for dialog generation (Serban et al., 2016; , which is: L CVAE = E q [log p(x|z, c)]−KL(q(z|x, c) p(z|c)) (11) Despite their similarities, we highlight the key differences that prohibit CVAE from achieving interpretable dialog generation.",
"First L CVAE encourages I(x, z|c) (Agakov, 2005), which learns z that capture context-dependent semantics.",
"More intuitively, z in CVAE is trained to generate x via p(x|z, c) so the meaning of learned z can only be interpreted along with its context c. Therefore this violates our goal of learning context-independent semantics.",
"Our methods learn q R (z|x) that only depends on x and trains q R separately to ensure the semantics of z are interpretable standalone.",
"Experiments and Results The proposed methods are evaluated on four datasets.",
"The first corpus is Penn Treebank (PTB) (Marcus et al., 1993) used to evaluate sentence VAEs (Bowman et al., 2015) .",
"We used the version pre-processed by Mikolov (Mikolov et al., 2010) .",
"The second dataset is the Stanford Multi-Domain Dialog (SMD) dataset that contains 3,031 human-Woz, task-oriented dialogs collected from 3 different domains (navigation, weather and scheduling) (Eric and Manning, 2017) .",
"The other two datasets are chat-oriented data: Daily Dialog (DD) and Switchboard (SW) (Godfrey and Holliman, 1997), which are used to test whether our methods can generalize beyond task-oriented dialogs but also to to open-domain chatting.",
"DD contains 13,118 multi-turn human-human dialogs annotated with dialog acts and emotions.",
"(Li et al., 2017) .",
"SW has 2,400 human-human telephone conversations that are annotated with topics and dialog acts.",
"SW is a more challenging dataset because it is transcribed from speech which contains complex spoken language phenomenon, e.g.",
"hesitation, self-repair etc.",
"Comparing Discrete Sentence Representation Models The first experiment used PTB and DD to evaluate the performance of the proposed methods in learning discrete sentence representations.",
"We implemented DI-VAE and DI-VST using GRU-RNN (Chung et al., 2014) and trained them using Adam (Kingma and Ba, 2014) .",
"Besides the proposed methods, the following baselines are compared.",
"Unregularized models: removing the KL(q|p) term from DI-VAE and DI-VST leads to a simple discrete autoencoder (DAE) and discrete skip thought (DST) with stochastic discrete hidden units.",
"ELBO models: the basic discrete sentence VAE (DVAE) or variational skip thought (DVST) that optimizes ELBO with regularization term KL(q(z|x) p(z)).",
"We found that standard training failed to learn informative latent actions for either DVAE or DVST because of the posterior collapse.",
"Therefore, KL-annealing (Bowman et al., 2015) and bag-of-word loss are used to force these two models learn meaningful representations.",
"We also include the results for VAE with continuous latent variables reported on the same PTB .",
"Additionally, we report the perplexity from a standard GRU-RNN language model (Zaremba et al., 2014) .",
"The evaluation metrics include reconstruction perplexity (PPL), KL(q(z) p(z)) and the mutual information between input data and latent vari-ables I (x, z) .",
"Intuitively a good model should achieve low perplexity and KL distance, and simultaneously achieve high I(x, z).",
"The discrete latent space for all models are M =20 and K=10.",
"Mini-batch size is 30.",
"Table 1 shows that all models achieve better perplexity than an RNNLM, which shows they manage to learn meaningful q(z|x).",
"First, for autoencoding models, DI-VAE is able to achieve the best results in all metrics compared other methods.",
"We found DAEs quickly learn to reconstruct the input but they are prone to overfitting during training, which leads to lower performance on the test data compared to DI-VAE.",
"Also, since there is no regularization term in the latent space, q(z) is very different from the p(z) which prohibits us from generating sentences from the latent space.",
"In fact, DI-VAE enjoys the same linear interpolation properties reported in (Bowman et al., 2015) (See Appendix A.2).",
"As for DVAEs, it achieves zero I(x, z) in standard training and only manages to learn some information when training with KL-annealing and bag-of-word loss.",
"On the other hand, our methods achieve robust performance without the need for additional processing.",
"Similarly, the proposed DI-VST is able to achieve the lowest PPL and similar KL compared to the strongly regularized DVST.",
"Interestingly, although DST is able to achieve the highest I(x, z), but PPL is not further improved.",
"These results confirm the effectiveness of the proposed BPR in terms of regularizing q(z) while learning meaningful posterior q(z|x).",
"In order to understand BPR's sensitivity to batch size N , a follow-up experiment varied the batch size from 2 to 60 (If N =1, DI-VAE is equivalent to DVAE).",
"Figure 2 show that as N increases, perplexity, I(x, z) monotonically improves, while KL(q p) only increases from 0 to 0.159.",
"After N > 30, the performance plateaus.",
"Therefore, using mini-batch is an efficient trade-off between q(z) estimation and computation speed.",
"The last experiment in this section investigates the relation between representation learning and the dimension of the latent space.",
"We set a fixed budget by restricting the maximum number of modes to be about 1000, i.e.",
"K M ≈ 1000.",
"We then vary the latent space size and report the same evaluation metrics.",
"Table 2 shows that models with multiple small latent variables perform significantly better than those with large and few latent variables.",
"K, M K M PPL KL(q p) I(x, z) 1000, 1 1000 75.61 0.032 0.335 10, 3 1000 71.42 0.071 0.607 4, 5 1024 68.43 0.088 0.809 Table 2 : DI-VAE on PTB with different latent dimensions under the same budget.",
"Interpreting Latent Actions The next question is to interpret the meaning of the learned latent action symbols.",
"To achieve this, the latent action of an utterance x n is obtained from a greedy mapping: a n = argmax k q R (z = k|x n ).",
"We set M =3 and K=5, so that there are at most 125 different latent actions, and each x n can now be represented by a 1 -a 2 -a 3 , e.g.",
"\"How are you?\"",
"→ 1-4-2.",
"Assuming that we have access to manually clustered data according to certain classes (e.g.",
"dialog acts), it is unfair to use classic cluster measures (Vinh et al., 2010) to evaluate the clusters from latent actions.",
"This is because the uniform prior p(z) evenly distribute the data to all possible latent actions, so that it is expected that frequent classes will be assigned to several latent actions.",
"Thus we utilize the homogeneity metric (Rosenberg and Hirschberg, 2007 ) that measures if each latent action contains only members of a single class.",
"We tested this on the SW and DD, which contain human annotated features and we report the latent actions' homogeneity w.r.t these features in Table 3 .",
"On DD, results show DI-VST SW DD Act Topic Act Emotion DI-VAE 0.48 0.08 0.18 0.09 DI-VST 0.33 0.13 0.34 0.12 works better than DI-VAE in terms of creating actions that are more coherent for emotion and dialog acts.",
"The results are interesting on SW since DI-VST performs worse on dialog acts than DI-VAE.",
"One reason is that the dialog acts in SW are more fine-grained (42 acts) than the ones in DD (5 acts) so that distinguishing utterances based on words in x is more important than the information in the neighbouring utterances.",
"We then apply the proposed methods to SMD which has no manual annotation and contains taskoriented dialogs.",
"Two experts are shown 5 randomly selected utterances from each latent action and are asked to give an action name that can describe as many of the utterances as possible.",
"Then an Amazon Mechanical Turk study is conducted to evaluate whether other utterances from the same latent action match these titles.",
"5 workers see the action name and a different group of 5 utterances from that latent action.",
"They are asked to select all utterances that belong to the given actions, which tests the homogeneity of the utterances falling in the same cluster.",
"Negative samples are included to prevent random selection.",
"Table 4 shows that both methods work well and DI-VST achieved better homogeneity than DI-VAE.",
"Since DI-VAE is trained to reconstruct its input and DI-VST is trained to model the context, they group utterances in different ways.",
"For example, DI-VST would group \"Can I get a restaurant\", \"I am looking for a restaurant\" into one action where Dialog Response Generation with Latent Actions Finally we implement an LAED as follows.",
"The encoder is a hierarchical recurrent encoder (Serban et al., 2016) with bi-directional GRU-RNNs as the utterance encoder and a second GRU-RNN as the discourse encoder.",
"The discourse encoder output its last hidden state h e |x| .",
"The decoder is another GRU-RNN and its initial state of the decoder is obtained by h d 0 = h e |x| + M m=1 e m (z m ), where z comes from the recognition network of the proposed methods.",
"The policy network π is a 2-layer multi-layer perceptron (MLP) that models p π (z|h e |x| ).",
"We use up to the previous 10 utterances as the dialog context and denote the LAED using DI-VAE latent actions as AE-ED and the one uses DI-VST as ST-ED.",
"First we need to confirm whether an LAED can generate responses that are consistent with the semantics of a given z.",
"To answer this, we use a pre-trained recognition network R to check if a generated response carries the attributes in the given action.",
"We generate dialog responses on a test dataset viax = F(z ∼ π(c), c) with greedy RNN decoding.",
"The generated responses are passed into the R and we measure attribute accuracy by countingx as correct if z = argmax k q R (k|x).",
"Table 6 : Results for attribute accuracy with and without attribute loss.",
"responses are highly consistent with the given latent actions.",
"Also, latent actions from DI-VAE achieve higher attribute accuracy than the ones from DI-VST, because z from auto-encoding is explicitly trained for x reconstruction.",
"Adding L attr is effective in forcing the decoder to take z into account during its generation, which helps the most in more challenging open-domain chatting data, e.g.",
"SW and DD.",
"The accuracy of ST-ED on SW is worse than the other two datasets.",
"The reason is that SW contains many short utterances that can be either a continuation of the same speaker or a new turn from the other speaker, whereas the responses in the other two domains are always followed by a different speaker.",
"The more complex context pattern in SW may require special treatment.",
"We leave it for future work.",
"The second experiment checks if the policy network π is able to predict the right latent action given just the dialog context.",
"We report both accuracy, i.e.",
"argmax k q R (k|x) = argmax k p π (k |c) and perplexity of p π (z|c).",
"The perplexity measure is more useful for open domain dialogs because decision-making in complex dialogs is often one-to-many given a similar context .",
"Table 7 shows the prediction scores on SMD AE-ED 3.045 (51.5% sys 52.4% usr 50.5%) ST-ED 1.695 (75.5% sys 82.1% usr 69.2%) DD SW AE-ED 4.47 (35.8%) 4.46 (31.68%) ST-ED 3.89 (47.5%) 3.68 (33.2%) Table 7 : Performance of policy network.",
"L attr is included in training.",
"the three dialog datasets.",
"These scores provide useful insights to understand the complexity of a dialog dataset.",
"For example, accuracy on opendomain chatting is harder than the task-oriented SMD data.",
"Also, it is intuitive that predicting system actions is easier than predicting user actions on SMD.",
"Also, in general the prediction scores for ST-ED are higher the ones for AE-ED.",
"The reason is related to our previous discussion about the granularity of the latent actions.",
"Since latent actions from DI-VST mainly model the the type of utterances used in certain types of context, it is easier for the policy network to predict latent actions from DI-VST.",
"Therefore, choosing the type of latent actions is a design choice and depends on the type of interpretability that is needed.",
"We finish with an example generated from the two variants of LAED on SMD as shown in Table 8 .",
"Given a dialog context, our systems are able to output a probability distribution over different latent actions that have interpretable meaning along with their natural language realizations.",
"c usr: Where does my friend live?",
"Model Action Generated Responses AE-ED give loc info -Tom is at 753 University Ave, and a road block.",
"p(z|c)=0.34 -Comfort Inn is at 7 miles away.",
"give user info -Your home address is 5671 barringer street.",
"p(z|c)=0.22 -Your home is at 10 ames street.",
"ST-ED give loc info -Jill's house is 8 miles away at 347 Alta Mesa Ave. p(z|c)=0.93 -Jill lives at 347 Alta Mesa Ave. Table 8 : Interpretable dialog generation on SMD with top probable latent actions.",
"AE-ED predicts more fine-grained but more error-prone actions.",
"Conclusion and Future Work This paper presents a novel unsupervised framework that enables the discovery of discrete latent actions and interpretable dialog response generation.",
"Our main contributions reside in the two sentence representation models DI-VAE and DI-VST, and their integration with the encoder decoder models.",
"Experiments show the proposed methods outperform strong baselines in learning discrete latent variables and showcase the effectiveness of interpretable dialog response generation.",
"Our findings also suggest promising future research directions, including learning better context-based latent actions and using reinforce-ment learning to adapt policy networks.",
"We believe that this work is an important step forward towards creating generative dialog models that can not only generalize to large unlabelled datasets in complex domains but also be explainable to human users."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"3.1",
"3.1.1",
"3.1.2",
"3.2",
"3.3",
"3.4",
"4",
"4.1",
"4.2",
"4.3",
"5"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Proposed Methods",
"Learning Sentence Representations from Auto-Encoding",
"Anti-Information Limitation of ELBO",
"VAE with Information Maximization and Batch Prior Regularization",
"Learning Sentence Representations from the Context",
"Integration with Encoder Decoders",
"Relationship with Conditional VAEs",
"Experiments and Results",
"Comparing Discrete Sentence Representation Models",
"Interpreting Latent Actions",
"Dialog Response Generation with Latent Actions",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-3#paper-964#slide-17 | What defines Interpretable Latent Actions | Definition: Latent action is a set of discrete variable that define the high-level attributes of
an utterance (sentence) X. Latent action is denoted as Z.
Z should capture salient sentence-level features about the response X.
The meaning of latent symbols Z should be independent of the context C.
If meaning of Z depends on C, then often impossible to interpret Z
Since the possible space of C is huge!
Conclusion: context-independent semantic ensures each assignment of z has the same
meaning in all context. | Definition: Latent action is a set of discrete variable that define the high-level attributes of
an utterance (sentence) X. Latent action is denoted as Z.
Z should capture salient sentence-level features about the response X.
The meaning of latent symbols Z should be independent of the context C.
If meaning of Z depends on C, then often impossible to interpret Z
Since the possible space of C is huge!
Conclusion: context-independent semantic ensures each assignment of z has the same
meaning in all context. | [] |
GEM-SciDuet-train-4#paper-965#slide-0 | 965 | Data Augmentation for Context-Sensitive Neural Lemmatization Using Inflection Tables and Raw Text | Lemmatization aims to reduce the sparse data problem by relating the inflected forms of a word to its dictionary form. Using context can help, both for unseen and ambiguous words. Yet most context-sensitive approaches require full lemma-annotated sentences for training, which may be scarce or unavailable in lowresource languages. In addition (as shown here), in a low-resource setting, a lemmatizer can learn more from n labeled examples of distinct words (types) than from n (contiguous) labeled tokens, since the latter contain far fewer distinct types. To combine the efficiency of type-based learning with the benefits of context, we propose a way to train a context-sensitive lemmatizer with little or no labeled corpus data, using inflection tables from the UniMorph project and raw text examples from Wikipedia that provide sentence contexts for the unambiguous UniMorph examples. Despite these being unambiguous examples, the model successfully generalizes from them, leading to improved results (both overall, and especially on unseen words) in comparison to a baseline that does not use context. Method Lematus (Bergmanis and Goldwater, 2018) is a neural sequence-to-sequence model with attention 2 Garrette et al. (2013) found the same for POS tagging. 3 Code and data: https://bitbucket.org/ tomsbergmanis/data_augumentation_um_wiki 4 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97
],
"paper_content_text": [
"Introduction Many lemmatizers work on isolated wordforms (Wicentowski, 2002; Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"Lemmatizing in context can improve accuracy on ambiguous and unseen words (Bergmanis and Goldwater, 2018) , but most systems for contextsensitive lemmatization must train on complete sentences labeled with POS and/or morphological tags as well as lemmas, and have only been tested with 20k-300k training tokens (Chrupała et al., 2008; Müller et al., 2015; Chakrabarty et al., 2017) .",
"1 1 The smallest of these corpora contains 20k tokens of Bengali annotated only with lemmas, which Chakrabarty et al.",
"(2017) reported took around two person months to create.",
"Intuitively, though, sentence-annotated data is inefficient for training a lemmatizer, especially in low-resource settings.",
"Training on (say) 1000 word types will provide far more information about a language's morphology than training on 1000 contiguous tokens, where fewer types are represented.",
"As noted above, sentence data can help with ambiguous and unseen words, but we show here that when data is scarce, this effect is small relative to the benefit of seeing more word types.",
"2 Motivated by this result, we propose a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive model.",
"3 We use Lematus (Bergmanis and Goldwater, 2018), a state-of-theart lemmatizer that learns from lemma-annotated words in their N -character contexts.",
"No predictions about surrounding words are used, so fully annotated training sentences are not needed.",
"We exploit this fact by combining two sources of training data: 1k lemma-annotated types (with contexts) from the Universal Dependency Treebank (UDT) v2.2 4 (Nivre et al., 2017) , plus examples obtained by finding unambiguous word-lemma pairs in inflection tables from the Universal Morphology (UM) project 5 and collecting sentence contexts for them from Wikipedia.",
"Although these examples are noisy and biased, we show that they improve lemmatization accuracy in experiments on 10 languages, and that the use of context helps, both overall and especially on unseen words.",
"inspired by the re-inflection model of Kann and Schütze (2016) , which won the 2016 SIGMOR-PHON shared task (Cotterell et al., 2016) .",
"It is built using the Nematus machine translation toolkit, 6 which uses the architecture of Sennrich et al.",
"(2017) : a 2-layer bidirectional GRU encoder and a 2-layer decoder with a conditional GRU (Sennrich et al., 2017) in the first layer and a GRU in the second layer.",
"Lematus takes as input a character sequence representing the wordform in its N -character context, and outputs the characters of the lemma.",
"Special input symbols are used to represent the left and right boundary of the target wordform (<lc>, <rc>) and other word boundaries (<s>).",
"For example, if N = 15, the system trained on Latvian would be expected to produce the characters of the lemma ceļš (meaning road) given input such as: s a k a <s> p aš v a l dī b u <lc> c e ļ u <rc> u n <s> i e l u <s> r eǵ i s t r When N = 0 (Lematus 0-ch), no context is used, making Lematus 0-ch comparable to other systems that do not model context (Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"In our experiments we use both Lematus 0-ch and Lematus 20-ch (20 characters of context), which was the best-performing system reported by Bergmanis and Goldwater (2018).",
"Data Augmentation Our data augmentation method uses UM inflection tables and creates additional training examples by finding Wikipedia sentences that use the inflected wordforms in context, pairing them with their lemma as shown in the inflection table.",
"However, we cannot use all the words in the tables because some of them are ambiguous: for example, Figure 1 shows that the form ceļi could be lemmatized either as ceļš or celis.",
"SG PL SG PL NOM ceļš ceļi celis ceļi GEN ceļa ceļu ceļa ceļu DAT ceļam ceļiem celim ceļiem ACC ceļu ceļus celi ceļus INS ceļu ceļiem celi ceļiem LOC ceļā ceļos celī ceļos VOC ceļ ceļi celi ceļi There are several other issues with this method that could potentially limit its usefulness.",
"First, the UM tables only include verbs, nouns and adjectives, whereas we test the system on UDT data, which includes all parts of speech.",
"Second, by excluding ambiguous forms, we may be restricting the added examples to a non-representative subset of the potential inflections, or the system may simply ignore the context because it isn't needed for these examples.",
"Finally, there are some annotation differences between UM and UDT.",
"7 Despite all of these issues, however, we show below that the added examples and their contexts do actually help.",
"Experimental Setup Baselines and Training Parameters We use four baselines: (1) Lemming 8 (Müller et al., 2015) is a context-sensitive system that uses log-linear models to jointly tag and lemmatize the data, and is trained on sentences annotated with both lemmas and POS tags.",
"(2) The hard monotonic attention model (HMAM) 9 (Makarov and Clematide, 2018b) is a neural sequence-tosequence model with a hard attention mechanism that advances through the sequence monotonically.",
"It is trained on word-lemma pairs (without context) 7 Recent efforts to unify the two resources have mostly focused on validating dataset schema (McCarthy et al., 2018) , leaving conflicts in word lemmas unresolved.",
"We estimated (by counting types that are unambiguous in each dataset but have different lemmas across them) that annotation inconsistencies affect up to 1% of types in the languages we used.",
"8 http://cistern.cis.lmu.de/lemming 9 https://github.com/ZurichNLP/ coling2018-neural-transition-basedmorphology with character-level alignments learned in a preprocessing step using an alignment model, and it has proved to be competitive in low resource scenarios.",
"(3) Our naive Baseline outputs the most frequent lemma (or one lemma at random from the options that are equally frequent) for words observed in training.",
"For unseen words it outputs the wordform itself.",
"(4) We also try a baseline data augmentation approach (AE Aug Baseline) inspired by Bergmanis et al.",
"(2017) and Kann and Schütze (2017) , who showed that adding training examples where the network simply learns to auto-encode corpus words can improve morphological inflection results in low-resource settings.",
"The AE Aug Baseline is a variant of Lematus 0-ch which augments the UDT lemmatization examples by auto-encoding the inflected forms of the UM examples (i.e., it just treats them as corpus words).",
"Comparing AE Aug Baseline to Lematus 0-ch augmented with UM lemma-inflection examples tells us whether using the UM lemma information helps more than simply auto-encoding more inflected examples.",
"To train the models we use the default settings for Lemming and the suggested lemmatization parameters for HMAM.",
"We mainly follow the hyperparameters used by Bergmanis and Goldwater (2018) for Lematus; details are in Appendix B.",
"Languages and Training Data We conduct preliminary experiments on five development languages: Estonian, Finnish, Latvian, Polish, and Russian.",
"In our final experiments we also add Bulgarian, Czech, Romanian, Swedish and Turkish.",
"We vary the amount and type of training data (types vs. tokens, UDT only, UM only, or UDT plus up to 10k UM examples), as described in Section 4.",
"To obtain N UM-based training examples, we select the first N unambiguous UM types (with their sentence contexts) from shuffled Wikipedia sentences.",
"For experiments with j > 1 examples per type, we first find all UM types with at least j sentence contexts in Wikipedia and then choose the N distinct types and their j contexts uniformly at random.",
"Evaluation To evaluate models' ability to lemmatize wordforms in their sentence context we follow Bergmanis and Goldwater (2018) and use the full UDT development and test sets.",
"Unlike Bergmanis and Goldwater (2018) who reported token level lemmatization exact match accuracy, we report type-level micro averaged lemmatization ex- act match accuracy.",
"This measure better reflects improvements on unseen words, which tend to be rare but are more important (since a most-frequentlemma baseline does very well on seen words, as shown by Bergmanis and Goldwater (2018) ).",
"We separately report performance on unseen and ambiguous tokens.",
"For a fair comparison across scenarios with different training sets, we count as unseen only words that are not ambiguous and are absent from all training sets/scenarios introduced in Section 4.",
"Due to the small training sets, between 70-90% of dev set types are classed as unseen in each language.",
"We define a type as ambiguous if the empirical entropy over its lemmas is greater than 0.1 in the full original UDT training splits.",
"10 According to this measure, only 1.2-5.3% of dev set types are classed as ambiguous in each language.",
"Significance Testing All systems are trained and tested on ten languages.",
"To test for statistically significant differences between the results of two systems we use a Monte Carlo method: for each set of results (i.e.",
"a set of 10 numerical values) we generate 10000 random samples, where each sample swaps the results of the two systems for each language with a probability of 0.5.",
"We then obtain a p-value as the proportion of samples for which the difference on average was at least as large as the difference observed in our experiments.",
"1k tokens vs. first 1k distinct types of the UDT training sets.",
"Table 2 shows that if only 1k examples are available, using types is clearly better for all systems.",
"Although Lematus does relatively poorly on the token data, it benefits the most from switching to types, putting it on par with HMAM and suggesting is it likely to benefit more from additional type data.",
"Lemming requires token-based data, but does worse than HMAM (a context-free method) in the token-based setting, and we also see no benefit from context in comparing Lematus 20-ch vs Lematus 0-ch.",
"So overall, in this very low-resource scenario with no data augmentation, context does not appear to help.",
"Using UM + Wikipedia Only We now try training only on UM + Wikipedia examples, rather than examples from UDT.",
"We use 1k, 2k or 5k unambiguous types from UM with a single example context from Wikipedia for each.",
"With 5k types we also try adding more example contexts (2, 3, or 5 examples for each type).",
"Figure 1 presents the results (for unseen words only).",
"As with the UDT experiments, there is little difference between Lematus 20-ch and Lematus 0ch in the smallest data setting.",
"However, when the number of training types increases to 5k, the benefits of context begin to show, with Lematus 20-ch yielding a 1.6% statistically significant (p < 0.001) improvement over Lematus 0-ch.",
"The results for increasing the number of examples per type are numerically higher than the one-example case, but the differences are not statistically significant.",
"It is worth noting that the accuracy even with 5k UM types is considerably lower than the accuracy of the model trained on only 1k UDT types (see Table 2 ).",
"We believe this discrepancy is due to the issues of biased/incomplete data noted above.",
"types with contexts from Wikipedia.",
"Table 3 summarizes the results, showing that despite the lower quality of the UM + Wikipedia examples, using them improves results of all systems, and more so with more examples.",
"Improvements are especially strong for unseen types, which constitute more than 70% of types in the dev set.",
"Furthermore, the benefit of the additional UM examples is above and beyond the effect of auto-encoding (AE Aug Baseline) for all systems in all data scenarios.",
"Considering the two context-free models, HMAM does better on the un-augmented 1k UDT data, but (as predicted by our results above) it benefits less from data augmentation than does Lematus 0-ch, so with added data they are statistically equivalent (p = 0.07 on the test set with 10k UM).",
"More importantly, Lematus 20-ch begins to outperform the context-free models with as few as 1k UM + Wikipedia examples, and the difference increases with more examples, eventually reaching over 4% better on the test set than the next best model (Lematus 0-ch) when 10k UM + Wikipedia examples are used (p < 0.001) This indicates that the system can learn useful contextual cues even from unambiguous training examples.",
"Finally, Figure 2 gives a breakdown of Lematus 20-ch dev set accuracy for individual languages, showing that data augmentation helps consistently, although results suggest diminishing returns.",
"Data Augmentation in Medium Resource Setting To examine the extent to which augmented data can help in the medium resource setting of 10k continuous tokens of UDT used in previous work, we follow Bergmanis and Goldwater (2018) and train Lematus 20-ch models for all ten languages using the first 10k tokens of UDT and compare them with models trained on 10k tokens of UDT augmented with 10k UM types.",
"To provide a better comparison of our results, we report both the type and the token level development set accuracy.",
"First of all, Table 4 shows that training on 10k continuous tokens of UDT yields a token level accuracy that is about 8% higher than when using the 1k types of UDT augmented with 10k UM types-the best-performing data augmentation systems (see Table 3 ).",
"Again, we believe this performance gap is due to the issues with the biased/incomplete data noted above.",
"For example, we analyzed errors that were unique to the model trained on the Latvian augmented data and found that 41% of the errors were due to wrongly lemmatized words other than nouns, verbs, and adjectives-the three POSs with available inflection tables in UM.",
"For instance, improperly lemmatized pronouns amounted to 14% of the errors on the Latvian dev set.",
"Table 4 also shows that UM examples with Wikipedia contexts benefit lemmatization not only in the low but also the medium resource setting, yielding statistically significant type and token level accuracy gains over models trained on 10k UDT continuous tokens alone (for both Unseen and All p < 0.001).",
"Conclusion We proposed a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive lemmatization model.",
"The proposed method uses Wikipedia sentences to provide contextualized examples for unambiguous inflection-lemma pairs from UniMorph tables.",
"These examples are noisy and biased, but nevertheless they improve lemmatization accuracy on all ten languages both in low (1k) and medium (10k) resource settings.",
"In particular, we showed that context is helpful, both overall and especially on unseen words-the first work we know of to demonstrate improvements from context in a very low-resource setting.",
"A Lematus Training Lematus is implemented using the Nematus machine translation toolkit 11 .",
"We use default training parameters of Lematus as specified by Bergmanis and Goldwater (2018) except for early stopping with patience (Prechelt, 1998) which we increase to 20.",
"Similar to Bergmanis and Goldwater (2018) we use the first epochs as a burn-in period, after which we validate the current model by its lemmatization exact match accuracy on the first 3k instances of development set and save this model if it performs better than the previous best model.",
"We choose a burn-in period of 20 and validation interval of 5 epochs for models that we train on datasets up to 2k instances and a burn-in period of 10 and validation interval of 2 epochs for others.",
"As we work with considerably smaller datasets than Bergmanis and Goldwater (2018) we reduce the effective model size and increase the rate of convergence by tying the input embeddings of the encoder, the decoder and the softmax output embeddings (Press and Wolf, 2017).",
"B Data Preparation Wikipedia database dumps contain XML structured articles that are formatted using the wikitext markup language.",
"To obtain wordforms in their sentence context we 1) use WikiExtractor 12 to extract plain text from Wikipedia database dumps, followed by scripts from Moses statistical machine translation system 13 (Koehn et al., 2007) to 2) split text into sentences (split-sentences.perl), and 3) extract separate tokens (tokenizer.perl).",
"Finally, we shuffle the extracted sentences to encourage homogeneous type distribution across the entire text.",
"Table 3 .",
"C Result Breakdown by Language"
]
} | {
"paper_header_number": [
"1",
"2.1",
"3",
"5"
],
"paper_header_content": [
"Introduction",
"Data Augmentation",
"Experimental Setup",
"Conclusion"
]
} | GEM-SciDuet-train-4#paper-965#slide-0 | Lemmatization | INST ar celu ar celiem
Latvian: cels (English: road) | INST ar celu ar celiem
Latvian: cels (English: road) | [] |
GEM-SciDuet-train-4#paper-965#slide-1 | 965 | Data Augmentation for Context-Sensitive Neural Lemmatization Using Inflection Tables and Raw Text | Lemmatization aims to reduce the sparse data problem by relating the inflected forms of a word to its dictionary form. Using context can help, both for unseen and ambiguous words. Yet most context-sensitive approaches require full lemma-annotated sentences for training, which may be scarce or unavailable in lowresource languages. In addition (as shown here), in a low-resource setting, a lemmatizer can learn more from n labeled examples of distinct words (types) than from n (contiguous) labeled tokens, since the latter contain far fewer distinct types. To combine the efficiency of type-based learning with the benefits of context, we propose a way to train a context-sensitive lemmatizer with little or no labeled corpus data, using inflection tables from the UniMorph project and raw text examples from Wikipedia that provide sentence contexts for the unambiguous UniMorph examples. Despite these being unambiguous examples, the model successfully generalizes from them, leading to improved results (both overall, and especially on unseen words) in comparison to a baseline that does not use context. Method Lematus (Bergmanis and Goldwater, 2018) is a neural sequence-to-sequence model with attention 2 Garrette et al. (2013) found the same for POS tagging. 3 Code and data: https://bitbucket.org/ tomsbergmanis/data_augumentation_um_wiki 4 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97
],
"paper_content_text": [
"Introduction Many lemmatizers work on isolated wordforms (Wicentowski, 2002; Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"Lemmatizing in context can improve accuracy on ambiguous and unseen words (Bergmanis and Goldwater, 2018) , but most systems for contextsensitive lemmatization must train on complete sentences labeled with POS and/or morphological tags as well as lemmas, and have only been tested with 20k-300k training tokens (Chrupała et al., 2008; Müller et al., 2015; Chakrabarty et al., 2017) .",
"1 1 The smallest of these corpora contains 20k tokens of Bengali annotated only with lemmas, which Chakrabarty et al.",
"(2017) reported took around two person months to create.",
"Intuitively, though, sentence-annotated data is inefficient for training a lemmatizer, especially in low-resource settings.",
"Training on (say) 1000 word types will provide far more information about a language's morphology than training on 1000 contiguous tokens, where fewer types are represented.",
"As noted above, sentence data can help with ambiguous and unseen words, but we show here that when data is scarce, this effect is small relative to the benefit of seeing more word types.",
"2 Motivated by this result, we propose a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive model.",
"3 We use Lematus (Bergmanis and Goldwater, 2018), a state-of-theart lemmatizer that learns from lemma-annotated words in their N -character contexts.",
"No predictions about surrounding words are used, so fully annotated training sentences are not needed.",
"We exploit this fact by combining two sources of training data: 1k lemma-annotated types (with contexts) from the Universal Dependency Treebank (UDT) v2.2 4 (Nivre et al., 2017) , plus examples obtained by finding unambiguous word-lemma pairs in inflection tables from the Universal Morphology (UM) project 5 and collecting sentence contexts for them from Wikipedia.",
"Although these examples are noisy and biased, we show that they improve lemmatization accuracy in experiments on 10 languages, and that the use of context helps, both overall and especially on unseen words.",
"inspired by the re-inflection model of Kann and Schütze (2016) , which won the 2016 SIGMOR-PHON shared task (Cotterell et al., 2016) .",
"It is built using the Nematus machine translation toolkit, 6 which uses the architecture of Sennrich et al.",
"(2017) : a 2-layer bidirectional GRU encoder and a 2-layer decoder with a conditional GRU (Sennrich et al., 2017) in the first layer and a GRU in the second layer.",
"Lematus takes as input a character sequence representing the wordform in its N -character context, and outputs the characters of the lemma.",
"Special input symbols are used to represent the left and right boundary of the target wordform (<lc>, <rc>) and other word boundaries (<s>).",
"For example, if N = 15, the system trained on Latvian would be expected to produce the characters of the lemma ceļš (meaning road) given input such as: s a k a <s> p aš v a l dī b u <lc> c e ļ u <rc> u n <s> i e l u <s> r eǵ i s t r When N = 0 (Lematus 0-ch), no context is used, making Lematus 0-ch comparable to other systems that do not model context (Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"In our experiments we use both Lematus 0-ch and Lematus 20-ch (20 characters of context), which was the best-performing system reported by Bergmanis and Goldwater (2018).",
"Data Augmentation Our data augmentation method uses UM inflection tables and creates additional training examples by finding Wikipedia sentences that use the inflected wordforms in context, pairing them with their lemma as shown in the inflection table.",
"However, we cannot use all the words in the tables because some of them are ambiguous: for example, Figure 1 shows that the form ceļi could be lemmatized either as ceļš or celis.",
"SG PL SG PL NOM ceļš ceļi celis ceļi GEN ceļa ceļu ceļa ceļu DAT ceļam ceļiem celim ceļiem ACC ceļu ceļus celi ceļus INS ceļu ceļiem celi ceļiem LOC ceļā ceļos celī ceļos VOC ceļ ceļi celi ceļi There are several other issues with this method that could potentially limit its usefulness.",
"First, the UM tables only include verbs, nouns and adjectives, whereas we test the system on UDT data, which includes all parts of speech.",
"Second, by excluding ambiguous forms, we may be restricting the added examples to a non-representative subset of the potential inflections, or the system may simply ignore the context because it isn't needed for these examples.",
"Finally, there are some annotation differences between UM and UDT.",
"7 Despite all of these issues, however, we show below that the added examples and their contexts do actually help.",
"Experimental Setup Baselines and Training Parameters We use four baselines: (1) Lemming 8 (Müller et al., 2015) is a context-sensitive system that uses log-linear models to jointly tag and lemmatize the data, and is trained on sentences annotated with both lemmas and POS tags.",
"(2) The hard monotonic attention model (HMAM) 9 (Makarov and Clematide, 2018b) is a neural sequence-tosequence model with a hard attention mechanism that advances through the sequence monotonically.",
"It is trained on word-lemma pairs (without context) 7 Recent efforts to unify the two resources have mostly focused on validating dataset schema (McCarthy et al., 2018) , leaving conflicts in word lemmas unresolved.",
"We estimated (by counting types that are unambiguous in each dataset but have different lemmas across them) that annotation inconsistencies affect up to 1% of types in the languages we used.",
"8 http://cistern.cis.lmu.de/lemming 9 https://github.com/ZurichNLP/ coling2018-neural-transition-basedmorphology with character-level alignments learned in a preprocessing step using an alignment model, and it has proved to be competitive in low resource scenarios.",
"(3) Our naive Baseline outputs the most frequent lemma (or one lemma at random from the options that are equally frequent) for words observed in training.",
"For unseen words it outputs the wordform itself.",
"(4) We also try a baseline data augmentation approach (AE Aug Baseline) inspired by Bergmanis et al.",
"(2017) and Kann and Schütze (2017) , who showed that adding training examples where the network simply learns to auto-encode corpus words can improve morphological inflection results in low-resource settings.",
"The AE Aug Baseline is a variant of Lematus 0-ch which augments the UDT lemmatization examples by auto-encoding the inflected forms of the UM examples (i.e., it just treats them as corpus words).",
"Comparing AE Aug Baseline to Lematus 0-ch augmented with UM lemma-inflection examples tells us whether using the UM lemma information helps more than simply auto-encoding more inflected examples.",
"To train the models we use the default settings for Lemming and the suggested lemmatization parameters for HMAM.",
"We mainly follow the hyperparameters used by Bergmanis and Goldwater (2018) for Lematus; details are in Appendix B.",
"Languages and Training Data We conduct preliminary experiments on five development languages: Estonian, Finnish, Latvian, Polish, and Russian.",
"In our final experiments we also add Bulgarian, Czech, Romanian, Swedish and Turkish.",
"We vary the amount and type of training data (types vs. tokens, UDT only, UM only, or UDT plus up to 10k UM examples), as described in Section 4.",
"To obtain N UM-based training examples, we select the first N unambiguous UM types (with their sentence contexts) from shuffled Wikipedia sentences.",
"For experiments with j > 1 examples per type, we first find all UM types with at least j sentence contexts in Wikipedia and then choose the N distinct types and their j contexts uniformly at random.",
"Evaluation To evaluate models' ability to lemmatize wordforms in their sentence context we follow Bergmanis and Goldwater (2018) and use the full UDT development and test sets.",
"Unlike Bergmanis and Goldwater (2018) who reported token level lemmatization exact match accuracy, we report type-level micro averaged lemmatization ex- act match accuracy.",
"This measure better reflects improvements on unseen words, which tend to be rare but are more important (since a most-frequentlemma baseline does very well on seen words, as shown by Bergmanis and Goldwater (2018) ).",
"We separately report performance on unseen and ambiguous tokens.",
"For a fair comparison across scenarios with different training sets, we count as unseen only words that are not ambiguous and are absent from all training sets/scenarios introduced in Section 4.",
"Due to the small training sets, between 70-90% of dev set types are classed as unseen in each language.",
"We define a type as ambiguous if the empirical entropy over its lemmas is greater than 0.1 in the full original UDT training splits.",
"10 According to this measure, only 1.2-5.3% of dev set types are classed as ambiguous in each language.",
"Significance Testing All systems are trained and tested on ten languages.",
"To test for statistically significant differences between the results of two systems we use a Monte Carlo method: for each set of results (i.e.",
"a set of 10 numerical values) we generate 10000 random samples, where each sample swaps the results of the two systems for each language with a probability of 0.5.",
"We then obtain a p-value as the proportion of samples for which the difference on average was at least as large as the difference observed in our experiments.",
"1k tokens vs. first 1k distinct types of the UDT training sets.",
"Table 2 shows that if only 1k examples are available, using types is clearly better for all systems.",
"Although Lematus does relatively poorly on the token data, it benefits the most from switching to types, putting it on par with HMAM and suggesting is it likely to benefit more from additional type data.",
"Lemming requires token-based data, but does worse than HMAM (a context-free method) in the token-based setting, and we also see no benefit from context in comparing Lematus 20-ch vs Lematus 0-ch.",
"So overall, in this very low-resource scenario with no data augmentation, context does not appear to help.",
"Using UM + Wikipedia Only We now try training only on UM + Wikipedia examples, rather than examples from UDT.",
"We use 1k, 2k or 5k unambiguous types from UM with a single example context from Wikipedia for each.",
"With 5k types we also try adding more example contexts (2, 3, or 5 examples for each type).",
"Figure 1 presents the results (for unseen words only).",
"As with the UDT experiments, there is little difference between Lematus 20-ch and Lematus 0ch in the smallest data setting.",
"However, when the number of training types increases to 5k, the benefits of context begin to show, with Lematus 20-ch yielding a 1.6% statistically significant (p < 0.001) improvement over Lematus 0-ch.",
"The results for increasing the number of examples per type are numerically higher than the one-example case, but the differences are not statistically significant.",
"It is worth noting that the accuracy even with 5k UM types is considerably lower than the accuracy of the model trained on only 1k UDT types (see Table 2 ).",
"We believe this discrepancy is due to the issues of biased/incomplete data noted above.",
"types with contexts from Wikipedia.",
"Table 3 summarizes the results, showing that despite the lower quality of the UM + Wikipedia examples, using them improves results of all systems, and more so with more examples.",
"Improvements are especially strong for unseen types, which constitute more than 70% of types in the dev set.",
"Furthermore, the benefit of the additional UM examples is above and beyond the effect of auto-encoding (AE Aug Baseline) for all systems in all data scenarios.",
"Considering the two context-free models, HMAM does better on the un-augmented 1k UDT data, but (as predicted by our results above) it benefits less from data augmentation than does Lematus 0-ch, so with added data they are statistically equivalent (p = 0.07 on the test set with 10k UM).",
"More importantly, Lematus 20-ch begins to outperform the context-free models with as few as 1k UM + Wikipedia examples, and the difference increases with more examples, eventually reaching over 4% better on the test set than the next best model (Lematus 0-ch) when 10k UM + Wikipedia examples are used (p < 0.001) This indicates that the system can learn useful contextual cues even from unambiguous training examples.",
"Finally, Figure 2 gives a breakdown of Lematus 20-ch dev set accuracy for individual languages, showing that data augmentation helps consistently, although results suggest diminishing returns.",
"Data Augmentation in Medium Resource Setting To examine the extent to which augmented data can help in the medium resource setting of 10k continuous tokens of UDT used in previous work, we follow Bergmanis and Goldwater (2018) and train Lematus 20-ch models for all ten languages using the first 10k tokens of UDT and compare them with models trained on 10k tokens of UDT augmented with 10k UM types.",
"To provide a better comparison of our results, we report both the type and the token level development set accuracy.",
"First of all, Table 4 shows that training on 10k continuous tokens of UDT yields a token level accuracy that is about 8% higher than when using the 1k types of UDT augmented with 10k UM types-the best-performing data augmentation systems (see Table 3 ).",
"Again, we believe this performance gap is due to the issues with the biased/incomplete data noted above.",
"For example, we analyzed errors that were unique to the model trained on the Latvian augmented data and found that 41% of the errors were due to wrongly lemmatized words other than nouns, verbs, and adjectives-the three POSs with available inflection tables in UM.",
"For instance, improperly lemmatized pronouns amounted to 14% of the errors on the Latvian dev set.",
"Table 4 also shows that UM examples with Wikipedia contexts benefit lemmatization not only in the low but also the medium resource setting, yielding statistically significant type and token level accuracy gains over models trained on 10k UDT continuous tokens alone (for both Unseen and All p < 0.001).",
"Conclusion We proposed a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive lemmatization model.",
"The proposed method uses Wikipedia sentences to provide contextualized examples for unambiguous inflection-lemma pairs from UniMorph tables.",
"These examples are noisy and biased, but nevertheless they improve lemmatization accuracy on all ten languages both in low (1k) and medium (10k) resource settings.",
"In particular, we showed that context is helpful, both overall and especially on unseen words-the first work we know of to demonstrate improvements from context in a very low-resource setting.",
"A Lematus Training Lematus is implemented using the Nematus machine translation toolkit 11 .",
"We use default training parameters of Lematus as specified by Bergmanis and Goldwater (2018) except for early stopping with patience (Prechelt, 1998) which we increase to 20.",
"Similar to Bergmanis and Goldwater (2018) we use the first epochs as a burn-in period, after which we validate the current model by its lemmatization exact match accuracy on the first 3k instances of development set and save this model if it performs better than the previous best model.",
"We choose a burn-in period of 20 and validation interval of 5 epochs for models that we train on datasets up to 2k instances and a burn-in period of 10 and validation interval of 2 epochs for others.",
"As we work with considerably smaller datasets than Bergmanis and Goldwater (2018) we reduce the effective model size and increase the rate of convergence by tying the input embeddings of the encoder, the decoder and the softmax output embeddings (Press and Wolf, 2017).",
"B Data Preparation Wikipedia database dumps contain XML structured articles that are formatted using the wikitext markup language.",
"To obtain wordforms in their sentence context we 1) use WikiExtractor 12 to extract plain text from Wikipedia database dumps, followed by scripts from Moses statistical machine translation system 13 (Koehn et al., 2007) to 2) split text into sentences (split-sentences.perl), and 3) extract separate tokens (tokenizer.perl).",
"Finally, we shuffle the extracted sentences to encourage homogeneous type distribution across the entire text.",
"Table 3 .",
"C Result Breakdown by Language"
]
} | {
"paper_header_number": [
"1",
"2.1",
"3",
"5"
],
"paper_header_content": [
"Introduction",
"Data Augmentation",
"Experimental Setup",
"Conclusion"
]
} | GEM-SciDuet-train-4#paper-965#slide-1 | Previous work | sentence context helps to lemmatize
ambiguous and unseen words
Bergmanis and Goldwater, 2018 | sentence context helps to lemmatize
ambiguous and unseen words
Bergmanis and Goldwater, 2018 | [] |
GEM-SciDuet-train-4#paper-965#slide-2 | 965 | Data Augmentation for Context-Sensitive Neural Lemmatization Using Inflection Tables and Raw Text | Lemmatization aims to reduce the sparse data problem by relating the inflected forms of a word to its dictionary form. Using context can help, both for unseen and ambiguous words. Yet most context-sensitive approaches require full lemma-annotated sentences for training, which may be scarce or unavailable in lowresource languages. In addition (as shown here), in a low-resource setting, a lemmatizer can learn more from n labeled examples of distinct words (types) than from n (contiguous) labeled tokens, since the latter contain far fewer distinct types. To combine the efficiency of type-based learning with the benefits of context, we propose a way to train a context-sensitive lemmatizer with little or no labeled corpus data, using inflection tables from the UniMorph project and raw text examples from Wikipedia that provide sentence contexts for the unambiguous UniMorph examples. Despite these being unambiguous examples, the model successfully generalizes from them, leading to improved results (both overall, and especially on unseen words) in comparison to a baseline that does not use context. Method Lematus (Bergmanis and Goldwater, 2018) is a neural sequence-to-sequence model with attention 2 Garrette et al. (2013) found the same for POS tagging. 3 Code and data: https://bitbucket.org/ tomsbergmanis/data_augumentation_um_wiki 4 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97
],
"paper_content_text": [
"Introduction Many lemmatizers work on isolated wordforms (Wicentowski, 2002; Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"Lemmatizing in context can improve accuracy on ambiguous and unseen words (Bergmanis and Goldwater, 2018) , but most systems for contextsensitive lemmatization must train on complete sentences labeled with POS and/or morphological tags as well as lemmas, and have only been tested with 20k-300k training tokens (Chrupała et al., 2008; Müller et al., 2015; Chakrabarty et al., 2017) .",
"1 1 The smallest of these corpora contains 20k tokens of Bengali annotated only with lemmas, which Chakrabarty et al.",
"(2017) reported took around two person months to create.",
"Intuitively, though, sentence-annotated data is inefficient for training a lemmatizer, especially in low-resource settings.",
"Training on (say) 1000 word types will provide far more information about a language's morphology than training on 1000 contiguous tokens, where fewer types are represented.",
"As noted above, sentence data can help with ambiguous and unseen words, but we show here that when data is scarce, this effect is small relative to the benefit of seeing more word types.",
"2 Motivated by this result, we propose a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive model.",
"3 We use Lematus (Bergmanis and Goldwater, 2018), a state-of-theart lemmatizer that learns from lemma-annotated words in their N -character contexts.",
"No predictions about surrounding words are used, so fully annotated training sentences are not needed.",
"We exploit this fact by combining two sources of training data: 1k lemma-annotated types (with contexts) from the Universal Dependency Treebank (UDT) v2.2 4 (Nivre et al., 2017) , plus examples obtained by finding unambiguous word-lemma pairs in inflection tables from the Universal Morphology (UM) project 5 and collecting sentence contexts for them from Wikipedia.",
"Although these examples are noisy and biased, we show that they improve lemmatization accuracy in experiments on 10 languages, and that the use of context helps, both overall and especially on unseen words.",
"inspired by the re-inflection model of Kann and Schütze (2016) , which won the 2016 SIGMOR-PHON shared task (Cotterell et al., 2016) .",
"It is built using the Nematus machine translation toolkit, 6 which uses the architecture of Sennrich et al.",
"(2017) : a 2-layer bidirectional GRU encoder and a 2-layer decoder with a conditional GRU (Sennrich et al., 2017) in the first layer and a GRU in the second layer.",
"Lematus takes as input a character sequence representing the wordform in its N -character context, and outputs the characters of the lemma.",
"Special input symbols are used to represent the left and right boundary of the target wordform (<lc>, <rc>) and other word boundaries (<s>).",
"For example, if N = 15, the system trained on Latvian would be expected to produce the characters of the lemma ceļš (meaning road) given input such as: s a k a <s> p aš v a l dī b u <lc> c e ļ u <rc> u n <s> i e l u <s> r eǵ i s t r When N = 0 (Lematus 0-ch), no context is used, making Lematus 0-ch comparable to other systems that do not model context (Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"In our experiments we use both Lematus 0-ch and Lematus 20-ch (20 characters of context), which was the best-performing system reported by Bergmanis and Goldwater (2018).",
"Data Augmentation Our data augmentation method uses UM inflection tables and creates additional training examples by finding Wikipedia sentences that use the inflected wordforms in context, pairing them with their lemma as shown in the inflection table.",
"However, we cannot use all the words in the tables because some of them are ambiguous: for example, Figure 1 shows that the form ceļi could be lemmatized either as ceļš or celis.",
"SG PL SG PL NOM ceļš ceļi celis ceļi GEN ceļa ceļu ceļa ceļu DAT ceļam ceļiem celim ceļiem ACC ceļu ceļus celi ceļus INS ceļu ceļiem celi ceļiem LOC ceļā ceļos celī ceļos VOC ceļ ceļi celi ceļi There are several other issues with this method that could potentially limit its usefulness.",
"First, the UM tables only include verbs, nouns and adjectives, whereas we test the system on UDT data, which includes all parts of speech.",
"Second, by excluding ambiguous forms, we may be restricting the added examples to a non-representative subset of the potential inflections, or the system may simply ignore the context because it isn't needed for these examples.",
"Finally, there are some annotation differences between UM and UDT.",
"7 Despite all of these issues, however, we show below that the added examples and their contexts do actually help.",
"Experimental Setup Baselines and Training Parameters We use four baselines: (1) Lemming 8 (Müller et al., 2015) is a context-sensitive system that uses log-linear models to jointly tag and lemmatize the data, and is trained on sentences annotated with both lemmas and POS tags.",
"(2) The hard monotonic attention model (HMAM) 9 (Makarov and Clematide, 2018b) is a neural sequence-tosequence model with a hard attention mechanism that advances through the sequence monotonically.",
"It is trained on word-lemma pairs (without context) 7 Recent efforts to unify the two resources have mostly focused on validating dataset schema (McCarthy et al., 2018) , leaving conflicts in word lemmas unresolved.",
"We estimated (by counting types that are unambiguous in each dataset but have different lemmas across them) that annotation inconsistencies affect up to 1% of types in the languages we used.",
"8 http://cistern.cis.lmu.de/lemming 9 https://github.com/ZurichNLP/ coling2018-neural-transition-basedmorphology with character-level alignments learned in a preprocessing step using an alignment model, and it has proved to be competitive in low resource scenarios.",
"(3) Our naive Baseline outputs the most frequent lemma (or one lemma at random from the options that are equally frequent) for words observed in training.",
"For unseen words it outputs the wordform itself.",
"(4) We also try a baseline data augmentation approach (AE Aug Baseline) inspired by Bergmanis et al.",
"(2017) and Kann and Schütze (2017) , who showed that adding training examples where the network simply learns to auto-encode corpus words can improve morphological inflection results in low-resource settings.",
"The AE Aug Baseline is a variant of Lematus 0-ch which augments the UDT lemmatization examples by auto-encoding the inflected forms of the UM examples (i.e., it just treats them as corpus words).",
"Comparing AE Aug Baseline to Lematus 0-ch augmented with UM lemma-inflection examples tells us whether using the UM lemma information helps more than simply auto-encoding more inflected examples.",
"To train the models we use the default settings for Lemming and the suggested lemmatization parameters for HMAM.",
"We mainly follow the hyperparameters used by Bergmanis and Goldwater (2018) for Lematus; details are in Appendix B.",
"Languages and Training Data We conduct preliminary experiments on five development languages: Estonian, Finnish, Latvian, Polish, and Russian.",
"In our final experiments we also add Bulgarian, Czech, Romanian, Swedish and Turkish.",
"We vary the amount and type of training data (types vs. tokens, UDT only, UM only, or UDT plus up to 10k UM examples), as described in Section 4.",
"To obtain N UM-based training examples, we select the first N unambiguous UM types (with their sentence contexts) from shuffled Wikipedia sentences.",
"For experiments with j > 1 examples per type, we first find all UM types with at least j sentence contexts in Wikipedia and then choose the N distinct types and their j contexts uniformly at random.",
"Evaluation To evaluate models' ability to lemmatize wordforms in their sentence context we follow Bergmanis and Goldwater (2018) and use the full UDT development and test sets.",
"Unlike Bergmanis and Goldwater (2018) who reported token level lemmatization exact match accuracy, we report type-level micro averaged lemmatization ex- act match accuracy.",
"This measure better reflects improvements on unseen words, which tend to be rare but are more important (since a most-frequentlemma baseline does very well on seen words, as shown by Bergmanis and Goldwater (2018) ).",
"We separately report performance on unseen and ambiguous tokens.",
"For a fair comparison across scenarios with different training sets, we count as unseen only words that are not ambiguous and are absent from all training sets/scenarios introduced in Section 4.",
"Due to the small training sets, between 70-90% of dev set types are classed as unseen in each language.",
"We define a type as ambiguous if the empirical entropy over its lemmas is greater than 0.1 in the full original UDT training splits.",
"10 According to this measure, only 1.2-5.3% of dev set types are classed as ambiguous in each language.",
"Significance Testing All systems are trained and tested on ten languages.",
"To test for statistically significant differences between the results of two systems we use a Monte Carlo method: for each set of results (i.e.",
"a set of 10 numerical values) we generate 10000 random samples, where each sample swaps the results of the two systems for each language with a probability of 0.5.",
"We then obtain a p-value as the proportion of samples for which the difference on average was at least as large as the difference observed in our experiments.",
"1k tokens vs. first 1k distinct types of the UDT training sets.",
"Table 2 shows that if only 1k examples are available, using types is clearly better for all systems.",
"Although Lematus does relatively poorly on the token data, it benefits the most from switching to types, putting it on par with HMAM and suggesting is it likely to benefit more from additional type data.",
"Lemming requires token-based data, but does worse than HMAM (a context-free method) in the token-based setting, and we also see no benefit from context in comparing Lematus 20-ch vs Lematus 0-ch.",
"So overall, in this very low-resource scenario with no data augmentation, context does not appear to help.",
"Using UM + Wikipedia Only We now try training only on UM + Wikipedia examples, rather than examples from UDT.",
"We use 1k, 2k or 5k unambiguous types from UM with a single example context from Wikipedia for each.",
"With 5k types we also try adding more example contexts (2, 3, or 5 examples for each type).",
"Figure 1 presents the results (for unseen words only).",
"As with the UDT experiments, there is little difference between Lematus 20-ch and Lematus 0ch in the smallest data setting.",
"However, when the number of training types increases to 5k, the benefits of context begin to show, with Lematus 20-ch yielding a 1.6% statistically significant (p < 0.001) improvement over Lematus 0-ch.",
"The results for increasing the number of examples per type are numerically higher than the one-example case, but the differences are not statistically significant.",
"It is worth noting that the accuracy even with 5k UM types is considerably lower than the accuracy of the model trained on only 1k UDT types (see Table 2 ).",
"We believe this discrepancy is due to the issues of biased/incomplete data noted above.",
"types with contexts from Wikipedia.",
"Table 3 summarizes the results, showing that despite the lower quality of the UM + Wikipedia examples, using them improves results of all systems, and more so with more examples.",
"Improvements are especially strong for unseen types, which constitute more than 70% of types in the dev set.",
"Furthermore, the benefit of the additional UM examples is above and beyond the effect of auto-encoding (AE Aug Baseline) for all systems in all data scenarios.",
"Considering the two context-free models, HMAM does better on the un-augmented 1k UDT data, but (as predicted by our results above) it benefits less from data augmentation than does Lematus 0-ch, so with added data they are statistically equivalent (p = 0.07 on the test set with 10k UM).",
"More importantly, Lematus 20-ch begins to outperform the context-free models with as few as 1k UM + Wikipedia examples, and the difference increases with more examples, eventually reaching over 4% better on the test set than the next best model (Lematus 0-ch) when 10k UM + Wikipedia examples are used (p < 0.001) This indicates that the system can learn useful contextual cues even from unambiguous training examples.",
"Finally, Figure 2 gives a breakdown of Lematus 20-ch dev set accuracy for individual languages, showing that data augmentation helps consistently, although results suggest diminishing returns.",
"Data Augmentation in Medium Resource Setting To examine the extent to which augmented data can help in the medium resource setting of 10k continuous tokens of UDT used in previous work, we follow Bergmanis and Goldwater (2018) and train Lematus 20-ch models for all ten languages using the first 10k tokens of UDT and compare them with models trained on 10k tokens of UDT augmented with 10k UM types.",
"To provide a better comparison of our results, we report both the type and the token level development set accuracy.",
"First of all, Table 4 shows that training on 10k continuous tokens of UDT yields a token level accuracy that is about 8% higher than when using the 1k types of UDT augmented with 10k UM types-the best-performing data augmentation systems (see Table 3 ).",
"Again, we believe this performance gap is due to the issues with the biased/incomplete data noted above.",
"For example, we analyzed errors that were unique to the model trained on the Latvian augmented data and found that 41% of the errors were due to wrongly lemmatized words other than nouns, verbs, and adjectives-the three POSs with available inflection tables in UM.",
"For instance, improperly lemmatized pronouns amounted to 14% of the errors on the Latvian dev set.",
"Table 4 also shows that UM examples with Wikipedia contexts benefit lemmatization not only in the low but also the medium resource setting, yielding statistically significant type and token level accuracy gains over models trained on 10k UDT continuous tokens alone (for both Unseen and All p < 0.001).",
"Conclusion We proposed a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive lemmatization model.",
"The proposed method uses Wikipedia sentences to provide contextualized examples for unambiguous inflection-lemma pairs from UniMorph tables.",
"These examples are noisy and biased, but nevertheless they improve lemmatization accuracy on all ten languages both in low (1k) and medium (10k) resource settings.",
"In particular, we showed that context is helpful, both overall and especially on unseen words-the first work we know of to demonstrate improvements from context in a very low-resource setting.",
"A Lematus Training Lematus is implemented using the Nematus machine translation toolkit 11 .",
"We use default training parameters of Lematus as specified by Bergmanis and Goldwater (2018) except for early stopping with patience (Prechelt, 1998) which we increase to 20.",
"Similar to Bergmanis and Goldwater (2018) we use the first epochs as a burn-in period, after which we validate the current model by its lemmatization exact match accuracy on the first 3k instances of development set and save this model if it performs better than the previous best model.",
"We choose a burn-in period of 20 and validation interval of 5 epochs for models that we train on datasets up to 2k instances and a burn-in period of 10 and validation interval of 2 epochs for others.",
"As we work with considerably smaller datasets than Bergmanis and Goldwater (2018) we reduce the effective model size and increase the rate of convergence by tying the input embeddings of the encoder, the decoder and the softmax output embeddings (Press and Wolf, 2017).",
"B Data Preparation Wikipedia database dumps contain XML structured articles that are formatted using the wikitext markup language.",
"To obtain wordforms in their sentence context we 1) use WikiExtractor 12 to extract plain text from Wikipedia database dumps, followed by scripts from Moses statistical machine translation system 13 (Koehn et al., 2007) to 2) split text into sentences (split-sentences.perl), and 3) extract separate tokens (tokenizer.perl).",
"Finally, we shuffle the extracted sentences to encourage homogeneous type distribution across the entire text.",
"Table 3 .",
"C Result Breakdown by Language"
]
} | {
"paper_header_number": [
"1",
"2.1",
"3",
"5"
],
"paper_header_content": [
"Introduction",
"Data Augmentation",
"Experimental Setup",
"Conclusion"
]
} | GEM-SciDuet-train-4#paper-965#slide-2 | Ambiguous words | A cels (road): NOUN, sing., ACC
B celis (knee): NOUN, plur., DAT | A cels (road): NOUN, sing., ACC
B celis (knee): NOUN, plur., DAT | [] |
GEM-SciDuet-train-4#paper-965#slide-3 | 965 | Data Augmentation for Context-Sensitive Neural Lemmatization Using Inflection Tables and Raw Text | Lemmatization aims to reduce the sparse data problem by relating the inflected forms of a word to its dictionary form. Using context can help, both for unseen and ambiguous words. Yet most context-sensitive approaches require full lemma-annotated sentences for training, which may be scarce or unavailable in lowresource languages. In addition (as shown here), in a low-resource setting, a lemmatizer can learn more from n labeled examples of distinct words (types) than from n (contiguous) labeled tokens, since the latter contain far fewer distinct types. To combine the efficiency of type-based learning with the benefits of context, we propose a way to train a context-sensitive lemmatizer with little or no labeled corpus data, using inflection tables from the UniMorph project and raw text examples from Wikipedia that provide sentence contexts for the unambiguous UniMorph examples. Despite these being unambiguous examples, the model successfully generalizes from them, leading to improved results (both overall, and especially on unseen words) in comparison to a baseline that does not use context. Method Lematus (Bergmanis and Goldwater, 2018) is a neural sequence-to-sequence model with attention 2 Garrette et al. (2013) found the same for POS tagging. 3 Code and data: https://bitbucket.org/ tomsbergmanis/data_augumentation_um_wiki 4 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97
],
"paper_content_text": [
"Introduction Many lemmatizers work on isolated wordforms (Wicentowski, 2002; Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"Lemmatizing in context can improve accuracy on ambiguous and unseen words (Bergmanis and Goldwater, 2018) , but most systems for contextsensitive lemmatization must train on complete sentences labeled with POS and/or morphological tags as well as lemmas, and have only been tested with 20k-300k training tokens (Chrupała et al., 2008; Müller et al., 2015; Chakrabarty et al., 2017) .",
"1 1 The smallest of these corpora contains 20k tokens of Bengali annotated only with lemmas, which Chakrabarty et al.",
"(2017) reported took around two person months to create.",
"Intuitively, though, sentence-annotated data is inefficient for training a lemmatizer, especially in low-resource settings.",
"Training on (say) 1000 word types will provide far more information about a language's morphology than training on 1000 contiguous tokens, where fewer types are represented.",
"As noted above, sentence data can help with ambiguous and unseen words, but we show here that when data is scarce, this effect is small relative to the benefit of seeing more word types.",
"2 Motivated by this result, we propose a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive model.",
"3 We use Lematus (Bergmanis and Goldwater, 2018), a state-of-theart lemmatizer that learns from lemma-annotated words in their N -character contexts.",
"No predictions about surrounding words are used, so fully annotated training sentences are not needed.",
"We exploit this fact by combining two sources of training data: 1k lemma-annotated types (with contexts) from the Universal Dependency Treebank (UDT) v2.2 4 (Nivre et al., 2017) , plus examples obtained by finding unambiguous word-lemma pairs in inflection tables from the Universal Morphology (UM) project 5 and collecting sentence contexts for them from Wikipedia.",
"Although these examples are noisy and biased, we show that they improve lemmatization accuracy in experiments on 10 languages, and that the use of context helps, both overall and especially on unseen words.",
"inspired by the re-inflection model of Kann and Schütze (2016) , which won the 2016 SIGMOR-PHON shared task (Cotterell et al., 2016) .",
"It is built using the Nematus machine translation toolkit, 6 which uses the architecture of Sennrich et al.",
"(2017) : a 2-layer bidirectional GRU encoder and a 2-layer decoder with a conditional GRU (Sennrich et al., 2017) in the first layer and a GRU in the second layer.",
"Lematus takes as input a character sequence representing the wordform in its N -character context, and outputs the characters of the lemma.",
"Special input symbols are used to represent the left and right boundary of the target wordform (<lc>, <rc>) and other word boundaries (<s>).",
"For example, if N = 15, the system trained on Latvian would be expected to produce the characters of the lemma ceļš (meaning road) given input such as: s a k a <s> p aš v a l dī b u <lc> c e ļ u <rc> u n <s> i e l u <s> r eǵ i s t r When N = 0 (Lematus 0-ch), no context is used, making Lematus 0-ch comparable to other systems that do not model context (Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"In our experiments we use both Lematus 0-ch and Lematus 20-ch (20 characters of context), which was the best-performing system reported by Bergmanis and Goldwater (2018).",
"Data Augmentation Our data augmentation method uses UM inflection tables and creates additional training examples by finding Wikipedia sentences that use the inflected wordforms in context, pairing them with their lemma as shown in the inflection table.",
"However, we cannot use all the words in the tables because some of them are ambiguous: for example, Figure 1 shows that the form ceļi could be lemmatized either as ceļš or celis.",
"SG PL SG PL NOM ceļš ceļi celis ceļi GEN ceļa ceļu ceļa ceļu DAT ceļam ceļiem celim ceļiem ACC ceļu ceļus celi ceļus INS ceļu ceļiem celi ceļiem LOC ceļā ceļos celī ceļos VOC ceļ ceļi celi ceļi There are several other issues with this method that could potentially limit its usefulness.",
"First, the UM tables only include verbs, nouns and adjectives, whereas we test the system on UDT data, which includes all parts of speech.",
"Second, by excluding ambiguous forms, we may be restricting the added examples to a non-representative subset of the potential inflections, or the system may simply ignore the context because it isn't needed for these examples.",
"Finally, there are some annotation differences between UM and UDT.",
"7 Despite all of these issues, however, we show below that the added examples and their contexts do actually help.",
"Experimental Setup Baselines and Training Parameters We use four baselines: (1) Lemming 8 (Müller et al., 2015) is a context-sensitive system that uses log-linear models to jointly tag and lemmatize the data, and is trained on sentences annotated with both lemmas and POS tags.",
"(2) The hard monotonic attention model (HMAM) 9 (Makarov and Clematide, 2018b) is a neural sequence-tosequence model with a hard attention mechanism that advances through the sequence monotonically.",
"It is trained on word-lemma pairs (without context) 7 Recent efforts to unify the two resources have mostly focused on validating dataset schema (McCarthy et al., 2018) , leaving conflicts in word lemmas unresolved.",
"We estimated (by counting types that are unambiguous in each dataset but have different lemmas across them) that annotation inconsistencies affect up to 1% of types in the languages we used.",
"8 http://cistern.cis.lmu.de/lemming 9 https://github.com/ZurichNLP/ coling2018-neural-transition-basedmorphology with character-level alignments learned in a preprocessing step using an alignment model, and it has proved to be competitive in low resource scenarios.",
"(3) Our naive Baseline outputs the most frequent lemma (or one lemma at random from the options that are equally frequent) for words observed in training.",
"For unseen words it outputs the wordform itself.",
"(4) We also try a baseline data augmentation approach (AE Aug Baseline) inspired by Bergmanis et al.",
"(2017) and Kann and Schütze (2017) , who showed that adding training examples where the network simply learns to auto-encode corpus words can improve morphological inflection results in low-resource settings.",
"The AE Aug Baseline is a variant of Lematus 0-ch which augments the UDT lemmatization examples by auto-encoding the inflected forms of the UM examples (i.e., it just treats them as corpus words).",
"Comparing AE Aug Baseline to Lematus 0-ch augmented with UM lemma-inflection examples tells us whether using the UM lemma information helps more than simply auto-encoding more inflected examples.",
"To train the models we use the default settings for Lemming and the suggested lemmatization parameters for HMAM.",
"We mainly follow the hyperparameters used by Bergmanis and Goldwater (2018) for Lematus; details are in Appendix B.",
"Languages and Training Data We conduct preliminary experiments on five development languages: Estonian, Finnish, Latvian, Polish, and Russian.",
"In our final experiments we also add Bulgarian, Czech, Romanian, Swedish and Turkish.",
"We vary the amount and type of training data (types vs. tokens, UDT only, UM only, or UDT plus up to 10k UM examples), as described in Section 4.",
"To obtain N UM-based training examples, we select the first N unambiguous UM types (with their sentence contexts) from shuffled Wikipedia sentences.",
"For experiments with j > 1 examples per type, we first find all UM types with at least j sentence contexts in Wikipedia and then choose the N distinct types and their j contexts uniformly at random.",
"Evaluation To evaluate models' ability to lemmatize wordforms in their sentence context we follow Bergmanis and Goldwater (2018) and use the full UDT development and test sets.",
"Unlike Bergmanis and Goldwater (2018) who reported token level lemmatization exact match accuracy, we report type-level micro averaged lemmatization ex- act match accuracy.",
"This measure better reflects improvements on unseen words, which tend to be rare but are more important (since a most-frequentlemma baseline does very well on seen words, as shown by Bergmanis and Goldwater (2018) ).",
"We separately report performance on unseen and ambiguous tokens.",
"For a fair comparison across scenarios with different training sets, we count as unseen only words that are not ambiguous and are absent from all training sets/scenarios introduced in Section 4.",
"Due to the small training sets, between 70-90% of dev set types are classed as unseen in each language.",
"We define a type as ambiguous if the empirical entropy over its lemmas is greater than 0.1 in the full original UDT training splits.",
"10 According to this measure, only 1.2-5.3% of dev set types are classed as ambiguous in each language.",
"Significance Testing All systems are trained and tested on ten languages.",
"To test for statistically significant differences between the results of two systems we use a Monte Carlo method: for each set of results (i.e.",
"a set of 10 numerical values) we generate 10000 random samples, where each sample swaps the results of the two systems for each language with a probability of 0.5.",
"We then obtain a p-value as the proportion of samples for which the difference on average was at least as large as the difference observed in our experiments.",
"1k tokens vs. first 1k distinct types of the UDT training sets.",
"Table 2 shows that if only 1k examples are available, using types is clearly better for all systems.",
"Although Lematus does relatively poorly on the token data, it benefits the most from switching to types, putting it on par with HMAM and suggesting is it likely to benefit more from additional type data.",
"Lemming requires token-based data, but does worse than HMAM (a context-free method) in the token-based setting, and we also see no benefit from context in comparing Lematus 20-ch vs Lematus 0-ch.",
"So overall, in this very low-resource scenario with no data augmentation, context does not appear to help.",
"Using UM + Wikipedia Only We now try training only on UM + Wikipedia examples, rather than examples from UDT.",
"We use 1k, 2k or 5k unambiguous types from UM with a single example context from Wikipedia for each.",
"With 5k types we also try adding more example contexts (2, 3, or 5 examples for each type).",
"Figure 1 presents the results (for unseen words only).",
"As with the UDT experiments, there is little difference between Lematus 20-ch and Lematus 0ch in the smallest data setting.",
"However, when the number of training types increases to 5k, the benefits of context begin to show, with Lematus 20-ch yielding a 1.6% statistically significant (p < 0.001) improvement over Lematus 0-ch.",
"The results for increasing the number of examples per type are numerically higher than the one-example case, but the differences are not statistically significant.",
"It is worth noting that the accuracy even with 5k UM types is considerably lower than the accuracy of the model trained on only 1k UDT types (see Table 2 ).",
"We believe this discrepancy is due to the issues of biased/incomplete data noted above.",
"types with contexts from Wikipedia.",
"Table 3 summarizes the results, showing that despite the lower quality of the UM + Wikipedia examples, using them improves results of all systems, and more so with more examples.",
"Improvements are especially strong for unseen types, which constitute more than 70% of types in the dev set.",
"Furthermore, the benefit of the additional UM examples is above and beyond the effect of auto-encoding (AE Aug Baseline) for all systems in all data scenarios.",
"Considering the two context-free models, HMAM does better on the un-augmented 1k UDT data, but (as predicted by our results above) it benefits less from data augmentation than does Lematus 0-ch, so with added data they are statistically equivalent (p = 0.07 on the test set with 10k UM).",
"More importantly, Lematus 20-ch begins to outperform the context-free models with as few as 1k UM + Wikipedia examples, and the difference increases with more examples, eventually reaching over 4% better on the test set than the next best model (Lematus 0-ch) when 10k UM + Wikipedia examples are used (p < 0.001) This indicates that the system can learn useful contextual cues even from unambiguous training examples.",
"Finally, Figure 2 gives a breakdown of Lematus 20-ch dev set accuracy for individual languages, showing that data augmentation helps consistently, although results suggest diminishing returns.",
"Data Augmentation in Medium Resource Setting To examine the extent to which augmented data can help in the medium resource setting of 10k continuous tokens of UDT used in previous work, we follow Bergmanis and Goldwater (2018) and train Lematus 20-ch models for all ten languages using the first 10k tokens of UDT and compare them with models trained on 10k tokens of UDT augmented with 10k UM types.",
"To provide a better comparison of our results, we report both the type and the token level development set accuracy.",
"First of all, Table 4 shows that training on 10k continuous tokens of UDT yields a token level accuracy that is about 8% higher than when using the 1k types of UDT augmented with 10k UM types-the best-performing data augmentation systems (see Table 3 ).",
"Again, we believe this performance gap is due to the issues with the biased/incomplete data noted above.",
"For example, we analyzed errors that were unique to the model trained on the Latvian augmented data and found that 41% of the errors were due to wrongly lemmatized words other than nouns, verbs, and adjectives-the three POSs with available inflection tables in UM.",
"For instance, improperly lemmatized pronouns amounted to 14% of the errors on the Latvian dev set.",
"Table 4 also shows that UM examples with Wikipedia contexts benefit lemmatization not only in the low but also the medium resource setting, yielding statistically significant type and token level accuracy gains over models trained on 10k UDT continuous tokens alone (for both Unseen and All p < 0.001).",
"Conclusion We proposed a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive lemmatization model.",
"The proposed method uses Wikipedia sentences to provide contextualized examples for unambiguous inflection-lemma pairs from UniMorph tables.",
"These examples are noisy and biased, but nevertheless they improve lemmatization accuracy on all ten languages both in low (1k) and medium (10k) resource settings.",
"In particular, we showed that context is helpful, both overall and especially on unseen words-the first work we know of to demonstrate improvements from context in a very low-resource setting.",
"A Lematus Training Lematus is implemented using the Nematus machine translation toolkit 11 .",
"We use default training parameters of Lematus as specified by Bergmanis and Goldwater (2018) except for early stopping with patience (Prechelt, 1998) which we increase to 20.",
"Similar to Bergmanis and Goldwater (2018) we use the first epochs as a burn-in period, after which we validate the current model by its lemmatization exact match accuracy on the first 3k instances of development set and save this model if it performs better than the previous best model.",
"We choose a burn-in period of 20 and validation interval of 5 epochs for models that we train on datasets up to 2k instances and a burn-in period of 10 and validation interval of 2 epochs for others.",
"As we work with considerably smaller datasets than Bergmanis and Goldwater (2018) we reduce the effective model size and increase the rate of convergence by tying the input embeddings of the encoder, the decoder and the softmax output embeddings (Press and Wolf, 2017).",
"B Data Preparation Wikipedia database dumps contain XML structured articles that are formatted using the wikitext markup language.",
"To obtain wordforms in their sentence context we 1) use WikiExtractor 12 to extract plain text from Wikipedia database dumps, followed by scripts from Moses statistical machine translation system 13 (Koehn et al., 2007) to 2) split text into sentences (split-sentences.perl), and 3) extract separate tokens (tokenizer.perl).",
"Finally, we shuffle the extracted sentences to encourage homogeneous type distribution across the entire text.",
"Table 3 .",
"C Result Breakdown by Language"
]
} | {
"paper_header_number": [
"1",
"2.1",
"3",
"5"
],
"paper_header_content": [
"Introduction",
"Data Augmentation",
"Experimental Setup",
"Conclusion"
]
} | GEM-SciDuet-train-4#paper-965#slide-3 | Learning from sentences | Lemma annotated sentences are scarce for low resource languages annotating sentences is slow
N types > N (contiguous) tokens | Lemma annotated sentences are scarce for low resource languages annotating sentences is slow
N types > N (contiguous) tokens | [] |
GEM-SciDuet-train-4#paper-965#slide-4 | 965 | Data Augmentation for Context-Sensitive Neural Lemmatization Using Inflection Tables and Raw Text | Lemmatization aims to reduce the sparse data problem by relating the inflected forms of a word to its dictionary form. Using context can help, both for unseen and ambiguous words. Yet most context-sensitive approaches require full lemma-annotated sentences for training, which may be scarce or unavailable in lowresource languages. In addition (as shown here), in a low-resource setting, a lemmatizer can learn more from n labeled examples of distinct words (types) than from n (contiguous) labeled tokens, since the latter contain far fewer distinct types. To combine the efficiency of type-based learning with the benefits of context, we propose a way to train a context-sensitive lemmatizer with little or no labeled corpus data, using inflection tables from the UniMorph project and raw text examples from Wikipedia that provide sentence contexts for the unambiguous UniMorph examples. Despite these being unambiguous examples, the model successfully generalizes from them, leading to improved results (both overall, and especially on unseen words) in comparison to a baseline that does not use context. Method Lematus (Bergmanis and Goldwater, 2018) is a neural sequence-to-sequence model with attention 2 Garrette et al. (2013) found the same for POS tagging. 3 Code and data: https://bitbucket.org/ tomsbergmanis/data_augumentation_um_wiki 4 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97
],
"paper_content_text": [
"Introduction Many lemmatizers work on isolated wordforms (Wicentowski, 2002; Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"Lemmatizing in context can improve accuracy on ambiguous and unseen words (Bergmanis and Goldwater, 2018) , but most systems for contextsensitive lemmatization must train on complete sentences labeled with POS and/or morphological tags as well as lemmas, and have only been tested with 20k-300k training tokens (Chrupała et al., 2008; Müller et al., 2015; Chakrabarty et al., 2017) .",
"1 1 The smallest of these corpora contains 20k tokens of Bengali annotated only with lemmas, which Chakrabarty et al.",
"(2017) reported took around two person months to create.",
"Intuitively, though, sentence-annotated data is inefficient for training a lemmatizer, especially in low-resource settings.",
"Training on (say) 1000 word types will provide far more information about a language's morphology than training on 1000 contiguous tokens, where fewer types are represented.",
"As noted above, sentence data can help with ambiguous and unseen words, but we show here that when data is scarce, this effect is small relative to the benefit of seeing more word types.",
"2 Motivated by this result, we propose a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive model.",
"3 We use Lematus (Bergmanis and Goldwater, 2018), a state-of-theart lemmatizer that learns from lemma-annotated words in their N -character contexts.",
"No predictions about surrounding words are used, so fully annotated training sentences are not needed.",
"We exploit this fact by combining two sources of training data: 1k lemma-annotated types (with contexts) from the Universal Dependency Treebank (UDT) v2.2 4 (Nivre et al., 2017) , plus examples obtained by finding unambiguous word-lemma pairs in inflection tables from the Universal Morphology (UM) project 5 and collecting sentence contexts for them from Wikipedia.",
"Although these examples are noisy and biased, we show that they improve lemmatization accuracy in experiments on 10 languages, and that the use of context helps, both overall and especially on unseen words.",
"inspired by the re-inflection model of Kann and Schütze (2016) , which won the 2016 SIGMOR-PHON shared task (Cotterell et al., 2016) .",
"It is built using the Nematus machine translation toolkit, 6 which uses the architecture of Sennrich et al.",
"(2017) : a 2-layer bidirectional GRU encoder and a 2-layer decoder with a conditional GRU (Sennrich et al., 2017) in the first layer and a GRU in the second layer.",
"Lematus takes as input a character sequence representing the wordform in its N -character context, and outputs the characters of the lemma.",
"Special input symbols are used to represent the left and right boundary of the target wordform (<lc>, <rc>) and other word boundaries (<s>).",
"For example, if N = 15, the system trained on Latvian would be expected to produce the characters of the lemma ceļš (meaning road) given input such as: s a k a <s> p aš v a l dī b u <lc> c e ļ u <rc> u n <s> i e l u <s> r eǵ i s t r When N = 0 (Lematus 0-ch), no context is used, making Lematus 0-ch comparable to other systems that do not model context (Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"In our experiments we use both Lematus 0-ch and Lematus 20-ch (20 characters of context), which was the best-performing system reported by Bergmanis and Goldwater (2018).",
"Data Augmentation Our data augmentation method uses UM inflection tables and creates additional training examples by finding Wikipedia sentences that use the inflected wordforms in context, pairing them with their lemma as shown in the inflection table.",
"However, we cannot use all the words in the tables because some of them are ambiguous: for example, Figure 1 shows that the form ceļi could be lemmatized either as ceļš or celis.",
"SG PL SG PL NOM ceļš ceļi celis ceļi GEN ceļa ceļu ceļa ceļu DAT ceļam ceļiem celim ceļiem ACC ceļu ceļus celi ceļus INS ceļu ceļiem celi ceļiem LOC ceļā ceļos celī ceļos VOC ceļ ceļi celi ceļi There are several other issues with this method that could potentially limit its usefulness.",
"First, the UM tables only include verbs, nouns and adjectives, whereas we test the system on UDT data, which includes all parts of speech.",
"Second, by excluding ambiguous forms, we may be restricting the added examples to a non-representative subset of the potential inflections, or the system may simply ignore the context because it isn't needed for these examples.",
"Finally, there are some annotation differences between UM and UDT.",
"7 Despite all of these issues, however, we show below that the added examples and their contexts do actually help.",
"Experimental Setup Baselines and Training Parameters We use four baselines: (1) Lemming 8 (Müller et al., 2015) is a context-sensitive system that uses log-linear models to jointly tag and lemmatize the data, and is trained on sentences annotated with both lemmas and POS tags.",
"(2) The hard monotonic attention model (HMAM) 9 (Makarov and Clematide, 2018b) is a neural sequence-tosequence model with a hard attention mechanism that advances through the sequence monotonically.",
"It is trained on word-lemma pairs (without context) 7 Recent efforts to unify the two resources have mostly focused on validating dataset schema (McCarthy et al., 2018) , leaving conflicts in word lemmas unresolved.",
"We estimated (by counting types that are unambiguous in each dataset but have different lemmas across them) that annotation inconsistencies affect up to 1% of types in the languages we used.",
"8 http://cistern.cis.lmu.de/lemming 9 https://github.com/ZurichNLP/ coling2018-neural-transition-basedmorphology with character-level alignments learned in a preprocessing step using an alignment model, and it has proved to be competitive in low resource scenarios.",
"(3) Our naive Baseline outputs the most frequent lemma (or one lemma at random from the options that are equally frequent) for words observed in training.",
"For unseen words it outputs the wordform itself.",
"(4) We also try a baseline data augmentation approach (AE Aug Baseline) inspired by Bergmanis et al.",
"(2017) and Kann and Schütze (2017) , who showed that adding training examples where the network simply learns to auto-encode corpus words can improve morphological inflection results in low-resource settings.",
"The AE Aug Baseline is a variant of Lematus 0-ch which augments the UDT lemmatization examples by auto-encoding the inflected forms of the UM examples (i.e., it just treats them as corpus words).",
"Comparing AE Aug Baseline to Lematus 0-ch augmented with UM lemma-inflection examples tells us whether using the UM lemma information helps more than simply auto-encoding more inflected examples.",
"To train the models we use the default settings for Lemming and the suggested lemmatization parameters for HMAM.",
"We mainly follow the hyperparameters used by Bergmanis and Goldwater (2018) for Lematus; details are in Appendix B.",
"Languages and Training Data We conduct preliminary experiments on five development languages: Estonian, Finnish, Latvian, Polish, and Russian.",
"In our final experiments we also add Bulgarian, Czech, Romanian, Swedish and Turkish.",
"We vary the amount and type of training data (types vs. tokens, UDT only, UM only, or UDT plus up to 10k UM examples), as described in Section 4.",
"To obtain N UM-based training examples, we select the first N unambiguous UM types (with their sentence contexts) from shuffled Wikipedia sentences.",
"For experiments with j > 1 examples per type, we first find all UM types with at least j sentence contexts in Wikipedia and then choose the N distinct types and their j contexts uniformly at random.",
"Evaluation To evaluate models' ability to lemmatize wordforms in their sentence context we follow Bergmanis and Goldwater (2018) and use the full UDT development and test sets.",
"Unlike Bergmanis and Goldwater (2018) who reported token level lemmatization exact match accuracy, we report type-level micro averaged lemmatization ex- act match accuracy.",
"This measure better reflects improvements on unseen words, which tend to be rare but are more important (since a most-frequentlemma baseline does very well on seen words, as shown by Bergmanis and Goldwater (2018) ).",
"We separately report performance on unseen and ambiguous tokens.",
"For a fair comparison across scenarios with different training sets, we count as unseen only words that are not ambiguous and are absent from all training sets/scenarios introduced in Section 4.",
"Due to the small training sets, between 70-90% of dev set types are classed as unseen in each language.",
"We define a type as ambiguous if the empirical entropy over its lemmas is greater than 0.1 in the full original UDT training splits.",
"10 According to this measure, only 1.2-5.3% of dev set types are classed as ambiguous in each language.",
"Significance Testing All systems are trained and tested on ten languages.",
"To test for statistically significant differences between the results of two systems we use a Monte Carlo method: for each set of results (i.e.",
"a set of 10 numerical values) we generate 10000 random samples, where each sample swaps the results of the two systems for each language with a probability of 0.5.",
"We then obtain a p-value as the proportion of samples for which the difference on average was at least as large as the difference observed in our experiments.",
"1k tokens vs. first 1k distinct types of the UDT training sets.",
"Table 2 shows that if only 1k examples are available, using types is clearly better for all systems.",
"Although Lematus does relatively poorly on the token data, it benefits the most from switching to types, putting it on par with HMAM and suggesting is it likely to benefit more from additional type data.",
"Lemming requires token-based data, but does worse than HMAM (a context-free method) in the token-based setting, and we also see no benefit from context in comparing Lematus 20-ch vs Lematus 0-ch.",
"So overall, in this very low-resource scenario with no data augmentation, context does not appear to help.",
"Using UM + Wikipedia Only We now try training only on UM + Wikipedia examples, rather than examples from UDT.",
"We use 1k, 2k or 5k unambiguous types from UM with a single example context from Wikipedia for each.",
"With 5k types we also try adding more example contexts (2, 3, or 5 examples for each type).",
"Figure 1 presents the results (for unseen words only).",
"As with the UDT experiments, there is little difference between Lematus 20-ch and Lematus 0ch in the smallest data setting.",
"However, when the number of training types increases to 5k, the benefits of context begin to show, with Lematus 20-ch yielding a 1.6% statistically significant (p < 0.001) improvement over Lematus 0-ch.",
"The results for increasing the number of examples per type are numerically higher than the one-example case, but the differences are not statistically significant.",
"It is worth noting that the accuracy even with 5k UM types is considerably lower than the accuracy of the model trained on only 1k UDT types (see Table 2 ).",
"We believe this discrepancy is due to the issues of biased/incomplete data noted above.",
"types with contexts from Wikipedia.",
"Table 3 summarizes the results, showing that despite the lower quality of the UM + Wikipedia examples, using them improves results of all systems, and more so with more examples.",
"Improvements are especially strong for unseen types, which constitute more than 70% of types in the dev set.",
"Furthermore, the benefit of the additional UM examples is above and beyond the effect of auto-encoding (AE Aug Baseline) for all systems in all data scenarios.",
"Considering the two context-free models, HMAM does better on the un-augmented 1k UDT data, but (as predicted by our results above) it benefits less from data augmentation than does Lematus 0-ch, so with added data they are statistically equivalent (p = 0.07 on the test set with 10k UM).",
"More importantly, Lematus 20-ch begins to outperform the context-free models with as few as 1k UM + Wikipedia examples, and the difference increases with more examples, eventually reaching over 4% better on the test set than the next best model (Lematus 0-ch) when 10k UM + Wikipedia examples are used (p < 0.001) This indicates that the system can learn useful contextual cues even from unambiguous training examples.",
"Finally, Figure 2 gives a breakdown of Lematus 20-ch dev set accuracy for individual languages, showing that data augmentation helps consistently, although results suggest diminishing returns.",
"Data Augmentation in Medium Resource Setting To examine the extent to which augmented data can help in the medium resource setting of 10k continuous tokens of UDT used in previous work, we follow Bergmanis and Goldwater (2018) and train Lematus 20-ch models for all ten languages using the first 10k tokens of UDT and compare them with models trained on 10k tokens of UDT augmented with 10k UM types.",
"To provide a better comparison of our results, we report both the type and the token level development set accuracy.",
"First of all, Table 4 shows that training on 10k continuous tokens of UDT yields a token level accuracy that is about 8% higher than when using the 1k types of UDT augmented with 10k UM types-the best-performing data augmentation systems (see Table 3 ).",
"Again, we believe this performance gap is due to the issues with the biased/incomplete data noted above.",
"For example, we analyzed errors that were unique to the model trained on the Latvian augmented data and found that 41% of the errors were due to wrongly lemmatized words other than nouns, verbs, and adjectives-the three POSs with available inflection tables in UM.",
"For instance, improperly lemmatized pronouns amounted to 14% of the errors on the Latvian dev set.",
"Table 4 also shows that UM examples with Wikipedia contexts benefit lemmatization not only in the low but also the medium resource setting, yielding statistically significant type and token level accuracy gains over models trained on 10k UDT continuous tokens alone (for both Unseen and All p < 0.001).",
"Conclusion We proposed a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive lemmatization model.",
"The proposed method uses Wikipedia sentences to provide contextualized examples for unambiguous inflection-lemma pairs from UniMorph tables.",
"These examples are noisy and biased, but nevertheless they improve lemmatization accuracy on all ten languages both in low (1k) and medium (10k) resource settings.",
"In particular, we showed that context is helpful, both overall and especially on unseen words-the first work we know of to demonstrate improvements from context in a very low-resource setting.",
"A Lematus Training Lematus is implemented using the Nematus machine translation toolkit 11 .",
"We use default training parameters of Lematus as specified by Bergmanis and Goldwater (2018) except for early stopping with patience (Prechelt, 1998) which we increase to 20.",
"Similar to Bergmanis and Goldwater (2018) we use the first epochs as a burn-in period, after which we validate the current model by its lemmatization exact match accuracy on the first 3k instances of development set and save this model if it performs better than the previous best model.",
"We choose a burn-in period of 20 and validation interval of 5 epochs for models that we train on datasets up to 2k instances and a burn-in period of 10 and validation interval of 2 epochs for others.",
"As we work with considerably smaller datasets than Bergmanis and Goldwater (2018) we reduce the effective model size and increase the rate of convergence by tying the input embeddings of the encoder, the decoder and the softmax output embeddings (Press and Wolf, 2017).",
"B Data Preparation Wikipedia database dumps contain XML structured articles that are formatted using the wikitext markup language.",
"To obtain wordforms in their sentence context we 1) use WikiExtractor 12 to extract plain text from Wikipedia database dumps, followed by scripts from Moses statistical machine translation system 13 (Koehn et al., 2007) to 2) split text into sentences (split-sentences.perl), and 3) extract separate tokens (tokenizer.perl).",
"Finally, we shuffle the extracted sentences to encourage homogeneous type distribution across the entire text.",
"Table 3 .",
"C Result Breakdown by Language"
]
} | {
"paper_header_number": [
"1",
"2.1",
"3",
"5"
],
"paper_header_content": [
"Introduction",
"Data Augmentation",
"Experimental Setup",
"Conclusion"
]
} | GEM-SciDuet-train-4#paper-965#slide-4 | N types N tokens | Training on 1k UDT tokens/types | Training on 1k UDT tokens/types | [] |
GEM-SciDuet-train-4#paper-965#slide-5 | 965 | Data Augmentation for Context-Sensitive Neural Lemmatization Using Inflection Tables and Raw Text | Lemmatization aims to reduce the sparse data problem by relating the inflected forms of a word to its dictionary form. Using context can help, both for unseen and ambiguous words. Yet most context-sensitive approaches require full lemma-annotated sentences for training, which may be scarce or unavailable in lowresource languages. In addition (as shown here), in a low-resource setting, a lemmatizer can learn more from n labeled examples of distinct words (types) than from n (contiguous) labeled tokens, since the latter contain far fewer distinct types. To combine the efficiency of type-based learning with the benefits of context, we propose a way to train a context-sensitive lemmatizer with little or no labeled corpus data, using inflection tables from the UniMorph project and raw text examples from Wikipedia that provide sentence contexts for the unambiguous UniMorph examples. Despite these being unambiguous examples, the model successfully generalizes from them, leading to improved results (both overall, and especially on unseen words) in comparison to a baseline that does not use context. Method Lematus (Bergmanis and Goldwater, 2018) is a neural sequence-to-sequence model with attention 2 Garrette et al. (2013) found the same for POS tagging. 3 Code and data: https://bitbucket.org/ tomsbergmanis/data_augumentation_um_wiki 4 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97
],
"paper_content_text": [
"Introduction Many lemmatizers work on isolated wordforms (Wicentowski, 2002; Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"Lemmatizing in context can improve accuracy on ambiguous and unseen words (Bergmanis and Goldwater, 2018) , but most systems for contextsensitive lemmatization must train on complete sentences labeled with POS and/or morphological tags as well as lemmas, and have only been tested with 20k-300k training tokens (Chrupała et al., 2008; Müller et al., 2015; Chakrabarty et al., 2017) .",
"1 1 The smallest of these corpora contains 20k tokens of Bengali annotated only with lemmas, which Chakrabarty et al.",
"(2017) reported took around two person months to create.",
"Intuitively, though, sentence-annotated data is inefficient for training a lemmatizer, especially in low-resource settings.",
"Training on (say) 1000 word types will provide far more information about a language's morphology than training on 1000 contiguous tokens, where fewer types are represented.",
"As noted above, sentence data can help with ambiguous and unseen words, but we show here that when data is scarce, this effect is small relative to the benefit of seeing more word types.",
"2 Motivated by this result, we propose a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive model.",
"3 We use Lematus (Bergmanis and Goldwater, 2018), a state-of-theart lemmatizer that learns from lemma-annotated words in their N -character contexts.",
"No predictions about surrounding words are used, so fully annotated training sentences are not needed.",
"We exploit this fact by combining two sources of training data: 1k lemma-annotated types (with contexts) from the Universal Dependency Treebank (UDT) v2.2 4 (Nivre et al., 2017) , plus examples obtained by finding unambiguous word-lemma pairs in inflection tables from the Universal Morphology (UM) project 5 and collecting sentence contexts for them from Wikipedia.",
"Although these examples are noisy and biased, we show that they improve lemmatization accuracy in experiments on 10 languages, and that the use of context helps, both overall and especially on unseen words.",
"inspired by the re-inflection model of Kann and Schütze (2016) , which won the 2016 SIGMOR-PHON shared task (Cotterell et al., 2016) .",
"It is built using the Nematus machine translation toolkit, 6 which uses the architecture of Sennrich et al.",
"(2017) : a 2-layer bidirectional GRU encoder and a 2-layer decoder with a conditional GRU (Sennrich et al., 2017) in the first layer and a GRU in the second layer.",
"Lematus takes as input a character sequence representing the wordform in its N -character context, and outputs the characters of the lemma.",
"Special input symbols are used to represent the left and right boundary of the target wordform (<lc>, <rc>) and other word boundaries (<s>).",
"For example, if N = 15, the system trained on Latvian would be expected to produce the characters of the lemma ceļš (meaning road) given input such as: s a k a <s> p aš v a l dī b u <lc> c e ļ u <rc> u n <s> i e l u <s> r eǵ i s t r When N = 0 (Lematus 0-ch), no context is used, making Lematus 0-ch comparable to other systems that do not model context (Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"In our experiments we use both Lematus 0-ch and Lematus 20-ch (20 characters of context), which was the best-performing system reported by Bergmanis and Goldwater (2018).",
"Data Augmentation Our data augmentation method uses UM inflection tables and creates additional training examples by finding Wikipedia sentences that use the inflected wordforms in context, pairing them with their lemma as shown in the inflection table.",
"However, we cannot use all the words in the tables because some of them are ambiguous: for example, Figure 1 shows that the form ceļi could be lemmatized either as ceļš or celis.",
"SG PL SG PL NOM ceļš ceļi celis ceļi GEN ceļa ceļu ceļa ceļu DAT ceļam ceļiem celim ceļiem ACC ceļu ceļus celi ceļus INS ceļu ceļiem celi ceļiem LOC ceļā ceļos celī ceļos VOC ceļ ceļi celi ceļi There are several other issues with this method that could potentially limit its usefulness.",
"First, the UM tables only include verbs, nouns and adjectives, whereas we test the system on UDT data, which includes all parts of speech.",
"Second, by excluding ambiguous forms, we may be restricting the added examples to a non-representative subset of the potential inflections, or the system may simply ignore the context because it isn't needed for these examples.",
"Finally, there are some annotation differences between UM and UDT.",
"7 Despite all of these issues, however, we show below that the added examples and their contexts do actually help.",
"Experimental Setup Baselines and Training Parameters We use four baselines: (1) Lemming 8 (Müller et al., 2015) is a context-sensitive system that uses log-linear models to jointly tag and lemmatize the data, and is trained on sentences annotated with both lemmas and POS tags.",
"(2) The hard monotonic attention model (HMAM) 9 (Makarov and Clematide, 2018b) is a neural sequence-tosequence model with a hard attention mechanism that advances through the sequence monotonically.",
"It is trained on word-lemma pairs (without context) 7 Recent efforts to unify the two resources have mostly focused on validating dataset schema (McCarthy et al., 2018) , leaving conflicts in word lemmas unresolved.",
"We estimated (by counting types that are unambiguous in each dataset but have different lemmas across them) that annotation inconsistencies affect up to 1% of types in the languages we used.",
"8 http://cistern.cis.lmu.de/lemming 9 https://github.com/ZurichNLP/ coling2018-neural-transition-basedmorphology with character-level alignments learned in a preprocessing step using an alignment model, and it has proved to be competitive in low resource scenarios.",
"(3) Our naive Baseline outputs the most frequent lemma (or one lemma at random from the options that are equally frequent) for words observed in training.",
"For unseen words it outputs the wordform itself.",
"(4) We also try a baseline data augmentation approach (AE Aug Baseline) inspired by Bergmanis et al.",
"(2017) and Kann and Schütze (2017) , who showed that adding training examples where the network simply learns to auto-encode corpus words can improve morphological inflection results in low-resource settings.",
"The AE Aug Baseline is a variant of Lematus 0-ch which augments the UDT lemmatization examples by auto-encoding the inflected forms of the UM examples (i.e., it just treats them as corpus words).",
"Comparing AE Aug Baseline to Lematus 0-ch augmented with UM lemma-inflection examples tells us whether using the UM lemma information helps more than simply auto-encoding more inflected examples.",
"To train the models we use the default settings for Lemming and the suggested lemmatization parameters for HMAM.",
"We mainly follow the hyperparameters used by Bergmanis and Goldwater (2018) for Lematus; details are in Appendix B.",
"Languages and Training Data We conduct preliminary experiments on five development languages: Estonian, Finnish, Latvian, Polish, and Russian.",
"In our final experiments we also add Bulgarian, Czech, Romanian, Swedish and Turkish.",
"We vary the amount and type of training data (types vs. tokens, UDT only, UM only, or UDT plus up to 10k UM examples), as described in Section 4.",
"To obtain N UM-based training examples, we select the first N unambiguous UM types (with their sentence contexts) from shuffled Wikipedia sentences.",
"For experiments with j > 1 examples per type, we first find all UM types with at least j sentence contexts in Wikipedia and then choose the N distinct types and their j contexts uniformly at random.",
"Evaluation To evaluate models' ability to lemmatize wordforms in their sentence context we follow Bergmanis and Goldwater (2018) and use the full UDT development and test sets.",
"Unlike Bergmanis and Goldwater (2018) who reported token level lemmatization exact match accuracy, we report type-level micro averaged lemmatization ex- act match accuracy.",
"This measure better reflects improvements on unseen words, which tend to be rare but are more important (since a most-frequentlemma baseline does very well on seen words, as shown by Bergmanis and Goldwater (2018) ).",
"We separately report performance on unseen and ambiguous tokens.",
"For a fair comparison across scenarios with different training sets, we count as unseen only words that are not ambiguous and are absent from all training sets/scenarios introduced in Section 4.",
"Due to the small training sets, between 70-90% of dev set types are classed as unseen in each language.",
"We define a type as ambiguous if the empirical entropy over its lemmas is greater than 0.1 in the full original UDT training splits.",
"10 According to this measure, only 1.2-5.3% of dev set types are classed as ambiguous in each language.",
"Significance Testing All systems are trained and tested on ten languages.",
"To test for statistically significant differences between the results of two systems we use a Monte Carlo method: for each set of results (i.e.",
"a set of 10 numerical values) we generate 10000 random samples, where each sample swaps the results of the two systems for each language with a probability of 0.5.",
"We then obtain a p-value as the proportion of samples for which the difference on average was at least as large as the difference observed in our experiments.",
"1k tokens vs. first 1k distinct types of the UDT training sets.",
"Table 2 shows that if only 1k examples are available, using types is clearly better for all systems.",
"Although Lematus does relatively poorly on the token data, it benefits the most from switching to types, putting it on par with HMAM and suggesting is it likely to benefit more from additional type data.",
"Lemming requires token-based data, but does worse than HMAM (a context-free method) in the token-based setting, and we also see no benefit from context in comparing Lematus 20-ch vs Lematus 0-ch.",
"So overall, in this very low-resource scenario with no data augmentation, context does not appear to help.",
"Using UM + Wikipedia Only We now try training only on UM + Wikipedia examples, rather than examples from UDT.",
"We use 1k, 2k or 5k unambiguous types from UM with a single example context from Wikipedia for each.",
"With 5k types we also try adding more example contexts (2, 3, or 5 examples for each type).",
"Figure 1 presents the results (for unseen words only).",
"As with the UDT experiments, there is little difference between Lematus 20-ch and Lematus 0ch in the smallest data setting.",
"However, when the number of training types increases to 5k, the benefits of context begin to show, with Lematus 20-ch yielding a 1.6% statistically significant (p < 0.001) improvement over Lematus 0-ch.",
"The results for increasing the number of examples per type are numerically higher than the one-example case, but the differences are not statistically significant.",
"It is worth noting that the accuracy even with 5k UM types is considerably lower than the accuracy of the model trained on only 1k UDT types (see Table 2 ).",
"We believe this discrepancy is due to the issues of biased/incomplete data noted above.",
"types with contexts from Wikipedia.",
"Table 3 summarizes the results, showing that despite the lower quality of the UM + Wikipedia examples, using them improves results of all systems, and more so with more examples.",
"Improvements are especially strong for unseen types, which constitute more than 70% of types in the dev set.",
"Furthermore, the benefit of the additional UM examples is above and beyond the effect of auto-encoding (AE Aug Baseline) for all systems in all data scenarios.",
"Considering the two context-free models, HMAM does better on the un-augmented 1k UDT data, but (as predicted by our results above) it benefits less from data augmentation than does Lematus 0-ch, so with added data they are statistically equivalent (p = 0.07 on the test set with 10k UM).",
"More importantly, Lematus 20-ch begins to outperform the context-free models with as few as 1k UM + Wikipedia examples, and the difference increases with more examples, eventually reaching over 4% better on the test set than the next best model (Lematus 0-ch) when 10k UM + Wikipedia examples are used (p < 0.001) This indicates that the system can learn useful contextual cues even from unambiguous training examples.",
"Finally, Figure 2 gives a breakdown of Lematus 20-ch dev set accuracy for individual languages, showing that data augmentation helps consistently, although results suggest diminishing returns.",
"Data Augmentation in Medium Resource Setting To examine the extent to which augmented data can help in the medium resource setting of 10k continuous tokens of UDT used in previous work, we follow Bergmanis and Goldwater (2018) and train Lematus 20-ch models for all ten languages using the first 10k tokens of UDT and compare them with models trained on 10k tokens of UDT augmented with 10k UM types.",
"To provide a better comparison of our results, we report both the type and the token level development set accuracy.",
"First of all, Table 4 shows that training on 10k continuous tokens of UDT yields a token level accuracy that is about 8% higher than when using the 1k types of UDT augmented with 10k UM types-the best-performing data augmentation systems (see Table 3 ).",
"Again, we believe this performance gap is due to the issues with the biased/incomplete data noted above.",
"For example, we analyzed errors that were unique to the model trained on the Latvian augmented data and found that 41% of the errors were due to wrongly lemmatized words other than nouns, verbs, and adjectives-the three POSs with available inflection tables in UM.",
"For instance, improperly lemmatized pronouns amounted to 14% of the errors on the Latvian dev set.",
"Table 4 also shows that UM examples with Wikipedia contexts benefit lemmatization not only in the low but also the medium resource setting, yielding statistically significant type and token level accuracy gains over models trained on 10k UDT continuous tokens alone (for both Unseen and All p < 0.001).",
"Conclusion We proposed a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive lemmatization model.",
"The proposed method uses Wikipedia sentences to provide contextualized examples for unambiguous inflection-lemma pairs from UniMorph tables.",
"These examples are noisy and biased, but nevertheless they improve lemmatization accuracy on all ten languages both in low (1k) and medium (10k) resource settings.",
"In particular, we showed that context is helpful, both overall and especially on unseen words-the first work we know of to demonstrate improvements from context in a very low-resource setting.",
"A Lematus Training Lematus is implemented using the Nematus machine translation toolkit 11 .",
"We use default training parameters of Lematus as specified by Bergmanis and Goldwater (2018) except for early stopping with patience (Prechelt, 1998) which we increase to 20.",
"Similar to Bergmanis and Goldwater (2018) we use the first epochs as a burn-in period, after which we validate the current model by its lemmatization exact match accuracy on the first 3k instances of development set and save this model if it performs better than the previous best model.",
"We choose a burn-in period of 20 and validation interval of 5 epochs for models that we train on datasets up to 2k instances and a burn-in period of 10 and validation interval of 2 epochs for others.",
"As we work with considerably smaller datasets than Bergmanis and Goldwater (2018) we reduce the effective model size and increase the rate of convergence by tying the input embeddings of the encoder, the decoder and the softmax output embeddings (Press and Wolf, 2017).",
"B Data Preparation Wikipedia database dumps contain XML structured articles that are formatted using the wikitext markup language.",
"To obtain wordforms in their sentence context we 1) use WikiExtractor 12 to extract plain text from Wikipedia database dumps, followed by scripts from Moses statistical machine translation system 13 (Koehn et al., 2007) to 2) split text into sentences (split-sentences.perl), and 3) extract separate tokens (tokenizer.perl).",
"Finally, we shuffle the extracted sentences to encourage homogeneous type distribution across the entire text.",
"Table 3 .",
"C Result Breakdown by Language"
]
} | {
"paper_header_number": [
"1",
"2.1",
"3",
"5"
],
"paper_header_content": [
"Introduction",
"Data Augmentation",
"Experimental Setup",
"Conclusion"
]
} | GEM-SciDuet-train-4#paper-965#slide-5 | Types in context | algorithms get smarter computers faster
Bergmanis and Goldwater, 2018 | algorithms get smarter computers faster
Bergmanis and Goldwater, 2018 | [] |
GEM-SciDuet-train-4#paper-965#slide-6 | 965 | Data Augmentation for Context-Sensitive Neural Lemmatization Using Inflection Tables and Raw Text | Lemmatization aims to reduce the sparse data problem by relating the inflected forms of a word to its dictionary form. Using context can help, both for unseen and ambiguous words. Yet most context-sensitive approaches require full lemma-annotated sentences for training, which may be scarce or unavailable in lowresource languages. In addition (as shown here), in a low-resource setting, a lemmatizer can learn more from n labeled examples of distinct words (types) than from n (contiguous) labeled tokens, since the latter contain far fewer distinct types. To combine the efficiency of type-based learning with the benefits of context, we propose a way to train a context-sensitive lemmatizer with little or no labeled corpus data, using inflection tables from the UniMorph project and raw text examples from Wikipedia that provide sentence contexts for the unambiguous UniMorph examples. Despite these being unambiguous examples, the model successfully generalizes from them, leading to improved results (both overall, and especially on unseen words) in comparison to a baseline that does not use context. Method Lematus (Bergmanis and Goldwater, 2018) is a neural sequence-to-sequence model with attention 2 Garrette et al. (2013) found the same for POS tagging. 3 Code and data: https://bitbucket.org/ tomsbergmanis/data_augumentation_um_wiki 4 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97
],
"paper_content_text": [
"Introduction Many lemmatizers work on isolated wordforms (Wicentowski, 2002; Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"Lemmatizing in context can improve accuracy on ambiguous and unseen words (Bergmanis and Goldwater, 2018) , but most systems for contextsensitive lemmatization must train on complete sentences labeled with POS and/or morphological tags as well as lemmas, and have only been tested with 20k-300k training tokens (Chrupała et al., 2008; Müller et al., 2015; Chakrabarty et al., 2017) .",
"1 1 The smallest of these corpora contains 20k tokens of Bengali annotated only with lemmas, which Chakrabarty et al.",
"(2017) reported took around two person months to create.",
"Intuitively, though, sentence-annotated data is inefficient for training a lemmatizer, especially in low-resource settings.",
"Training on (say) 1000 word types will provide far more information about a language's morphology than training on 1000 contiguous tokens, where fewer types are represented.",
"As noted above, sentence data can help with ambiguous and unseen words, but we show here that when data is scarce, this effect is small relative to the benefit of seeing more word types.",
"2 Motivated by this result, we propose a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive model.",
"3 We use Lematus (Bergmanis and Goldwater, 2018), a state-of-theart lemmatizer that learns from lemma-annotated words in their N -character contexts.",
"No predictions about surrounding words are used, so fully annotated training sentences are not needed.",
"We exploit this fact by combining two sources of training data: 1k lemma-annotated types (with contexts) from the Universal Dependency Treebank (UDT) v2.2 4 (Nivre et al., 2017) , plus examples obtained by finding unambiguous word-lemma pairs in inflection tables from the Universal Morphology (UM) project 5 and collecting sentence contexts for them from Wikipedia.",
"Although these examples are noisy and biased, we show that they improve lemmatization accuracy in experiments on 10 languages, and that the use of context helps, both overall and especially on unseen words.",
"inspired by the re-inflection model of Kann and Schütze (2016) , which won the 2016 SIGMOR-PHON shared task (Cotterell et al., 2016) .",
"It is built using the Nematus machine translation toolkit, 6 which uses the architecture of Sennrich et al.",
"(2017) : a 2-layer bidirectional GRU encoder and a 2-layer decoder with a conditional GRU (Sennrich et al., 2017) in the first layer and a GRU in the second layer.",
"Lematus takes as input a character sequence representing the wordform in its N -character context, and outputs the characters of the lemma.",
"Special input symbols are used to represent the left and right boundary of the target wordform (<lc>, <rc>) and other word boundaries (<s>).",
"For example, if N = 15, the system trained on Latvian would be expected to produce the characters of the lemma ceļš (meaning road) given input such as: s a k a <s> p aš v a l dī b u <lc> c e ļ u <rc> u n <s> i e l u <s> r eǵ i s t r When N = 0 (Lematus 0-ch), no context is used, making Lematus 0-ch comparable to other systems that do not model context (Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"In our experiments we use both Lematus 0-ch and Lematus 20-ch (20 characters of context), which was the best-performing system reported by Bergmanis and Goldwater (2018).",
"Data Augmentation Our data augmentation method uses UM inflection tables and creates additional training examples by finding Wikipedia sentences that use the inflected wordforms in context, pairing them with their lemma as shown in the inflection table.",
"However, we cannot use all the words in the tables because some of them are ambiguous: for example, Figure 1 shows that the form ceļi could be lemmatized either as ceļš or celis.",
"SG PL SG PL NOM ceļš ceļi celis ceļi GEN ceļa ceļu ceļa ceļu DAT ceļam ceļiem celim ceļiem ACC ceļu ceļus celi ceļus INS ceļu ceļiem celi ceļiem LOC ceļā ceļos celī ceļos VOC ceļ ceļi celi ceļi There are several other issues with this method that could potentially limit its usefulness.",
"First, the UM tables only include verbs, nouns and adjectives, whereas we test the system on UDT data, which includes all parts of speech.",
"Second, by excluding ambiguous forms, we may be restricting the added examples to a non-representative subset of the potential inflections, or the system may simply ignore the context because it isn't needed for these examples.",
"Finally, there are some annotation differences between UM and UDT.",
"7 Despite all of these issues, however, we show below that the added examples and their contexts do actually help.",
"Experimental Setup Baselines and Training Parameters We use four baselines: (1) Lemming 8 (Müller et al., 2015) is a context-sensitive system that uses log-linear models to jointly tag and lemmatize the data, and is trained on sentences annotated with both lemmas and POS tags.",
"(2) The hard monotonic attention model (HMAM) 9 (Makarov and Clematide, 2018b) is a neural sequence-tosequence model with a hard attention mechanism that advances through the sequence monotonically.",
"It is trained on word-lemma pairs (without context) 7 Recent efforts to unify the two resources have mostly focused on validating dataset schema (McCarthy et al., 2018) , leaving conflicts in word lemmas unresolved.",
"We estimated (by counting types that are unambiguous in each dataset but have different lemmas across them) that annotation inconsistencies affect up to 1% of types in the languages we used.",
"8 http://cistern.cis.lmu.de/lemming 9 https://github.com/ZurichNLP/ coling2018-neural-transition-basedmorphology with character-level alignments learned in a preprocessing step using an alignment model, and it has proved to be competitive in low resource scenarios.",
"(3) Our naive Baseline outputs the most frequent lemma (or one lemma at random from the options that are equally frequent) for words observed in training.",
"For unseen words it outputs the wordform itself.",
"(4) We also try a baseline data augmentation approach (AE Aug Baseline) inspired by Bergmanis et al.",
"(2017) and Kann and Schütze (2017) , who showed that adding training examples where the network simply learns to auto-encode corpus words can improve morphological inflection results in low-resource settings.",
"The AE Aug Baseline is a variant of Lematus 0-ch which augments the UDT lemmatization examples by auto-encoding the inflected forms of the UM examples (i.e., it just treats them as corpus words).",
"Comparing AE Aug Baseline to Lematus 0-ch augmented with UM lemma-inflection examples tells us whether using the UM lemma information helps more than simply auto-encoding more inflected examples.",
"To train the models we use the default settings for Lemming and the suggested lemmatization parameters for HMAM.",
"We mainly follow the hyperparameters used by Bergmanis and Goldwater (2018) for Lematus; details are in Appendix B.",
"Languages and Training Data We conduct preliminary experiments on five development languages: Estonian, Finnish, Latvian, Polish, and Russian.",
"In our final experiments we also add Bulgarian, Czech, Romanian, Swedish and Turkish.",
"We vary the amount and type of training data (types vs. tokens, UDT only, UM only, or UDT plus up to 10k UM examples), as described in Section 4.",
"To obtain N UM-based training examples, we select the first N unambiguous UM types (with their sentence contexts) from shuffled Wikipedia sentences.",
"For experiments with j > 1 examples per type, we first find all UM types with at least j sentence contexts in Wikipedia and then choose the N distinct types and their j contexts uniformly at random.",
"Evaluation To evaluate models' ability to lemmatize wordforms in their sentence context we follow Bergmanis and Goldwater (2018) and use the full UDT development and test sets.",
"Unlike Bergmanis and Goldwater (2018) who reported token level lemmatization exact match accuracy, we report type-level micro averaged lemmatization ex- act match accuracy.",
"This measure better reflects improvements on unseen words, which tend to be rare but are more important (since a most-frequentlemma baseline does very well on seen words, as shown by Bergmanis and Goldwater (2018) ).",
"We separately report performance on unseen and ambiguous tokens.",
"For a fair comparison across scenarios with different training sets, we count as unseen only words that are not ambiguous and are absent from all training sets/scenarios introduced in Section 4.",
"Due to the small training sets, between 70-90% of dev set types are classed as unseen in each language.",
"We define a type as ambiguous if the empirical entropy over its lemmas is greater than 0.1 in the full original UDT training splits.",
"10 According to this measure, only 1.2-5.3% of dev set types are classed as ambiguous in each language.",
"Significance Testing All systems are trained and tested on ten languages.",
"To test for statistically significant differences between the results of two systems we use a Monte Carlo method: for each set of results (i.e.",
"a set of 10 numerical values) we generate 10000 random samples, where each sample swaps the results of the two systems for each language with a probability of 0.5.",
"We then obtain a p-value as the proportion of samples for which the difference on average was at least as large as the difference observed in our experiments.",
"1k tokens vs. first 1k distinct types of the UDT training sets.",
"Table 2 shows that if only 1k examples are available, using types is clearly better for all systems.",
"Although Lematus does relatively poorly on the token data, it benefits the most from switching to types, putting it on par with HMAM and suggesting is it likely to benefit more from additional type data.",
"Lemming requires token-based data, but does worse than HMAM (a context-free method) in the token-based setting, and we also see no benefit from context in comparing Lematus 20-ch vs Lematus 0-ch.",
"So overall, in this very low-resource scenario with no data augmentation, context does not appear to help.",
"Using UM + Wikipedia Only We now try training only on UM + Wikipedia examples, rather than examples from UDT.",
"We use 1k, 2k or 5k unambiguous types from UM with a single example context from Wikipedia for each.",
"With 5k types we also try adding more example contexts (2, 3, or 5 examples for each type).",
"Figure 1 presents the results (for unseen words only).",
"As with the UDT experiments, there is little difference between Lematus 20-ch and Lematus 0ch in the smallest data setting.",
"However, when the number of training types increases to 5k, the benefits of context begin to show, with Lematus 20-ch yielding a 1.6% statistically significant (p < 0.001) improvement over Lematus 0-ch.",
"The results for increasing the number of examples per type are numerically higher than the one-example case, but the differences are not statistically significant.",
"It is worth noting that the accuracy even with 5k UM types is considerably lower than the accuracy of the model trained on only 1k UDT types (see Table 2 ).",
"We believe this discrepancy is due to the issues of biased/incomplete data noted above.",
"types with contexts from Wikipedia.",
"Table 3 summarizes the results, showing that despite the lower quality of the UM + Wikipedia examples, using them improves results of all systems, and more so with more examples.",
"Improvements are especially strong for unseen types, which constitute more than 70% of types in the dev set.",
"Furthermore, the benefit of the additional UM examples is above and beyond the effect of auto-encoding (AE Aug Baseline) for all systems in all data scenarios.",
"Considering the two context-free models, HMAM does better on the un-augmented 1k UDT data, but (as predicted by our results above) it benefits less from data augmentation than does Lematus 0-ch, so with added data they are statistically equivalent (p = 0.07 on the test set with 10k UM).",
"More importantly, Lematus 20-ch begins to outperform the context-free models with as few as 1k UM + Wikipedia examples, and the difference increases with more examples, eventually reaching over 4% better on the test set than the next best model (Lematus 0-ch) when 10k UM + Wikipedia examples are used (p < 0.001) This indicates that the system can learn useful contextual cues even from unambiguous training examples.",
"Finally, Figure 2 gives a breakdown of Lematus 20-ch dev set accuracy for individual languages, showing that data augmentation helps consistently, although results suggest diminishing returns.",
"Data Augmentation in Medium Resource Setting To examine the extent to which augmented data can help in the medium resource setting of 10k continuous tokens of UDT used in previous work, we follow Bergmanis and Goldwater (2018) and train Lematus 20-ch models for all ten languages using the first 10k tokens of UDT and compare them with models trained on 10k tokens of UDT augmented with 10k UM types.",
"To provide a better comparison of our results, we report both the type and the token level development set accuracy.",
"First of all, Table 4 shows that training on 10k continuous tokens of UDT yields a token level accuracy that is about 8% higher than when using the 1k types of UDT augmented with 10k UM types-the best-performing data augmentation systems (see Table 3 ).",
"Again, we believe this performance gap is due to the issues with the biased/incomplete data noted above.",
"For example, we analyzed errors that were unique to the model trained on the Latvian augmented data and found that 41% of the errors were due to wrongly lemmatized words other than nouns, verbs, and adjectives-the three POSs with available inflection tables in UM.",
"For instance, improperly lemmatized pronouns amounted to 14% of the errors on the Latvian dev set.",
"Table 4 also shows that UM examples with Wikipedia contexts benefit lemmatization not only in the low but also the medium resource setting, yielding statistically significant type and token level accuracy gains over models trained on 10k UDT continuous tokens alone (for both Unseen and All p < 0.001).",
"Conclusion We proposed a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive lemmatization model.",
"The proposed method uses Wikipedia sentences to provide contextualized examples for unambiguous inflection-lemma pairs from UniMorph tables.",
"These examples are noisy and biased, but nevertheless they improve lemmatization accuracy on all ten languages both in low (1k) and medium (10k) resource settings.",
"In particular, we showed that context is helpful, both overall and especially on unseen words-the first work we know of to demonstrate improvements from context in a very low-resource setting.",
"A Lematus Training Lematus is implemented using the Nematus machine translation toolkit 11 .",
"We use default training parameters of Lematus as specified by Bergmanis and Goldwater (2018) except for early stopping with patience (Prechelt, 1998) which we increase to 20.",
"Similar to Bergmanis and Goldwater (2018) we use the first epochs as a burn-in period, after which we validate the current model by its lemmatization exact match accuracy on the first 3k instances of development set and save this model if it performs better than the previous best model.",
"We choose a burn-in period of 20 and validation interval of 5 epochs for models that we train on datasets up to 2k instances and a burn-in period of 10 and validation interval of 2 epochs for others.",
"As we work with considerably smaller datasets than Bergmanis and Goldwater (2018) we reduce the effective model size and increase the rate of convergence by tying the input embeddings of the encoder, the decoder and the softmax output embeddings (Press and Wolf, 2017).",
"B Data Preparation Wikipedia database dumps contain XML structured articles that are formatted using the wikitext markup language.",
"To obtain wordforms in their sentence context we 1) use WikiExtractor 12 to extract plain text from Wikipedia database dumps, followed by scripts from Moses statistical machine translation system 13 (Koehn et al., 2007) to 2) split text into sentences (split-sentences.perl), and 3) extract separate tokens (tokenizer.perl).",
"Finally, we shuffle the extracted sentences to encourage homogeneous type distribution across the entire text.",
"Table 3 .",
"C Result Breakdown by Language"
]
} | {
"paper_header_number": [
"1",
"2.1",
"3",
"5"
],
"paper_header_content": [
"Introduction",
"Data Augmentation",
"Experimental Setup",
"Conclusion"
]
} | GEM-SciDuet-train-4#paper-965#slide-6 | Proposal Data Augmentation | ...to get types in context | ...to get types in context | [] |
GEM-SciDuet-train-4#paper-965#slide-7 | 965 | Data Augmentation for Context-Sensitive Neural Lemmatization Using Inflection Tables and Raw Text | Lemmatization aims to reduce the sparse data problem by relating the inflected forms of a word to its dictionary form. Using context can help, both for unseen and ambiguous words. Yet most context-sensitive approaches require full lemma-annotated sentences for training, which may be scarce or unavailable in lowresource languages. In addition (as shown here), in a low-resource setting, a lemmatizer can learn more from n labeled examples of distinct words (types) than from n (contiguous) labeled tokens, since the latter contain far fewer distinct types. To combine the efficiency of type-based learning with the benefits of context, we propose a way to train a context-sensitive lemmatizer with little or no labeled corpus data, using inflection tables from the UniMorph project and raw text examples from Wikipedia that provide sentence contexts for the unambiguous UniMorph examples. Despite these being unambiguous examples, the model successfully generalizes from them, leading to improved results (both overall, and especially on unseen words) in comparison to a baseline that does not use context. Method Lematus (Bergmanis and Goldwater, 2018) is a neural sequence-to-sequence model with attention 2 Garrette et al. (2013) found the same for POS tagging. 3 Code and data: https://bitbucket.org/ tomsbergmanis/data_augumentation_um_wiki 4 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97
],
"paper_content_text": [
"Introduction Many lemmatizers work on isolated wordforms (Wicentowski, 2002; Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"Lemmatizing in context can improve accuracy on ambiguous and unseen words (Bergmanis and Goldwater, 2018) , but most systems for contextsensitive lemmatization must train on complete sentences labeled with POS and/or morphological tags as well as lemmas, and have only been tested with 20k-300k training tokens (Chrupała et al., 2008; Müller et al., 2015; Chakrabarty et al., 2017) .",
"1 1 The smallest of these corpora contains 20k tokens of Bengali annotated only with lemmas, which Chakrabarty et al.",
"(2017) reported took around two person months to create.",
"Intuitively, though, sentence-annotated data is inefficient for training a lemmatizer, especially in low-resource settings.",
"Training on (say) 1000 word types will provide far more information about a language's morphology than training on 1000 contiguous tokens, where fewer types are represented.",
"As noted above, sentence data can help with ambiguous and unseen words, but we show here that when data is scarce, this effect is small relative to the benefit of seeing more word types.",
"2 Motivated by this result, we propose a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive model.",
"3 We use Lematus (Bergmanis and Goldwater, 2018), a state-of-theart lemmatizer that learns from lemma-annotated words in their N -character contexts.",
"No predictions about surrounding words are used, so fully annotated training sentences are not needed.",
"We exploit this fact by combining two sources of training data: 1k lemma-annotated types (with contexts) from the Universal Dependency Treebank (UDT) v2.2 4 (Nivre et al., 2017) , plus examples obtained by finding unambiguous word-lemma pairs in inflection tables from the Universal Morphology (UM) project 5 and collecting sentence contexts for them from Wikipedia.",
"Although these examples are noisy and biased, we show that they improve lemmatization accuracy in experiments on 10 languages, and that the use of context helps, both overall and especially on unseen words.",
"inspired by the re-inflection model of Kann and Schütze (2016) , which won the 2016 SIGMOR-PHON shared task (Cotterell et al., 2016) .",
"It is built using the Nematus machine translation toolkit, 6 which uses the architecture of Sennrich et al.",
"(2017) : a 2-layer bidirectional GRU encoder and a 2-layer decoder with a conditional GRU (Sennrich et al., 2017) in the first layer and a GRU in the second layer.",
"Lematus takes as input a character sequence representing the wordform in its N -character context, and outputs the characters of the lemma.",
"Special input symbols are used to represent the left and right boundary of the target wordform (<lc>, <rc>) and other word boundaries (<s>).",
"For example, if N = 15, the system trained on Latvian would be expected to produce the characters of the lemma ceļš (meaning road) given input such as: s a k a <s> p aš v a l dī b u <lc> c e ļ u <rc> u n <s> i e l u <s> r eǵ i s t r When N = 0 (Lematus 0-ch), no context is used, making Lematus 0-ch comparable to other systems that do not model context (Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"In our experiments we use both Lematus 0-ch and Lematus 20-ch (20 characters of context), which was the best-performing system reported by Bergmanis and Goldwater (2018).",
"Data Augmentation Our data augmentation method uses UM inflection tables and creates additional training examples by finding Wikipedia sentences that use the inflected wordforms in context, pairing them with their lemma as shown in the inflection table.",
"However, we cannot use all the words in the tables because some of them are ambiguous: for example, Figure 1 shows that the form ceļi could be lemmatized either as ceļš or celis.",
"SG PL SG PL NOM ceļš ceļi celis ceļi GEN ceļa ceļu ceļa ceļu DAT ceļam ceļiem celim ceļiem ACC ceļu ceļus celi ceļus INS ceļu ceļiem celi ceļiem LOC ceļā ceļos celī ceļos VOC ceļ ceļi celi ceļi There are several other issues with this method that could potentially limit its usefulness.",
"First, the UM tables only include verbs, nouns and adjectives, whereas we test the system on UDT data, which includes all parts of speech.",
"Second, by excluding ambiguous forms, we may be restricting the added examples to a non-representative subset of the potential inflections, or the system may simply ignore the context because it isn't needed for these examples.",
"Finally, there are some annotation differences between UM and UDT.",
"7 Despite all of these issues, however, we show below that the added examples and their contexts do actually help.",
"Experimental Setup Baselines and Training Parameters We use four baselines: (1) Lemming 8 (Müller et al., 2015) is a context-sensitive system that uses log-linear models to jointly tag and lemmatize the data, and is trained on sentences annotated with both lemmas and POS tags.",
"(2) The hard monotonic attention model (HMAM) 9 (Makarov and Clematide, 2018b) is a neural sequence-tosequence model with a hard attention mechanism that advances through the sequence monotonically.",
"It is trained on word-lemma pairs (without context) 7 Recent efforts to unify the two resources have mostly focused on validating dataset schema (McCarthy et al., 2018) , leaving conflicts in word lemmas unresolved.",
"We estimated (by counting types that are unambiguous in each dataset but have different lemmas across them) that annotation inconsistencies affect up to 1% of types in the languages we used.",
"8 http://cistern.cis.lmu.de/lemming 9 https://github.com/ZurichNLP/ coling2018-neural-transition-basedmorphology with character-level alignments learned in a preprocessing step using an alignment model, and it has proved to be competitive in low resource scenarios.",
"(3) Our naive Baseline outputs the most frequent lemma (or one lemma at random from the options that are equally frequent) for words observed in training.",
"For unseen words it outputs the wordform itself.",
"(4) We also try a baseline data augmentation approach (AE Aug Baseline) inspired by Bergmanis et al.",
"(2017) and Kann and Schütze (2017) , who showed that adding training examples where the network simply learns to auto-encode corpus words can improve morphological inflection results in low-resource settings.",
"The AE Aug Baseline is a variant of Lematus 0-ch which augments the UDT lemmatization examples by auto-encoding the inflected forms of the UM examples (i.e., it just treats them as corpus words).",
"Comparing AE Aug Baseline to Lematus 0-ch augmented with UM lemma-inflection examples tells us whether using the UM lemma information helps more than simply auto-encoding more inflected examples.",
"To train the models we use the default settings for Lemming and the suggested lemmatization parameters for HMAM.",
"We mainly follow the hyperparameters used by Bergmanis and Goldwater (2018) for Lematus; details are in Appendix B.",
"Languages and Training Data We conduct preliminary experiments on five development languages: Estonian, Finnish, Latvian, Polish, and Russian.",
"In our final experiments we also add Bulgarian, Czech, Romanian, Swedish and Turkish.",
"We vary the amount and type of training data (types vs. tokens, UDT only, UM only, or UDT plus up to 10k UM examples), as described in Section 4.",
"To obtain N UM-based training examples, we select the first N unambiguous UM types (with their sentence contexts) from shuffled Wikipedia sentences.",
"For experiments with j > 1 examples per type, we first find all UM types with at least j sentence contexts in Wikipedia and then choose the N distinct types and their j contexts uniformly at random.",
"Evaluation To evaluate models' ability to lemmatize wordforms in their sentence context we follow Bergmanis and Goldwater (2018) and use the full UDT development and test sets.",
"Unlike Bergmanis and Goldwater (2018) who reported token level lemmatization exact match accuracy, we report type-level micro averaged lemmatization ex- act match accuracy.",
"This measure better reflects improvements on unseen words, which tend to be rare but are more important (since a most-frequentlemma baseline does very well on seen words, as shown by Bergmanis and Goldwater (2018) ).",
"We separately report performance on unseen and ambiguous tokens.",
"For a fair comparison across scenarios with different training sets, we count as unseen only words that are not ambiguous and are absent from all training sets/scenarios introduced in Section 4.",
"Due to the small training sets, between 70-90% of dev set types are classed as unseen in each language.",
"We define a type as ambiguous if the empirical entropy over its lemmas is greater than 0.1 in the full original UDT training splits.",
"10 According to this measure, only 1.2-5.3% of dev set types are classed as ambiguous in each language.",
"Significance Testing All systems are trained and tested on ten languages.",
"To test for statistically significant differences between the results of two systems we use a Monte Carlo method: for each set of results (i.e.",
"a set of 10 numerical values) we generate 10000 random samples, where each sample swaps the results of the two systems for each language with a probability of 0.5.",
"We then obtain a p-value as the proportion of samples for which the difference on average was at least as large as the difference observed in our experiments.",
"1k tokens vs. first 1k distinct types of the UDT training sets.",
"Table 2 shows that if only 1k examples are available, using types is clearly better for all systems.",
"Although Lematus does relatively poorly on the token data, it benefits the most from switching to types, putting it on par with HMAM and suggesting is it likely to benefit more from additional type data.",
"Lemming requires token-based data, but does worse than HMAM (a context-free method) in the token-based setting, and we also see no benefit from context in comparing Lematus 20-ch vs Lematus 0-ch.",
"So overall, in this very low-resource scenario with no data augmentation, context does not appear to help.",
"Using UM + Wikipedia Only We now try training only on UM + Wikipedia examples, rather than examples from UDT.",
"We use 1k, 2k or 5k unambiguous types from UM with a single example context from Wikipedia for each.",
"With 5k types we also try adding more example contexts (2, 3, or 5 examples for each type).",
"Figure 1 presents the results (for unseen words only).",
"As with the UDT experiments, there is little difference between Lematus 20-ch and Lematus 0ch in the smallest data setting.",
"However, when the number of training types increases to 5k, the benefits of context begin to show, with Lematus 20-ch yielding a 1.6% statistically significant (p < 0.001) improvement over Lematus 0-ch.",
"The results for increasing the number of examples per type are numerically higher than the one-example case, but the differences are not statistically significant.",
"It is worth noting that the accuracy even with 5k UM types is considerably lower than the accuracy of the model trained on only 1k UDT types (see Table 2 ).",
"We believe this discrepancy is due to the issues of biased/incomplete data noted above.",
"types with contexts from Wikipedia.",
"Table 3 summarizes the results, showing that despite the lower quality of the UM + Wikipedia examples, using them improves results of all systems, and more so with more examples.",
"Improvements are especially strong for unseen types, which constitute more than 70% of types in the dev set.",
"Furthermore, the benefit of the additional UM examples is above and beyond the effect of auto-encoding (AE Aug Baseline) for all systems in all data scenarios.",
"Considering the two context-free models, HMAM does better on the un-augmented 1k UDT data, but (as predicted by our results above) it benefits less from data augmentation than does Lematus 0-ch, so with added data they are statistically equivalent (p = 0.07 on the test set with 10k UM).",
"More importantly, Lematus 20-ch begins to outperform the context-free models with as few as 1k UM + Wikipedia examples, and the difference increases with more examples, eventually reaching over 4% better on the test set than the next best model (Lematus 0-ch) when 10k UM + Wikipedia examples are used (p < 0.001) This indicates that the system can learn useful contextual cues even from unambiguous training examples.",
"Finally, Figure 2 gives a breakdown of Lematus 20-ch dev set accuracy for individual languages, showing that data augmentation helps consistently, although results suggest diminishing returns.",
"Data Augmentation in Medium Resource Setting To examine the extent to which augmented data can help in the medium resource setting of 10k continuous tokens of UDT used in previous work, we follow Bergmanis and Goldwater (2018) and train Lematus 20-ch models for all ten languages using the first 10k tokens of UDT and compare them with models trained on 10k tokens of UDT augmented with 10k UM types.",
"To provide a better comparison of our results, we report both the type and the token level development set accuracy.",
"First of all, Table 4 shows that training on 10k continuous tokens of UDT yields a token level accuracy that is about 8% higher than when using the 1k types of UDT augmented with 10k UM types-the best-performing data augmentation systems (see Table 3 ).",
"Again, we believe this performance gap is due to the issues with the biased/incomplete data noted above.",
"For example, we analyzed errors that were unique to the model trained on the Latvian augmented data and found that 41% of the errors were due to wrongly lemmatized words other than nouns, verbs, and adjectives-the three POSs with available inflection tables in UM.",
"For instance, improperly lemmatized pronouns amounted to 14% of the errors on the Latvian dev set.",
"Table 4 also shows that UM examples with Wikipedia contexts benefit lemmatization not only in the low but also the medium resource setting, yielding statistically significant type and token level accuracy gains over models trained on 10k UDT continuous tokens alone (for both Unseen and All p < 0.001).",
"Conclusion We proposed a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive lemmatization model.",
"The proposed method uses Wikipedia sentences to provide contextualized examples for unambiguous inflection-lemma pairs from UniMorph tables.",
"These examples are noisy and biased, but nevertheless they improve lemmatization accuracy on all ten languages both in low (1k) and medium (10k) resource settings.",
"In particular, we showed that context is helpful, both overall and especially on unseen words-the first work we know of to demonstrate improvements from context in a very low-resource setting.",
"A Lematus Training Lematus is implemented using the Nematus machine translation toolkit 11 .",
"We use default training parameters of Lematus as specified by Bergmanis and Goldwater (2018) except for early stopping with patience (Prechelt, 1998) which we increase to 20.",
"Similar to Bergmanis and Goldwater (2018) we use the first epochs as a burn-in period, after which we validate the current model by its lemmatization exact match accuracy on the first 3k instances of development set and save this model if it performs better than the previous best model.",
"We choose a burn-in period of 20 and validation interval of 5 epochs for models that we train on datasets up to 2k instances and a burn-in period of 10 and validation interval of 2 epochs for others.",
"As we work with considerably smaller datasets than Bergmanis and Goldwater (2018) we reduce the effective model size and increase the rate of convergence by tying the input embeddings of the encoder, the decoder and the softmax output embeddings (Press and Wolf, 2017).",
"B Data Preparation Wikipedia database dumps contain XML structured articles that are formatted using the wikitext markup language.",
"To obtain wordforms in their sentence context we 1) use WikiExtractor 12 to extract plain text from Wikipedia database dumps, followed by scripts from Moses statistical machine translation system 13 (Koehn et al., 2007) to 2) split text into sentences (split-sentences.perl), and 3) extract separate tokens (tokenizer.perl).",
"Finally, we shuffle the extracted sentences to encourage homogeneous type distribution across the entire text.",
"Table 3 .",
"C Result Breakdown by Language"
]
} | {
"paper_header_number": [
"1",
"2.1",
"3",
"5"
],
"paper_header_content": [
"Introduction",
"Data Augmentation",
"Experimental Setup",
"Conclusion"
]
} | GEM-SciDuet-train-4#paper-965#slide-7 | Method Data Augmentation | Inflection cels cela N;LOC;SG
Dzives pedeja cela pavadot musu cels
Context cels cela N;LOC;SG
Lemma cels cela N;LOC;SG | Inflection cels cela N;LOC;SG
Dzives pedeja cela pavadot musu cels
Context cels cela N;LOC;SG
Lemma cels cela N;LOC;SG | [] |
GEM-SciDuet-train-4#paper-965#slide-8 | 965 | Data Augmentation for Context-Sensitive Neural Lemmatization Using Inflection Tables and Raw Text | Lemmatization aims to reduce the sparse data problem by relating the inflected forms of a word to its dictionary form. Using context can help, both for unseen and ambiguous words. Yet most context-sensitive approaches require full lemma-annotated sentences for training, which may be scarce or unavailable in lowresource languages. In addition (as shown here), in a low-resource setting, a lemmatizer can learn more from n labeled examples of distinct words (types) than from n (contiguous) labeled tokens, since the latter contain far fewer distinct types. To combine the efficiency of type-based learning with the benefits of context, we propose a way to train a context-sensitive lemmatizer with little or no labeled corpus data, using inflection tables from the UniMorph project and raw text examples from Wikipedia that provide sentence contexts for the unambiguous UniMorph examples. Despite these being unambiguous examples, the model successfully generalizes from them, leading to improved results (both overall, and especially on unseen words) in comparison to a baseline that does not use context. Method Lematus (Bergmanis and Goldwater, 2018) is a neural sequence-to-sequence model with attention 2 Garrette et al. (2013) found the same for POS tagging. 3 Code and data: https://bitbucket.org/ tomsbergmanis/data_augumentation_um_wiki 4 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97
],
"paper_content_text": [
"Introduction Many lemmatizers work on isolated wordforms (Wicentowski, 2002; Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"Lemmatizing in context can improve accuracy on ambiguous and unseen words (Bergmanis and Goldwater, 2018) , but most systems for contextsensitive lemmatization must train on complete sentences labeled with POS and/or morphological tags as well as lemmas, and have only been tested with 20k-300k training tokens (Chrupała et al., 2008; Müller et al., 2015; Chakrabarty et al., 2017) .",
"1 1 The smallest of these corpora contains 20k tokens of Bengali annotated only with lemmas, which Chakrabarty et al.",
"(2017) reported took around two person months to create.",
"Intuitively, though, sentence-annotated data is inefficient for training a lemmatizer, especially in low-resource settings.",
"Training on (say) 1000 word types will provide far more information about a language's morphology than training on 1000 contiguous tokens, where fewer types are represented.",
"As noted above, sentence data can help with ambiguous and unseen words, but we show here that when data is scarce, this effect is small relative to the benefit of seeing more word types.",
"2 Motivated by this result, we propose a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive model.",
"3 We use Lematus (Bergmanis and Goldwater, 2018), a state-of-theart lemmatizer that learns from lemma-annotated words in their N -character contexts.",
"No predictions about surrounding words are used, so fully annotated training sentences are not needed.",
"We exploit this fact by combining two sources of training data: 1k lemma-annotated types (with contexts) from the Universal Dependency Treebank (UDT) v2.2 4 (Nivre et al., 2017) , plus examples obtained by finding unambiguous word-lemma pairs in inflection tables from the Universal Morphology (UM) project 5 and collecting sentence contexts for them from Wikipedia.",
"Although these examples are noisy and biased, we show that they improve lemmatization accuracy in experiments on 10 languages, and that the use of context helps, both overall and especially on unseen words.",
"inspired by the re-inflection model of Kann and Schütze (2016) , which won the 2016 SIGMOR-PHON shared task (Cotterell et al., 2016) .",
"It is built using the Nematus machine translation toolkit, 6 which uses the architecture of Sennrich et al.",
"(2017) : a 2-layer bidirectional GRU encoder and a 2-layer decoder with a conditional GRU (Sennrich et al., 2017) in the first layer and a GRU in the second layer.",
"Lematus takes as input a character sequence representing the wordform in its N -character context, and outputs the characters of the lemma.",
"Special input symbols are used to represent the left and right boundary of the target wordform (<lc>, <rc>) and other word boundaries (<s>).",
"For example, if N = 15, the system trained on Latvian would be expected to produce the characters of the lemma ceļš (meaning road) given input such as: s a k a <s> p aš v a l dī b u <lc> c e ļ u <rc> u n <s> i e l u <s> r eǵ i s t r When N = 0 (Lematus 0-ch), no context is used, making Lematus 0-ch comparable to other systems that do not model context (Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"In our experiments we use both Lematus 0-ch and Lematus 20-ch (20 characters of context), which was the best-performing system reported by Bergmanis and Goldwater (2018).",
"Data Augmentation Our data augmentation method uses UM inflection tables and creates additional training examples by finding Wikipedia sentences that use the inflected wordforms in context, pairing them with their lemma as shown in the inflection table.",
"However, we cannot use all the words in the tables because some of them are ambiguous: for example, Figure 1 shows that the form ceļi could be lemmatized either as ceļš or celis.",
"SG PL SG PL NOM ceļš ceļi celis ceļi GEN ceļa ceļu ceļa ceļu DAT ceļam ceļiem celim ceļiem ACC ceļu ceļus celi ceļus INS ceļu ceļiem celi ceļiem LOC ceļā ceļos celī ceļos VOC ceļ ceļi celi ceļi There are several other issues with this method that could potentially limit its usefulness.",
"First, the UM tables only include verbs, nouns and adjectives, whereas we test the system on UDT data, which includes all parts of speech.",
"Second, by excluding ambiguous forms, we may be restricting the added examples to a non-representative subset of the potential inflections, or the system may simply ignore the context because it isn't needed for these examples.",
"Finally, there are some annotation differences between UM and UDT.",
"7 Despite all of these issues, however, we show below that the added examples and their contexts do actually help.",
"Experimental Setup Baselines and Training Parameters We use four baselines: (1) Lemming 8 (Müller et al., 2015) is a context-sensitive system that uses log-linear models to jointly tag and lemmatize the data, and is trained on sentences annotated with both lemmas and POS tags.",
"(2) The hard monotonic attention model (HMAM) 9 (Makarov and Clematide, 2018b) is a neural sequence-tosequence model with a hard attention mechanism that advances through the sequence monotonically.",
"It is trained on word-lemma pairs (without context) 7 Recent efforts to unify the two resources have mostly focused on validating dataset schema (McCarthy et al., 2018) , leaving conflicts in word lemmas unresolved.",
"We estimated (by counting types that are unambiguous in each dataset but have different lemmas across them) that annotation inconsistencies affect up to 1% of types in the languages we used.",
"8 http://cistern.cis.lmu.de/lemming 9 https://github.com/ZurichNLP/ coling2018-neural-transition-basedmorphology with character-level alignments learned in a preprocessing step using an alignment model, and it has proved to be competitive in low resource scenarios.",
"(3) Our naive Baseline outputs the most frequent lemma (or one lemma at random from the options that are equally frequent) for words observed in training.",
"For unseen words it outputs the wordform itself.",
"(4) We also try a baseline data augmentation approach (AE Aug Baseline) inspired by Bergmanis et al.",
"(2017) and Kann and Schütze (2017) , who showed that adding training examples where the network simply learns to auto-encode corpus words can improve morphological inflection results in low-resource settings.",
"The AE Aug Baseline is a variant of Lematus 0-ch which augments the UDT lemmatization examples by auto-encoding the inflected forms of the UM examples (i.e., it just treats them as corpus words).",
"Comparing AE Aug Baseline to Lematus 0-ch augmented with UM lemma-inflection examples tells us whether using the UM lemma information helps more than simply auto-encoding more inflected examples.",
"To train the models we use the default settings for Lemming and the suggested lemmatization parameters for HMAM.",
"We mainly follow the hyperparameters used by Bergmanis and Goldwater (2018) for Lematus; details are in Appendix B.",
"Languages and Training Data We conduct preliminary experiments on five development languages: Estonian, Finnish, Latvian, Polish, and Russian.",
"In our final experiments we also add Bulgarian, Czech, Romanian, Swedish and Turkish.",
"We vary the amount and type of training data (types vs. tokens, UDT only, UM only, or UDT plus up to 10k UM examples), as described in Section 4.",
"To obtain N UM-based training examples, we select the first N unambiguous UM types (with their sentence contexts) from shuffled Wikipedia sentences.",
"For experiments with j > 1 examples per type, we first find all UM types with at least j sentence contexts in Wikipedia and then choose the N distinct types and their j contexts uniformly at random.",
"Evaluation To evaluate models' ability to lemmatize wordforms in their sentence context we follow Bergmanis and Goldwater (2018) and use the full UDT development and test sets.",
"Unlike Bergmanis and Goldwater (2018) who reported token level lemmatization exact match accuracy, we report type-level micro averaged lemmatization ex- act match accuracy.",
"This measure better reflects improvements on unseen words, which tend to be rare but are more important (since a most-frequentlemma baseline does very well on seen words, as shown by Bergmanis and Goldwater (2018) ).",
"We separately report performance on unseen and ambiguous tokens.",
"For a fair comparison across scenarios with different training sets, we count as unseen only words that are not ambiguous and are absent from all training sets/scenarios introduced in Section 4.",
"Due to the small training sets, between 70-90% of dev set types are classed as unseen in each language.",
"We define a type as ambiguous if the empirical entropy over its lemmas is greater than 0.1 in the full original UDT training splits.",
"10 According to this measure, only 1.2-5.3% of dev set types are classed as ambiguous in each language.",
"Significance Testing All systems are trained and tested on ten languages.",
"To test for statistically significant differences between the results of two systems we use a Monte Carlo method: for each set of results (i.e.",
"a set of 10 numerical values) we generate 10000 random samples, where each sample swaps the results of the two systems for each language with a probability of 0.5.",
"We then obtain a p-value as the proportion of samples for which the difference on average was at least as large as the difference observed in our experiments.",
"1k tokens vs. first 1k distinct types of the UDT training sets.",
"Table 2 shows that if only 1k examples are available, using types is clearly better for all systems.",
"Although Lematus does relatively poorly on the token data, it benefits the most from switching to types, putting it on par with HMAM and suggesting is it likely to benefit more from additional type data.",
"Lemming requires token-based data, but does worse than HMAM (a context-free method) in the token-based setting, and we also see no benefit from context in comparing Lematus 20-ch vs Lematus 0-ch.",
"So overall, in this very low-resource scenario with no data augmentation, context does not appear to help.",
"Using UM + Wikipedia Only We now try training only on UM + Wikipedia examples, rather than examples from UDT.",
"We use 1k, 2k or 5k unambiguous types from UM with a single example context from Wikipedia for each.",
"With 5k types we also try adding more example contexts (2, 3, or 5 examples for each type).",
"Figure 1 presents the results (for unseen words only).",
"As with the UDT experiments, there is little difference between Lematus 20-ch and Lematus 0ch in the smallest data setting.",
"However, when the number of training types increases to 5k, the benefits of context begin to show, with Lematus 20-ch yielding a 1.6% statistically significant (p < 0.001) improvement over Lematus 0-ch.",
"The results for increasing the number of examples per type are numerically higher than the one-example case, but the differences are not statistically significant.",
"It is worth noting that the accuracy even with 5k UM types is considerably lower than the accuracy of the model trained on only 1k UDT types (see Table 2 ).",
"We believe this discrepancy is due to the issues of biased/incomplete data noted above.",
"types with contexts from Wikipedia.",
"Table 3 summarizes the results, showing that despite the lower quality of the UM + Wikipedia examples, using them improves results of all systems, and more so with more examples.",
"Improvements are especially strong for unseen types, which constitute more than 70% of types in the dev set.",
"Furthermore, the benefit of the additional UM examples is above and beyond the effect of auto-encoding (AE Aug Baseline) for all systems in all data scenarios.",
"Considering the two context-free models, HMAM does better on the un-augmented 1k UDT data, but (as predicted by our results above) it benefits less from data augmentation than does Lematus 0-ch, so with added data they are statistically equivalent (p = 0.07 on the test set with 10k UM).",
"More importantly, Lematus 20-ch begins to outperform the context-free models with as few as 1k UM + Wikipedia examples, and the difference increases with more examples, eventually reaching over 4% better on the test set than the next best model (Lematus 0-ch) when 10k UM + Wikipedia examples are used (p < 0.001) This indicates that the system can learn useful contextual cues even from unambiguous training examples.",
"Finally, Figure 2 gives a breakdown of Lematus 20-ch dev set accuracy for individual languages, showing that data augmentation helps consistently, although results suggest diminishing returns.",
"Data Augmentation in Medium Resource Setting To examine the extent to which augmented data can help in the medium resource setting of 10k continuous tokens of UDT used in previous work, we follow Bergmanis and Goldwater (2018) and train Lematus 20-ch models for all ten languages using the first 10k tokens of UDT and compare them with models trained on 10k tokens of UDT augmented with 10k UM types.",
"To provide a better comparison of our results, we report both the type and the token level development set accuracy.",
"First of all, Table 4 shows that training on 10k continuous tokens of UDT yields a token level accuracy that is about 8% higher than when using the 1k types of UDT augmented with 10k UM types-the best-performing data augmentation systems (see Table 3 ).",
"Again, we believe this performance gap is due to the issues with the biased/incomplete data noted above.",
"For example, we analyzed errors that were unique to the model trained on the Latvian augmented data and found that 41% of the errors were due to wrongly lemmatized words other than nouns, verbs, and adjectives-the three POSs with available inflection tables in UM.",
"For instance, improperly lemmatized pronouns amounted to 14% of the errors on the Latvian dev set.",
"Table 4 also shows that UM examples with Wikipedia contexts benefit lemmatization not only in the low but also the medium resource setting, yielding statistically significant type and token level accuracy gains over models trained on 10k UDT continuous tokens alone (for both Unseen and All p < 0.001).",
"Conclusion We proposed a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive lemmatization model.",
"The proposed method uses Wikipedia sentences to provide contextualized examples for unambiguous inflection-lemma pairs from UniMorph tables.",
"These examples are noisy and biased, but nevertheless they improve lemmatization accuracy on all ten languages both in low (1k) and medium (10k) resource settings.",
"In particular, we showed that context is helpful, both overall and especially on unseen words-the first work we know of to demonstrate improvements from context in a very low-resource setting.",
"A Lematus Training Lematus is implemented using the Nematus machine translation toolkit 11 .",
"We use default training parameters of Lematus as specified by Bergmanis and Goldwater (2018) except for early stopping with patience (Prechelt, 1998) which we increase to 20.",
"Similar to Bergmanis and Goldwater (2018) we use the first epochs as a burn-in period, after which we validate the current model by its lemmatization exact match accuracy on the first 3k instances of development set and save this model if it performs better than the previous best model.",
"We choose a burn-in period of 20 and validation interval of 5 epochs for models that we train on datasets up to 2k instances and a burn-in period of 10 and validation interval of 2 epochs for others.",
"As we work with considerably smaller datasets than Bergmanis and Goldwater (2018) we reduce the effective model size and increase the rate of convergence by tying the input embeddings of the encoder, the decoder and the softmax output embeddings (Press and Wolf, 2017).",
"B Data Preparation Wikipedia database dumps contain XML structured articles that are formatted using the wikitext markup language.",
"To obtain wordforms in their sentence context we 1) use WikiExtractor 12 to extract plain text from Wikipedia database dumps, followed by scripts from Moses statistical machine translation system 13 (Koehn et al., 2007) to 2) split text into sentences (split-sentences.perl), and 3) extract separate tokens (tokenizer.perl).",
"Finally, we shuffle the extracted sentences to encourage homogeneous type distribution across the entire text.",
"Table 3 .",
"C Result Breakdown by Language"
]
} | {
"paper_header_number": [
"1",
"2.1",
"3",
"5"
],
"paper_header_content": [
"Introduction",
"Data Augmentation",
"Experimental Setup",
"Conclusion"
]
} | GEM-SciDuet-train-4#paper-965#slide-8 | Inflection Tables | INST ar celu ar celiem
Latvian: cels (English: road)
ACC celu celiem celus
celt (build) celot (travel) celis (knee) | INST ar celu ar celiem
Latvian: cels (English: road)
ACC celu celiem celus
celt (build) celot (travel) celis (knee) | [] |
GEM-SciDuet-train-4#paper-965#slide-9 | 965 | Data Augmentation for Context-Sensitive Neural Lemmatization Using Inflection Tables and Raw Text | Lemmatization aims to reduce the sparse data problem by relating the inflected forms of a word to its dictionary form. Using context can help, both for unseen and ambiguous words. Yet most context-sensitive approaches require full lemma-annotated sentences for training, which may be scarce or unavailable in lowresource languages. In addition (as shown here), in a low-resource setting, a lemmatizer can learn more from n labeled examples of distinct words (types) than from n (contiguous) labeled tokens, since the latter contain far fewer distinct types. To combine the efficiency of type-based learning with the benefits of context, we propose a way to train a context-sensitive lemmatizer with little or no labeled corpus data, using inflection tables from the UniMorph project and raw text examples from Wikipedia that provide sentence contexts for the unambiguous UniMorph examples. Despite these being unambiguous examples, the model successfully generalizes from them, leading to improved results (both overall, and especially on unseen words) in comparison to a baseline that does not use context. Method Lematus (Bergmanis and Goldwater, 2018) is a neural sequence-to-sequence model with attention 2 Garrette et al. (2013) found the same for POS tagging. 3 Code and data: https://bitbucket.org/ tomsbergmanis/data_augumentation_um_wiki 4 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97
],
"paper_content_text": [
"Introduction Many lemmatizers work on isolated wordforms (Wicentowski, 2002; Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"Lemmatizing in context can improve accuracy on ambiguous and unseen words (Bergmanis and Goldwater, 2018) , but most systems for contextsensitive lemmatization must train on complete sentences labeled with POS and/or morphological tags as well as lemmas, and have only been tested with 20k-300k training tokens (Chrupała et al., 2008; Müller et al., 2015; Chakrabarty et al., 2017) .",
"1 1 The smallest of these corpora contains 20k tokens of Bengali annotated only with lemmas, which Chakrabarty et al.",
"(2017) reported took around two person months to create.",
"Intuitively, though, sentence-annotated data is inefficient for training a lemmatizer, especially in low-resource settings.",
"Training on (say) 1000 word types will provide far more information about a language's morphology than training on 1000 contiguous tokens, where fewer types are represented.",
"As noted above, sentence data can help with ambiguous and unseen words, but we show here that when data is scarce, this effect is small relative to the benefit of seeing more word types.",
"2 Motivated by this result, we propose a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive model.",
"3 We use Lematus (Bergmanis and Goldwater, 2018), a state-of-theart lemmatizer that learns from lemma-annotated words in their N -character contexts.",
"No predictions about surrounding words are used, so fully annotated training sentences are not needed.",
"We exploit this fact by combining two sources of training data: 1k lemma-annotated types (with contexts) from the Universal Dependency Treebank (UDT) v2.2 4 (Nivre et al., 2017) , plus examples obtained by finding unambiguous word-lemma pairs in inflection tables from the Universal Morphology (UM) project 5 and collecting sentence contexts for them from Wikipedia.",
"Although these examples are noisy and biased, we show that they improve lemmatization accuracy in experiments on 10 languages, and that the use of context helps, both overall and especially on unseen words.",
"inspired by the re-inflection model of Kann and Schütze (2016) , which won the 2016 SIGMOR-PHON shared task (Cotterell et al., 2016) .",
"It is built using the Nematus machine translation toolkit, 6 which uses the architecture of Sennrich et al.",
"(2017) : a 2-layer bidirectional GRU encoder and a 2-layer decoder with a conditional GRU (Sennrich et al., 2017) in the first layer and a GRU in the second layer.",
"Lematus takes as input a character sequence representing the wordform in its N -character context, and outputs the characters of the lemma.",
"Special input symbols are used to represent the left and right boundary of the target wordform (<lc>, <rc>) and other word boundaries (<s>).",
"For example, if N = 15, the system trained on Latvian would be expected to produce the characters of the lemma ceļš (meaning road) given input such as: s a k a <s> p aš v a l dī b u <lc> c e ļ u <rc> u n <s> i e l u <s> r eǵ i s t r When N = 0 (Lematus 0-ch), no context is used, making Lematus 0-ch comparable to other systems that do not model context (Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"In our experiments we use both Lematus 0-ch and Lematus 20-ch (20 characters of context), which was the best-performing system reported by Bergmanis and Goldwater (2018).",
"Data Augmentation Our data augmentation method uses UM inflection tables and creates additional training examples by finding Wikipedia sentences that use the inflected wordforms in context, pairing them with their lemma as shown in the inflection table.",
"However, we cannot use all the words in the tables because some of them are ambiguous: for example, Figure 1 shows that the form ceļi could be lemmatized either as ceļš or celis.",
"SG PL SG PL NOM ceļš ceļi celis ceļi GEN ceļa ceļu ceļa ceļu DAT ceļam ceļiem celim ceļiem ACC ceļu ceļus celi ceļus INS ceļu ceļiem celi ceļiem LOC ceļā ceļos celī ceļos VOC ceļ ceļi celi ceļi There are several other issues with this method that could potentially limit its usefulness.",
"First, the UM tables only include verbs, nouns and adjectives, whereas we test the system on UDT data, which includes all parts of speech.",
"Second, by excluding ambiguous forms, we may be restricting the added examples to a non-representative subset of the potential inflections, or the system may simply ignore the context because it isn't needed for these examples.",
"Finally, there are some annotation differences between UM and UDT.",
"7 Despite all of these issues, however, we show below that the added examples and their contexts do actually help.",
"Experimental Setup Baselines and Training Parameters We use four baselines: (1) Lemming 8 (Müller et al., 2015) is a context-sensitive system that uses log-linear models to jointly tag and lemmatize the data, and is trained on sentences annotated with both lemmas and POS tags.",
"(2) The hard monotonic attention model (HMAM) 9 (Makarov and Clematide, 2018b) is a neural sequence-tosequence model with a hard attention mechanism that advances through the sequence monotonically.",
"It is trained on word-lemma pairs (without context) 7 Recent efforts to unify the two resources have mostly focused on validating dataset schema (McCarthy et al., 2018) , leaving conflicts in word lemmas unresolved.",
"We estimated (by counting types that are unambiguous in each dataset but have different lemmas across them) that annotation inconsistencies affect up to 1% of types in the languages we used.",
"8 http://cistern.cis.lmu.de/lemming 9 https://github.com/ZurichNLP/ coling2018-neural-transition-basedmorphology with character-level alignments learned in a preprocessing step using an alignment model, and it has proved to be competitive in low resource scenarios.",
"(3) Our naive Baseline outputs the most frequent lemma (or one lemma at random from the options that are equally frequent) for words observed in training.",
"For unseen words it outputs the wordform itself.",
"(4) We also try a baseline data augmentation approach (AE Aug Baseline) inspired by Bergmanis et al.",
"(2017) and Kann and Schütze (2017) , who showed that adding training examples where the network simply learns to auto-encode corpus words can improve morphological inflection results in low-resource settings.",
"The AE Aug Baseline is a variant of Lematus 0-ch which augments the UDT lemmatization examples by auto-encoding the inflected forms of the UM examples (i.e., it just treats them as corpus words).",
"Comparing AE Aug Baseline to Lematus 0-ch augmented with UM lemma-inflection examples tells us whether using the UM lemma information helps more than simply auto-encoding more inflected examples.",
"To train the models we use the default settings for Lemming and the suggested lemmatization parameters for HMAM.",
"We mainly follow the hyperparameters used by Bergmanis and Goldwater (2018) for Lematus; details are in Appendix B.",
"Languages and Training Data We conduct preliminary experiments on five development languages: Estonian, Finnish, Latvian, Polish, and Russian.",
"In our final experiments we also add Bulgarian, Czech, Romanian, Swedish and Turkish.",
"We vary the amount and type of training data (types vs. tokens, UDT only, UM only, or UDT plus up to 10k UM examples), as described in Section 4.",
"To obtain N UM-based training examples, we select the first N unambiguous UM types (with their sentence contexts) from shuffled Wikipedia sentences.",
"For experiments with j > 1 examples per type, we first find all UM types with at least j sentence contexts in Wikipedia and then choose the N distinct types and their j contexts uniformly at random.",
"Evaluation To evaluate models' ability to lemmatize wordforms in their sentence context we follow Bergmanis and Goldwater (2018) and use the full UDT development and test sets.",
"Unlike Bergmanis and Goldwater (2018) who reported token level lemmatization exact match accuracy, we report type-level micro averaged lemmatization ex- act match accuracy.",
"This measure better reflects improvements on unseen words, which tend to be rare but are more important (since a most-frequentlemma baseline does very well on seen words, as shown by Bergmanis and Goldwater (2018) ).",
"We separately report performance on unseen and ambiguous tokens.",
"For a fair comparison across scenarios with different training sets, we count as unseen only words that are not ambiguous and are absent from all training sets/scenarios introduced in Section 4.",
"Due to the small training sets, between 70-90% of dev set types are classed as unseen in each language.",
"We define a type as ambiguous if the empirical entropy over its lemmas is greater than 0.1 in the full original UDT training splits.",
"10 According to this measure, only 1.2-5.3% of dev set types are classed as ambiguous in each language.",
"Significance Testing All systems are trained and tested on ten languages.",
"To test for statistically significant differences between the results of two systems we use a Monte Carlo method: for each set of results (i.e.",
"a set of 10 numerical values) we generate 10000 random samples, where each sample swaps the results of the two systems for each language with a probability of 0.5.",
"We then obtain a p-value as the proportion of samples for which the difference on average was at least as large as the difference observed in our experiments.",
"1k tokens vs. first 1k distinct types of the UDT training sets.",
"Table 2 shows that if only 1k examples are available, using types is clearly better for all systems.",
"Although Lematus does relatively poorly on the token data, it benefits the most from switching to types, putting it on par with HMAM and suggesting is it likely to benefit more from additional type data.",
"Lemming requires token-based data, but does worse than HMAM (a context-free method) in the token-based setting, and we also see no benefit from context in comparing Lematus 20-ch vs Lematus 0-ch.",
"So overall, in this very low-resource scenario with no data augmentation, context does not appear to help.",
"Using UM + Wikipedia Only We now try training only on UM + Wikipedia examples, rather than examples from UDT.",
"We use 1k, 2k or 5k unambiguous types from UM with a single example context from Wikipedia for each.",
"With 5k types we also try adding more example contexts (2, 3, or 5 examples for each type).",
"Figure 1 presents the results (for unseen words only).",
"As with the UDT experiments, there is little difference between Lematus 20-ch and Lematus 0ch in the smallest data setting.",
"However, when the number of training types increases to 5k, the benefits of context begin to show, with Lematus 20-ch yielding a 1.6% statistically significant (p < 0.001) improvement over Lematus 0-ch.",
"The results for increasing the number of examples per type are numerically higher than the one-example case, but the differences are not statistically significant.",
"It is worth noting that the accuracy even with 5k UM types is considerably lower than the accuracy of the model trained on only 1k UDT types (see Table 2 ).",
"We believe this discrepancy is due to the issues of biased/incomplete data noted above.",
"types with contexts from Wikipedia.",
"Table 3 summarizes the results, showing that despite the lower quality of the UM + Wikipedia examples, using them improves results of all systems, and more so with more examples.",
"Improvements are especially strong for unseen types, which constitute more than 70% of types in the dev set.",
"Furthermore, the benefit of the additional UM examples is above and beyond the effect of auto-encoding (AE Aug Baseline) for all systems in all data scenarios.",
"Considering the two context-free models, HMAM does better on the un-augmented 1k UDT data, but (as predicted by our results above) it benefits less from data augmentation than does Lematus 0-ch, so with added data they are statistically equivalent (p = 0.07 on the test set with 10k UM).",
"More importantly, Lematus 20-ch begins to outperform the context-free models with as few as 1k UM + Wikipedia examples, and the difference increases with more examples, eventually reaching over 4% better on the test set than the next best model (Lematus 0-ch) when 10k UM + Wikipedia examples are used (p < 0.001) This indicates that the system can learn useful contextual cues even from unambiguous training examples.",
"Finally, Figure 2 gives a breakdown of Lematus 20-ch dev set accuracy for individual languages, showing that data augmentation helps consistently, although results suggest diminishing returns.",
"Data Augmentation in Medium Resource Setting To examine the extent to which augmented data can help in the medium resource setting of 10k continuous tokens of UDT used in previous work, we follow Bergmanis and Goldwater (2018) and train Lematus 20-ch models for all ten languages using the first 10k tokens of UDT and compare them with models trained on 10k tokens of UDT augmented with 10k UM types.",
"To provide a better comparison of our results, we report both the type and the token level development set accuracy.",
"First of all, Table 4 shows that training on 10k continuous tokens of UDT yields a token level accuracy that is about 8% higher than when using the 1k types of UDT augmented with 10k UM types-the best-performing data augmentation systems (see Table 3 ).",
"Again, we believe this performance gap is due to the issues with the biased/incomplete data noted above.",
"For example, we analyzed errors that were unique to the model trained on the Latvian augmented data and found that 41% of the errors were due to wrongly lemmatized words other than nouns, verbs, and adjectives-the three POSs with available inflection tables in UM.",
"For instance, improperly lemmatized pronouns amounted to 14% of the errors on the Latvian dev set.",
"Table 4 also shows that UM examples with Wikipedia contexts benefit lemmatization not only in the low but also the medium resource setting, yielding statistically significant type and token level accuracy gains over models trained on 10k UDT continuous tokens alone (for both Unseen and All p < 0.001).",
"Conclusion We proposed a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive lemmatization model.",
"The proposed method uses Wikipedia sentences to provide contextualized examples for unambiguous inflection-lemma pairs from UniMorph tables.",
"These examples are noisy and biased, but nevertheless they improve lemmatization accuracy on all ten languages both in low (1k) and medium (10k) resource settings.",
"In particular, we showed that context is helpful, both overall and especially on unseen words-the first work we know of to demonstrate improvements from context in a very low-resource setting.",
"A Lematus Training Lematus is implemented using the Nematus machine translation toolkit 11 .",
"We use default training parameters of Lematus as specified by Bergmanis and Goldwater (2018) except for early stopping with patience (Prechelt, 1998) which we increase to 20.",
"Similar to Bergmanis and Goldwater (2018) we use the first epochs as a burn-in period, after which we validate the current model by its lemmatization exact match accuracy on the first 3k instances of development set and save this model if it performs better than the previous best model.",
"We choose a burn-in period of 20 and validation interval of 5 epochs for models that we train on datasets up to 2k instances and a burn-in period of 10 and validation interval of 2 epochs for others.",
"As we work with considerably smaller datasets than Bergmanis and Goldwater (2018) we reduce the effective model size and increase the rate of convergence by tying the input embeddings of the encoder, the decoder and the softmax output embeddings (Press and Wolf, 2017).",
"B Data Preparation Wikipedia database dumps contain XML structured articles that are formatted using the wikitext markup language.",
"To obtain wordforms in their sentence context we 1) use WikiExtractor 12 to extract plain text from Wikipedia database dumps, followed by scripts from Moses statistical machine translation system 13 (Koehn et al., 2007) to 2) split text into sentences (split-sentences.perl), and 3) extract separate tokens (tokenizer.perl).",
"Finally, we shuffle the extracted sentences to encourage homogeneous type distribution across the entire text.",
"Table 3 .",
"C Result Breakdown by Language"
]
} | {
"paper_header_number": [
"1",
"2.1",
"3",
"5"
],
"paper_header_content": [
"Introduction",
"Data Augmentation",
"Experimental Setup",
"Conclusion"
]
} | GEM-SciDuet-train-4#paper-965#slide-9 | Key question | If ambiguous words enforce the use of context:
Is context still useful in the absence of ambiguous forms? | If ambiguous words enforce the use of context:
Is context still useful in the absence of ambiguous forms? | [] |
GEM-SciDuet-train-4#paper-965#slide-10 | 965 | Data Augmentation for Context-Sensitive Neural Lemmatization Using Inflection Tables and Raw Text | Lemmatization aims to reduce the sparse data problem by relating the inflected forms of a word to its dictionary form. Using context can help, both for unseen and ambiguous words. Yet most context-sensitive approaches require full lemma-annotated sentences for training, which may be scarce or unavailable in lowresource languages. In addition (as shown here), in a low-resource setting, a lemmatizer can learn more from n labeled examples of distinct words (types) than from n (contiguous) labeled tokens, since the latter contain far fewer distinct types. To combine the efficiency of type-based learning with the benefits of context, we propose a way to train a context-sensitive lemmatizer with little or no labeled corpus data, using inflection tables from the UniMorph project and raw text examples from Wikipedia that provide sentence contexts for the unambiguous UniMorph examples. Despite these being unambiguous examples, the model successfully generalizes from them, leading to improved results (both overall, and especially on unseen words) in comparison to a baseline that does not use context. Method Lematus (Bergmanis and Goldwater, 2018) is a neural sequence-to-sequence model with attention 2 Garrette et al. (2013) found the same for POS tagging. 3 Code and data: https://bitbucket.org/ tomsbergmanis/data_augumentation_um_wiki 4 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97
],
"paper_content_text": [
"Introduction Many lemmatizers work on isolated wordforms (Wicentowski, 2002; Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"Lemmatizing in context can improve accuracy on ambiguous and unseen words (Bergmanis and Goldwater, 2018) , but most systems for contextsensitive lemmatization must train on complete sentences labeled with POS and/or morphological tags as well as lemmas, and have only been tested with 20k-300k training tokens (Chrupała et al., 2008; Müller et al., 2015; Chakrabarty et al., 2017) .",
"1 1 The smallest of these corpora contains 20k tokens of Bengali annotated only with lemmas, which Chakrabarty et al.",
"(2017) reported took around two person months to create.",
"Intuitively, though, sentence-annotated data is inefficient for training a lemmatizer, especially in low-resource settings.",
"Training on (say) 1000 word types will provide far more information about a language's morphology than training on 1000 contiguous tokens, where fewer types are represented.",
"As noted above, sentence data can help with ambiguous and unseen words, but we show here that when data is scarce, this effect is small relative to the benefit of seeing more word types.",
"2 Motivated by this result, we propose a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive model.",
"3 We use Lematus (Bergmanis and Goldwater, 2018), a state-of-theart lemmatizer that learns from lemma-annotated words in their N -character contexts.",
"No predictions about surrounding words are used, so fully annotated training sentences are not needed.",
"We exploit this fact by combining two sources of training data: 1k lemma-annotated types (with contexts) from the Universal Dependency Treebank (UDT) v2.2 4 (Nivre et al., 2017) , plus examples obtained by finding unambiguous word-lemma pairs in inflection tables from the Universal Morphology (UM) project 5 and collecting sentence contexts for them from Wikipedia.",
"Although these examples are noisy and biased, we show that they improve lemmatization accuracy in experiments on 10 languages, and that the use of context helps, both overall and especially on unseen words.",
"inspired by the re-inflection model of Kann and Schütze (2016) , which won the 2016 SIGMOR-PHON shared task (Cotterell et al., 2016) .",
"It is built using the Nematus machine translation toolkit, 6 which uses the architecture of Sennrich et al.",
"(2017) : a 2-layer bidirectional GRU encoder and a 2-layer decoder with a conditional GRU (Sennrich et al., 2017) in the first layer and a GRU in the second layer.",
"Lematus takes as input a character sequence representing the wordform in its N -character context, and outputs the characters of the lemma.",
"Special input symbols are used to represent the left and right boundary of the target wordform (<lc>, <rc>) and other word boundaries (<s>).",
"For example, if N = 15, the system trained on Latvian would be expected to produce the characters of the lemma ceļš (meaning road) given input such as: s a k a <s> p aš v a l dī b u <lc> c e ļ u <rc> u n <s> i e l u <s> r eǵ i s t r When N = 0 (Lematus 0-ch), no context is used, making Lematus 0-ch comparable to other systems that do not model context (Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"In our experiments we use both Lematus 0-ch and Lematus 20-ch (20 characters of context), which was the best-performing system reported by Bergmanis and Goldwater (2018).",
"Data Augmentation Our data augmentation method uses UM inflection tables and creates additional training examples by finding Wikipedia sentences that use the inflected wordforms in context, pairing them with their lemma as shown in the inflection table.",
"However, we cannot use all the words in the tables because some of them are ambiguous: for example, Figure 1 shows that the form ceļi could be lemmatized either as ceļš or celis.",
"SG PL SG PL NOM ceļš ceļi celis ceļi GEN ceļa ceļu ceļa ceļu DAT ceļam ceļiem celim ceļiem ACC ceļu ceļus celi ceļus INS ceļu ceļiem celi ceļiem LOC ceļā ceļos celī ceļos VOC ceļ ceļi celi ceļi There are several other issues with this method that could potentially limit its usefulness.",
"First, the UM tables only include verbs, nouns and adjectives, whereas we test the system on UDT data, which includes all parts of speech.",
"Second, by excluding ambiguous forms, we may be restricting the added examples to a non-representative subset of the potential inflections, or the system may simply ignore the context because it isn't needed for these examples.",
"Finally, there are some annotation differences between UM and UDT.",
"7 Despite all of these issues, however, we show below that the added examples and their contexts do actually help.",
"Experimental Setup Baselines and Training Parameters We use four baselines: (1) Lemming 8 (Müller et al., 2015) is a context-sensitive system that uses log-linear models to jointly tag and lemmatize the data, and is trained on sentences annotated with both lemmas and POS tags.",
"(2) The hard monotonic attention model (HMAM) 9 (Makarov and Clematide, 2018b) is a neural sequence-tosequence model with a hard attention mechanism that advances through the sequence monotonically.",
"It is trained on word-lemma pairs (without context) 7 Recent efforts to unify the two resources have mostly focused on validating dataset schema (McCarthy et al., 2018) , leaving conflicts in word lemmas unresolved.",
"We estimated (by counting types that are unambiguous in each dataset but have different lemmas across them) that annotation inconsistencies affect up to 1% of types in the languages we used.",
"8 http://cistern.cis.lmu.de/lemming 9 https://github.com/ZurichNLP/ coling2018-neural-transition-basedmorphology with character-level alignments learned in a preprocessing step using an alignment model, and it has proved to be competitive in low resource scenarios.",
"(3) Our naive Baseline outputs the most frequent lemma (or one lemma at random from the options that are equally frequent) for words observed in training.",
"For unseen words it outputs the wordform itself.",
"(4) We also try a baseline data augmentation approach (AE Aug Baseline) inspired by Bergmanis et al.",
"(2017) and Kann and Schütze (2017) , who showed that adding training examples where the network simply learns to auto-encode corpus words can improve morphological inflection results in low-resource settings.",
"The AE Aug Baseline is a variant of Lematus 0-ch which augments the UDT lemmatization examples by auto-encoding the inflected forms of the UM examples (i.e., it just treats them as corpus words).",
"Comparing AE Aug Baseline to Lematus 0-ch augmented with UM lemma-inflection examples tells us whether using the UM lemma information helps more than simply auto-encoding more inflected examples.",
"To train the models we use the default settings for Lemming and the suggested lemmatization parameters for HMAM.",
"We mainly follow the hyperparameters used by Bergmanis and Goldwater (2018) for Lematus; details are in Appendix B.",
"Languages and Training Data We conduct preliminary experiments on five development languages: Estonian, Finnish, Latvian, Polish, and Russian.",
"In our final experiments we also add Bulgarian, Czech, Romanian, Swedish and Turkish.",
"We vary the amount and type of training data (types vs. tokens, UDT only, UM only, or UDT plus up to 10k UM examples), as described in Section 4.",
"To obtain N UM-based training examples, we select the first N unambiguous UM types (with their sentence contexts) from shuffled Wikipedia sentences.",
"For experiments with j > 1 examples per type, we first find all UM types with at least j sentence contexts in Wikipedia and then choose the N distinct types and their j contexts uniformly at random.",
"Evaluation To evaluate models' ability to lemmatize wordforms in their sentence context we follow Bergmanis and Goldwater (2018) and use the full UDT development and test sets.",
"Unlike Bergmanis and Goldwater (2018) who reported token level lemmatization exact match accuracy, we report type-level micro averaged lemmatization ex- act match accuracy.",
"This measure better reflects improvements on unseen words, which tend to be rare but are more important (since a most-frequentlemma baseline does very well on seen words, as shown by Bergmanis and Goldwater (2018) ).",
"We separately report performance on unseen and ambiguous tokens.",
"For a fair comparison across scenarios with different training sets, we count as unseen only words that are not ambiguous and are absent from all training sets/scenarios introduced in Section 4.",
"Due to the small training sets, between 70-90% of dev set types are classed as unseen in each language.",
"We define a type as ambiguous if the empirical entropy over its lemmas is greater than 0.1 in the full original UDT training splits.",
"10 According to this measure, only 1.2-5.3% of dev set types are classed as ambiguous in each language.",
"Significance Testing All systems are trained and tested on ten languages.",
"To test for statistically significant differences between the results of two systems we use a Monte Carlo method: for each set of results (i.e.",
"a set of 10 numerical values) we generate 10000 random samples, where each sample swaps the results of the two systems for each language with a probability of 0.5.",
"We then obtain a p-value as the proportion of samples for which the difference on average was at least as large as the difference observed in our experiments.",
"1k tokens vs. first 1k distinct types of the UDT training sets.",
"Table 2 shows that if only 1k examples are available, using types is clearly better for all systems.",
"Although Lematus does relatively poorly on the token data, it benefits the most from switching to types, putting it on par with HMAM and suggesting is it likely to benefit more from additional type data.",
"Lemming requires token-based data, but does worse than HMAM (a context-free method) in the token-based setting, and we also see no benefit from context in comparing Lematus 20-ch vs Lematus 0-ch.",
"So overall, in this very low-resource scenario with no data augmentation, context does not appear to help.",
"Using UM + Wikipedia Only We now try training only on UM + Wikipedia examples, rather than examples from UDT.",
"We use 1k, 2k or 5k unambiguous types from UM with a single example context from Wikipedia for each.",
"With 5k types we also try adding more example contexts (2, 3, or 5 examples for each type).",
"Figure 1 presents the results (for unseen words only).",
"As with the UDT experiments, there is little difference between Lematus 20-ch and Lematus 0ch in the smallest data setting.",
"However, when the number of training types increases to 5k, the benefits of context begin to show, with Lematus 20-ch yielding a 1.6% statistically significant (p < 0.001) improvement over Lematus 0-ch.",
"The results for increasing the number of examples per type are numerically higher than the one-example case, but the differences are not statistically significant.",
"It is worth noting that the accuracy even with 5k UM types is considerably lower than the accuracy of the model trained on only 1k UDT types (see Table 2 ).",
"We believe this discrepancy is due to the issues of biased/incomplete data noted above.",
"types with contexts from Wikipedia.",
"Table 3 summarizes the results, showing that despite the lower quality of the UM + Wikipedia examples, using them improves results of all systems, and more so with more examples.",
"Improvements are especially strong for unseen types, which constitute more than 70% of types in the dev set.",
"Furthermore, the benefit of the additional UM examples is above and beyond the effect of auto-encoding (AE Aug Baseline) for all systems in all data scenarios.",
"Considering the two context-free models, HMAM does better on the un-augmented 1k UDT data, but (as predicted by our results above) it benefits less from data augmentation than does Lematus 0-ch, so with added data they are statistically equivalent (p = 0.07 on the test set with 10k UM).",
"More importantly, Lematus 20-ch begins to outperform the context-free models with as few as 1k UM + Wikipedia examples, and the difference increases with more examples, eventually reaching over 4% better on the test set than the next best model (Lematus 0-ch) when 10k UM + Wikipedia examples are used (p < 0.001) This indicates that the system can learn useful contextual cues even from unambiguous training examples.",
"Finally, Figure 2 gives a breakdown of Lematus 20-ch dev set accuracy for individual languages, showing that data augmentation helps consistently, although results suggest diminishing returns.",
"Data Augmentation in Medium Resource Setting To examine the extent to which augmented data can help in the medium resource setting of 10k continuous tokens of UDT used in previous work, we follow Bergmanis and Goldwater (2018) and train Lematus 20-ch models for all ten languages using the first 10k tokens of UDT and compare them with models trained on 10k tokens of UDT augmented with 10k UM types.",
"To provide a better comparison of our results, we report both the type and the token level development set accuracy.",
"First of all, Table 4 shows that training on 10k continuous tokens of UDT yields a token level accuracy that is about 8% higher than when using the 1k types of UDT augmented with 10k UM types-the best-performing data augmentation systems (see Table 3 ).",
"Again, we believe this performance gap is due to the issues with the biased/incomplete data noted above.",
"For example, we analyzed errors that were unique to the model trained on the Latvian augmented data and found that 41% of the errors were due to wrongly lemmatized words other than nouns, verbs, and adjectives-the three POSs with available inflection tables in UM.",
"For instance, improperly lemmatized pronouns amounted to 14% of the errors on the Latvian dev set.",
"Table 4 also shows that UM examples with Wikipedia contexts benefit lemmatization not only in the low but also the medium resource setting, yielding statistically significant type and token level accuracy gains over models trained on 10k UDT continuous tokens alone (for both Unseen and All p < 0.001).",
"Conclusion We proposed a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive lemmatization model.",
"The proposed method uses Wikipedia sentences to provide contextualized examples for unambiguous inflection-lemma pairs from UniMorph tables.",
"These examples are noisy and biased, but nevertheless they improve lemmatization accuracy on all ten languages both in low (1k) and medium (10k) resource settings.",
"In particular, we showed that context is helpful, both overall and especially on unseen words-the first work we know of to demonstrate improvements from context in a very low-resource setting.",
"A Lematus Training Lematus is implemented using the Nematus machine translation toolkit 11 .",
"We use default training parameters of Lematus as specified by Bergmanis and Goldwater (2018) except for early stopping with patience (Prechelt, 1998) which we increase to 20.",
"Similar to Bergmanis and Goldwater (2018) we use the first epochs as a burn-in period, after which we validate the current model by its lemmatization exact match accuracy on the first 3k instances of development set and save this model if it performs better than the previous best model.",
"We choose a burn-in period of 20 and validation interval of 5 epochs for models that we train on datasets up to 2k instances and a burn-in period of 10 and validation interval of 2 epochs for others.",
"As we work with considerably smaller datasets than Bergmanis and Goldwater (2018) we reduce the effective model size and increase the rate of convergence by tying the input embeddings of the encoder, the decoder and the softmax output embeddings (Press and Wolf, 2017).",
"B Data Preparation Wikipedia database dumps contain XML structured articles that are formatted using the wikitext markup language.",
"To obtain wordforms in their sentence context we 1) use WikiExtractor 12 to extract plain text from Wikipedia database dumps, followed by scripts from Moses statistical machine translation system 13 (Koehn et al., 2007) to 2) split text into sentences (split-sentences.perl), and 3) extract separate tokens (tokenizer.perl).",
"Finally, we shuffle the extracted sentences to encourage homogeneous type distribution across the entire text.",
"Table 3 .",
"C Result Breakdown by Language"
]
} | {
"paper_header_number": [
"1",
"2.1",
"3",
"5"
],
"paper_header_content": [
"Introduction",
"Data Augmentation",
"Experimental Setup",
"Conclusion"
]
} | GEM-SciDuet-train-4#paper-965#slide-10 | Experiments | Train: 1k types from universal dependency corpus
UniMorph in Wikipedia contexts
Estonian, Finnish, Latvian, Polish,
Romanian, Russian, Swedish, Turkish
Metric: type level macro average accuracy
Test: on standard splits of universal dependency corpus | Train: 1k types from universal dependency corpus
UniMorph in Wikipedia contexts
Estonian, Finnish, Latvian, Polish,
Romanian, Russian, Swedish, Turkish
Metric: type level macro average accuracy
Test: on standard splits of universal dependency corpus | [] |
GEM-SciDuet-train-4#paper-965#slide-12 | 965 | Data Augmentation for Context-Sensitive Neural Lemmatization Using Inflection Tables and Raw Text | Lemmatization aims to reduce the sparse data problem by relating the inflected forms of a word to its dictionary form. Using context can help, both for unseen and ambiguous words. Yet most context-sensitive approaches require full lemma-annotated sentences for training, which may be scarce or unavailable in lowresource languages. In addition (as shown here), in a low-resource setting, a lemmatizer can learn more from n labeled examples of distinct words (types) than from n (contiguous) labeled tokens, since the latter contain far fewer distinct types. To combine the efficiency of type-based learning with the benefits of context, we propose a way to train a context-sensitive lemmatizer with little or no labeled corpus data, using inflection tables from the UniMorph project and raw text examples from Wikipedia that provide sentence contexts for the unambiguous UniMorph examples. Despite these being unambiguous examples, the model successfully generalizes from them, leading to improved results (both overall, and especially on unseen words) in comparison to a baseline that does not use context. Method Lematus (Bergmanis and Goldwater, 2018) is a neural sequence-to-sequence model with attention 2 Garrette et al. (2013) found the same for POS tagging. 3 Code and data: https://bitbucket.org/ tomsbergmanis/data_augumentation_um_wiki 4 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97
],
"paper_content_text": [
"Introduction Many lemmatizers work on isolated wordforms (Wicentowski, 2002; Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"Lemmatizing in context can improve accuracy on ambiguous and unseen words (Bergmanis and Goldwater, 2018) , but most systems for contextsensitive lemmatization must train on complete sentences labeled with POS and/or morphological tags as well as lemmas, and have only been tested with 20k-300k training tokens (Chrupała et al., 2008; Müller et al., 2015; Chakrabarty et al., 2017) .",
"1 1 The smallest of these corpora contains 20k tokens of Bengali annotated only with lemmas, which Chakrabarty et al.",
"(2017) reported took around two person months to create.",
"Intuitively, though, sentence-annotated data is inefficient for training a lemmatizer, especially in low-resource settings.",
"Training on (say) 1000 word types will provide far more information about a language's morphology than training on 1000 contiguous tokens, where fewer types are represented.",
"As noted above, sentence data can help with ambiguous and unseen words, but we show here that when data is scarce, this effect is small relative to the benefit of seeing more word types.",
"2 Motivated by this result, we propose a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive model.",
"3 We use Lematus (Bergmanis and Goldwater, 2018), a state-of-theart lemmatizer that learns from lemma-annotated words in their N -character contexts.",
"No predictions about surrounding words are used, so fully annotated training sentences are not needed.",
"We exploit this fact by combining two sources of training data: 1k lemma-annotated types (with contexts) from the Universal Dependency Treebank (UDT) v2.2 4 (Nivre et al., 2017) , plus examples obtained by finding unambiguous word-lemma pairs in inflection tables from the Universal Morphology (UM) project 5 and collecting sentence contexts for them from Wikipedia.",
"Although these examples are noisy and biased, we show that they improve lemmatization accuracy in experiments on 10 languages, and that the use of context helps, both overall and especially on unseen words.",
"inspired by the re-inflection model of Kann and Schütze (2016) , which won the 2016 SIGMOR-PHON shared task (Cotterell et al., 2016) .",
"It is built using the Nematus machine translation toolkit, 6 which uses the architecture of Sennrich et al.",
"(2017) : a 2-layer bidirectional GRU encoder and a 2-layer decoder with a conditional GRU (Sennrich et al., 2017) in the first layer and a GRU in the second layer.",
"Lematus takes as input a character sequence representing the wordform in its N -character context, and outputs the characters of the lemma.",
"Special input symbols are used to represent the left and right boundary of the target wordform (<lc>, <rc>) and other word boundaries (<s>).",
"For example, if N = 15, the system trained on Latvian would be expected to produce the characters of the lemma ceļš (meaning road) given input such as: s a k a <s> p aš v a l dī b u <lc> c e ļ u <rc> u n <s> i e l u <s> r eǵ i s t r When N = 0 (Lematus 0-ch), no context is used, making Lematus 0-ch comparable to other systems that do not model context (Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"In our experiments we use both Lematus 0-ch and Lematus 20-ch (20 characters of context), which was the best-performing system reported by Bergmanis and Goldwater (2018).",
"Data Augmentation Our data augmentation method uses UM inflection tables and creates additional training examples by finding Wikipedia sentences that use the inflected wordforms in context, pairing them with their lemma as shown in the inflection table.",
"However, we cannot use all the words in the tables because some of them are ambiguous: for example, Figure 1 shows that the form ceļi could be lemmatized either as ceļš or celis.",
"SG PL SG PL NOM ceļš ceļi celis ceļi GEN ceļa ceļu ceļa ceļu DAT ceļam ceļiem celim ceļiem ACC ceļu ceļus celi ceļus INS ceļu ceļiem celi ceļiem LOC ceļā ceļos celī ceļos VOC ceļ ceļi celi ceļi There are several other issues with this method that could potentially limit its usefulness.",
"First, the UM tables only include verbs, nouns and adjectives, whereas we test the system on UDT data, which includes all parts of speech.",
"Second, by excluding ambiguous forms, we may be restricting the added examples to a non-representative subset of the potential inflections, or the system may simply ignore the context because it isn't needed for these examples.",
"Finally, there are some annotation differences between UM and UDT.",
"7 Despite all of these issues, however, we show below that the added examples and their contexts do actually help.",
"Experimental Setup Baselines and Training Parameters We use four baselines: (1) Lemming 8 (Müller et al., 2015) is a context-sensitive system that uses log-linear models to jointly tag and lemmatize the data, and is trained on sentences annotated with both lemmas and POS tags.",
"(2) The hard monotonic attention model (HMAM) 9 (Makarov and Clematide, 2018b) is a neural sequence-tosequence model with a hard attention mechanism that advances through the sequence monotonically.",
"It is trained on word-lemma pairs (without context) 7 Recent efforts to unify the two resources have mostly focused on validating dataset schema (McCarthy et al., 2018) , leaving conflicts in word lemmas unresolved.",
"We estimated (by counting types that are unambiguous in each dataset but have different lemmas across them) that annotation inconsistencies affect up to 1% of types in the languages we used.",
"8 http://cistern.cis.lmu.de/lemming 9 https://github.com/ZurichNLP/ coling2018-neural-transition-basedmorphology with character-level alignments learned in a preprocessing step using an alignment model, and it has proved to be competitive in low resource scenarios.",
"(3) Our naive Baseline outputs the most frequent lemma (or one lemma at random from the options that are equally frequent) for words observed in training.",
"For unseen words it outputs the wordform itself.",
"(4) We also try a baseline data augmentation approach (AE Aug Baseline) inspired by Bergmanis et al.",
"(2017) and Kann and Schütze (2017) , who showed that adding training examples where the network simply learns to auto-encode corpus words can improve morphological inflection results in low-resource settings.",
"The AE Aug Baseline is a variant of Lematus 0-ch which augments the UDT lemmatization examples by auto-encoding the inflected forms of the UM examples (i.e., it just treats them as corpus words).",
"Comparing AE Aug Baseline to Lematus 0-ch augmented with UM lemma-inflection examples tells us whether using the UM lemma information helps more than simply auto-encoding more inflected examples.",
"To train the models we use the default settings for Lemming and the suggested lemmatization parameters for HMAM.",
"We mainly follow the hyperparameters used by Bergmanis and Goldwater (2018) for Lematus; details are in Appendix B.",
"Languages and Training Data We conduct preliminary experiments on five development languages: Estonian, Finnish, Latvian, Polish, and Russian.",
"In our final experiments we also add Bulgarian, Czech, Romanian, Swedish and Turkish.",
"We vary the amount and type of training data (types vs. tokens, UDT only, UM only, or UDT plus up to 10k UM examples), as described in Section 4.",
"To obtain N UM-based training examples, we select the first N unambiguous UM types (with their sentence contexts) from shuffled Wikipedia sentences.",
"For experiments with j > 1 examples per type, we first find all UM types with at least j sentence contexts in Wikipedia and then choose the N distinct types and their j contexts uniformly at random.",
"Evaluation To evaluate models' ability to lemmatize wordforms in their sentence context we follow Bergmanis and Goldwater (2018) and use the full UDT development and test sets.",
"Unlike Bergmanis and Goldwater (2018) who reported token level lemmatization exact match accuracy, we report type-level micro averaged lemmatization ex- act match accuracy.",
"This measure better reflects improvements on unseen words, which tend to be rare but are more important (since a most-frequentlemma baseline does very well on seen words, as shown by Bergmanis and Goldwater (2018) ).",
"We separately report performance on unseen and ambiguous tokens.",
"For a fair comparison across scenarios with different training sets, we count as unseen only words that are not ambiguous and are absent from all training sets/scenarios introduced in Section 4.",
"Due to the small training sets, between 70-90% of dev set types are classed as unseen in each language.",
"We define a type as ambiguous if the empirical entropy over its lemmas is greater than 0.1 in the full original UDT training splits.",
"10 According to this measure, only 1.2-5.3% of dev set types are classed as ambiguous in each language.",
"Significance Testing All systems are trained and tested on ten languages.",
"To test for statistically significant differences between the results of two systems we use a Monte Carlo method: for each set of results (i.e.",
"a set of 10 numerical values) we generate 10000 random samples, where each sample swaps the results of the two systems for each language with a probability of 0.5.",
"We then obtain a p-value as the proportion of samples for which the difference on average was at least as large as the difference observed in our experiments.",
"1k tokens vs. first 1k distinct types of the UDT training sets.",
"Table 2 shows that if only 1k examples are available, using types is clearly better for all systems.",
"Although Lematus does relatively poorly on the token data, it benefits the most from switching to types, putting it on par with HMAM and suggesting is it likely to benefit more from additional type data.",
"Lemming requires token-based data, but does worse than HMAM (a context-free method) in the token-based setting, and we also see no benefit from context in comparing Lematus 20-ch vs Lematus 0-ch.",
"So overall, in this very low-resource scenario with no data augmentation, context does not appear to help.",
"Using UM + Wikipedia Only We now try training only on UM + Wikipedia examples, rather than examples from UDT.",
"We use 1k, 2k or 5k unambiguous types from UM with a single example context from Wikipedia for each.",
"With 5k types we also try adding more example contexts (2, 3, or 5 examples for each type).",
"Figure 1 presents the results (for unseen words only).",
"As with the UDT experiments, there is little difference between Lematus 20-ch and Lematus 0ch in the smallest data setting.",
"However, when the number of training types increases to 5k, the benefits of context begin to show, with Lematus 20-ch yielding a 1.6% statistically significant (p < 0.001) improvement over Lematus 0-ch.",
"The results for increasing the number of examples per type are numerically higher than the one-example case, but the differences are not statistically significant.",
"It is worth noting that the accuracy even with 5k UM types is considerably lower than the accuracy of the model trained on only 1k UDT types (see Table 2 ).",
"We believe this discrepancy is due to the issues of biased/incomplete data noted above.",
"types with contexts from Wikipedia.",
"Table 3 summarizes the results, showing that despite the lower quality of the UM + Wikipedia examples, using them improves results of all systems, and more so with more examples.",
"Improvements are especially strong for unseen types, which constitute more than 70% of types in the dev set.",
"Furthermore, the benefit of the additional UM examples is above and beyond the effect of auto-encoding (AE Aug Baseline) for all systems in all data scenarios.",
"Considering the two context-free models, HMAM does better on the un-augmented 1k UDT data, but (as predicted by our results above) it benefits less from data augmentation than does Lematus 0-ch, so with added data they are statistically equivalent (p = 0.07 on the test set with 10k UM).",
"More importantly, Lematus 20-ch begins to outperform the context-free models with as few as 1k UM + Wikipedia examples, and the difference increases with more examples, eventually reaching over 4% better on the test set than the next best model (Lematus 0-ch) when 10k UM + Wikipedia examples are used (p < 0.001) This indicates that the system can learn useful contextual cues even from unambiguous training examples.",
"Finally, Figure 2 gives a breakdown of Lematus 20-ch dev set accuracy for individual languages, showing that data augmentation helps consistently, although results suggest diminishing returns.",
"Data Augmentation in Medium Resource Setting To examine the extent to which augmented data can help in the medium resource setting of 10k continuous tokens of UDT used in previous work, we follow Bergmanis and Goldwater (2018) and train Lematus 20-ch models for all ten languages using the first 10k tokens of UDT and compare them with models trained on 10k tokens of UDT augmented with 10k UM types.",
"To provide a better comparison of our results, we report both the type and the token level development set accuracy.",
"First of all, Table 4 shows that training on 10k continuous tokens of UDT yields a token level accuracy that is about 8% higher than when using the 1k types of UDT augmented with 10k UM types-the best-performing data augmentation systems (see Table 3 ).",
"Again, we believe this performance gap is due to the issues with the biased/incomplete data noted above.",
"For example, we analyzed errors that were unique to the model trained on the Latvian augmented data and found that 41% of the errors were due to wrongly lemmatized words other than nouns, verbs, and adjectives-the three POSs with available inflection tables in UM.",
"For instance, improperly lemmatized pronouns amounted to 14% of the errors on the Latvian dev set.",
"Table 4 also shows that UM examples with Wikipedia contexts benefit lemmatization not only in the low but also the medium resource setting, yielding statistically significant type and token level accuracy gains over models trained on 10k UDT continuous tokens alone (for both Unseen and All p < 0.001).",
"Conclusion We proposed a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive lemmatization model.",
"The proposed method uses Wikipedia sentences to provide contextualized examples for unambiguous inflection-lemma pairs from UniMorph tables.",
"These examples are noisy and biased, but nevertheless they improve lemmatization accuracy on all ten languages both in low (1k) and medium (10k) resource settings.",
"In particular, we showed that context is helpful, both overall and especially on unseen words-the first work we know of to demonstrate improvements from context in a very low-resource setting.",
"A Lematus Training Lematus is implemented using the Nematus machine translation toolkit 11 .",
"We use default training parameters of Lematus as specified by Bergmanis and Goldwater (2018) except for early stopping with patience (Prechelt, 1998) which we increase to 20.",
"Similar to Bergmanis and Goldwater (2018) we use the first epochs as a burn-in period, after which we validate the current model by its lemmatization exact match accuracy on the first 3k instances of development set and save this model if it performs better than the previous best model.",
"We choose a burn-in period of 20 and validation interval of 5 epochs for models that we train on datasets up to 2k instances and a burn-in period of 10 and validation interval of 2 epochs for others.",
"As we work with considerably smaller datasets than Bergmanis and Goldwater (2018) we reduce the effective model size and increase the rate of convergence by tying the input embeddings of the encoder, the decoder and the softmax output embeddings (Press and Wolf, 2017).",
"B Data Preparation Wikipedia database dumps contain XML structured articles that are formatted using the wikitext markup language.",
"To obtain wordforms in their sentence context we 1) use WikiExtractor 12 to extract plain text from Wikipedia database dumps, followed by scripts from Moses statistical machine translation system 13 (Koehn et al., 2007) to 2) split text into sentences (split-sentences.perl), and 3) extract separate tokens (tokenizer.perl).",
"Finally, we shuffle the extracted sentences to encourage homogeneous type distribution across the entire text.",
"Table 3 .",
"C Result Breakdown by Language"
]
} | {
"paper_header_number": [
"1",
"2.1",
"3",
"5"
],
"paper_header_content": [
"Introduction",
"Data Augmentation",
"Experimental Setup",
"Conclusion"
]
} | GEM-SciDuet-train-4#paper-965#slide-12 | Does model learn from context | context vs no context | context vs no context | [] |
GEM-SciDuet-train-4#paper-965#slide-13 | 965 | Data Augmentation for Context-Sensitive Neural Lemmatization Using Inflection Tables and Raw Text | Lemmatization aims to reduce the sparse data problem by relating the inflected forms of a word to its dictionary form. Using context can help, both for unseen and ambiguous words. Yet most context-sensitive approaches require full lemma-annotated sentences for training, which may be scarce or unavailable in lowresource languages. In addition (as shown here), in a low-resource setting, a lemmatizer can learn more from n labeled examples of distinct words (types) than from n (contiguous) labeled tokens, since the latter contain far fewer distinct types. To combine the efficiency of type-based learning with the benefits of context, we propose a way to train a context-sensitive lemmatizer with little or no labeled corpus data, using inflection tables from the UniMorph project and raw text examples from Wikipedia that provide sentence contexts for the unambiguous UniMorph examples. Despite these being unambiguous examples, the model successfully generalizes from them, leading to improved results (both overall, and especially on unseen words) in comparison to a baseline that does not use context. Method Lematus (Bergmanis and Goldwater, 2018) is a neural sequence-to-sequence model with attention 2 Garrette et al. (2013) found the same for POS tagging. 3 Code and data: https://bitbucket.org/ tomsbergmanis/data_augumentation_um_wiki 4 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97
],
"paper_content_text": [
"Introduction Many lemmatizers work on isolated wordforms (Wicentowski, 2002; Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"Lemmatizing in context can improve accuracy on ambiguous and unseen words (Bergmanis and Goldwater, 2018) , but most systems for contextsensitive lemmatization must train on complete sentences labeled with POS and/or morphological tags as well as lemmas, and have only been tested with 20k-300k training tokens (Chrupała et al., 2008; Müller et al., 2015; Chakrabarty et al., 2017) .",
"1 1 The smallest of these corpora contains 20k tokens of Bengali annotated only with lemmas, which Chakrabarty et al.",
"(2017) reported took around two person months to create.",
"Intuitively, though, sentence-annotated data is inefficient for training a lemmatizer, especially in low-resource settings.",
"Training on (say) 1000 word types will provide far more information about a language's morphology than training on 1000 contiguous tokens, where fewer types are represented.",
"As noted above, sentence data can help with ambiguous and unseen words, but we show here that when data is scarce, this effect is small relative to the benefit of seeing more word types.",
"2 Motivated by this result, we propose a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive model.",
"3 We use Lematus (Bergmanis and Goldwater, 2018), a state-of-theart lemmatizer that learns from lemma-annotated words in their N -character contexts.",
"No predictions about surrounding words are used, so fully annotated training sentences are not needed.",
"We exploit this fact by combining two sources of training data: 1k lemma-annotated types (with contexts) from the Universal Dependency Treebank (UDT) v2.2 4 (Nivre et al., 2017) , plus examples obtained by finding unambiguous word-lemma pairs in inflection tables from the Universal Morphology (UM) project 5 and collecting sentence contexts for them from Wikipedia.",
"Although these examples are noisy and biased, we show that they improve lemmatization accuracy in experiments on 10 languages, and that the use of context helps, both overall and especially on unseen words.",
"inspired by the re-inflection model of Kann and Schütze (2016) , which won the 2016 SIGMOR-PHON shared task (Cotterell et al., 2016) .",
"It is built using the Nematus machine translation toolkit, 6 which uses the architecture of Sennrich et al.",
"(2017) : a 2-layer bidirectional GRU encoder and a 2-layer decoder with a conditional GRU (Sennrich et al., 2017) in the first layer and a GRU in the second layer.",
"Lematus takes as input a character sequence representing the wordform in its N -character context, and outputs the characters of the lemma.",
"Special input symbols are used to represent the left and right boundary of the target wordform (<lc>, <rc>) and other word boundaries (<s>).",
"For example, if N = 15, the system trained on Latvian would be expected to produce the characters of the lemma ceļš (meaning road) given input such as: s a k a <s> p aš v a l dī b u <lc> c e ļ u <rc> u n <s> i e l u <s> r eǵ i s t r When N = 0 (Lematus 0-ch), no context is used, making Lematus 0-ch comparable to other systems that do not model context (Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"In our experiments we use both Lematus 0-ch and Lematus 20-ch (20 characters of context), which was the best-performing system reported by Bergmanis and Goldwater (2018).",
"Data Augmentation Our data augmentation method uses UM inflection tables and creates additional training examples by finding Wikipedia sentences that use the inflected wordforms in context, pairing them with their lemma as shown in the inflection table.",
"However, we cannot use all the words in the tables because some of them are ambiguous: for example, Figure 1 shows that the form ceļi could be lemmatized either as ceļš or celis.",
"SG PL SG PL NOM ceļš ceļi celis ceļi GEN ceļa ceļu ceļa ceļu DAT ceļam ceļiem celim ceļiem ACC ceļu ceļus celi ceļus INS ceļu ceļiem celi ceļiem LOC ceļā ceļos celī ceļos VOC ceļ ceļi celi ceļi There are several other issues with this method that could potentially limit its usefulness.",
"First, the UM tables only include verbs, nouns and adjectives, whereas we test the system on UDT data, which includes all parts of speech.",
"Second, by excluding ambiguous forms, we may be restricting the added examples to a non-representative subset of the potential inflections, or the system may simply ignore the context because it isn't needed for these examples.",
"Finally, there are some annotation differences between UM and UDT.",
"7 Despite all of these issues, however, we show below that the added examples and their contexts do actually help.",
"Experimental Setup Baselines and Training Parameters We use four baselines: (1) Lemming 8 (Müller et al., 2015) is a context-sensitive system that uses log-linear models to jointly tag and lemmatize the data, and is trained on sentences annotated with both lemmas and POS tags.",
"(2) The hard monotonic attention model (HMAM) 9 (Makarov and Clematide, 2018b) is a neural sequence-tosequence model with a hard attention mechanism that advances through the sequence monotonically.",
"It is trained on word-lemma pairs (without context) 7 Recent efforts to unify the two resources have mostly focused on validating dataset schema (McCarthy et al., 2018) , leaving conflicts in word lemmas unresolved.",
"We estimated (by counting types that are unambiguous in each dataset but have different lemmas across them) that annotation inconsistencies affect up to 1% of types in the languages we used.",
"8 http://cistern.cis.lmu.de/lemming 9 https://github.com/ZurichNLP/ coling2018-neural-transition-basedmorphology with character-level alignments learned in a preprocessing step using an alignment model, and it has proved to be competitive in low resource scenarios.",
"(3) Our naive Baseline outputs the most frequent lemma (or one lemma at random from the options that are equally frequent) for words observed in training.",
"For unseen words it outputs the wordform itself.",
"(4) We also try a baseline data augmentation approach (AE Aug Baseline) inspired by Bergmanis et al.",
"(2017) and Kann and Schütze (2017) , who showed that adding training examples where the network simply learns to auto-encode corpus words can improve morphological inflection results in low-resource settings.",
"The AE Aug Baseline is a variant of Lematus 0-ch which augments the UDT lemmatization examples by auto-encoding the inflected forms of the UM examples (i.e., it just treats them as corpus words).",
"Comparing AE Aug Baseline to Lematus 0-ch augmented with UM lemma-inflection examples tells us whether using the UM lemma information helps more than simply auto-encoding more inflected examples.",
"To train the models we use the default settings for Lemming and the suggested lemmatization parameters for HMAM.",
"We mainly follow the hyperparameters used by Bergmanis and Goldwater (2018) for Lematus; details are in Appendix B.",
"Languages and Training Data We conduct preliminary experiments on five development languages: Estonian, Finnish, Latvian, Polish, and Russian.",
"In our final experiments we also add Bulgarian, Czech, Romanian, Swedish and Turkish.",
"We vary the amount and type of training data (types vs. tokens, UDT only, UM only, or UDT plus up to 10k UM examples), as described in Section 4.",
"To obtain N UM-based training examples, we select the first N unambiguous UM types (with their sentence contexts) from shuffled Wikipedia sentences.",
"For experiments with j > 1 examples per type, we first find all UM types with at least j sentence contexts in Wikipedia and then choose the N distinct types and their j contexts uniformly at random.",
"Evaluation To evaluate models' ability to lemmatize wordforms in their sentence context we follow Bergmanis and Goldwater (2018) and use the full UDT development and test sets.",
"Unlike Bergmanis and Goldwater (2018) who reported token level lemmatization exact match accuracy, we report type-level micro averaged lemmatization ex- act match accuracy.",
"This measure better reflects improvements on unseen words, which tend to be rare but are more important (since a most-frequentlemma baseline does very well on seen words, as shown by Bergmanis and Goldwater (2018) ).",
"We separately report performance on unseen and ambiguous tokens.",
"For a fair comparison across scenarios with different training sets, we count as unseen only words that are not ambiguous and are absent from all training sets/scenarios introduced in Section 4.",
"Due to the small training sets, between 70-90% of dev set types are classed as unseen in each language.",
"We define a type as ambiguous if the empirical entropy over its lemmas is greater than 0.1 in the full original UDT training splits.",
"10 According to this measure, only 1.2-5.3% of dev set types are classed as ambiguous in each language.",
"Significance Testing All systems are trained and tested on ten languages.",
"To test for statistically significant differences between the results of two systems we use a Monte Carlo method: for each set of results (i.e.",
"a set of 10 numerical values) we generate 10000 random samples, where each sample swaps the results of the two systems for each language with a probability of 0.5.",
"We then obtain a p-value as the proportion of samples for which the difference on average was at least as large as the difference observed in our experiments.",
"1k tokens vs. first 1k distinct types of the UDT training sets.",
"Table 2 shows that if only 1k examples are available, using types is clearly better for all systems.",
"Although Lematus does relatively poorly on the token data, it benefits the most from switching to types, putting it on par with HMAM and suggesting is it likely to benefit more from additional type data.",
"Lemming requires token-based data, but does worse than HMAM (a context-free method) in the token-based setting, and we also see no benefit from context in comparing Lematus 20-ch vs Lematus 0-ch.",
"So overall, in this very low-resource scenario with no data augmentation, context does not appear to help.",
"Using UM + Wikipedia Only We now try training only on UM + Wikipedia examples, rather than examples from UDT.",
"We use 1k, 2k or 5k unambiguous types from UM with a single example context from Wikipedia for each.",
"With 5k types we also try adding more example contexts (2, 3, or 5 examples for each type).",
"Figure 1 presents the results (for unseen words only).",
"As with the UDT experiments, there is little difference between Lematus 20-ch and Lematus 0ch in the smallest data setting.",
"However, when the number of training types increases to 5k, the benefits of context begin to show, with Lematus 20-ch yielding a 1.6% statistically significant (p < 0.001) improvement over Lematus 0-ch.",
"The results for increasing the number of examples per type are numerically higher than the one-example case, but the differences are not statistically significant.",
"It is worth noting that the accuracy even with 5k UM types is considerably lower than the accuracy of the model trained on only 1k UDT types (see Table 2 ).",
"We believe this discrepancy is due to the issues of biased/incomplete data noted above.",
"types with contexts from Wikipedia.",
"Table 3 summarizes the results, showing that despite the lower quality of the UM + Wikipedia examples, using them improves results of all systems, and more so with more examples.",
"Improvements are especially strong for unseen types, which constitute more than 70% of types in the dev set.",
"Furthermore, the benefit of the additional UM examples is above and beyond the effect of auto-encoding (AE Aug Baseline) for all systems in all data scenarios.",
"Considering the two context-free models, HMAM does better on the un-augmented 1k UDT data, but (as predicted by our results above) it benefits less from data augmentation than does Lematus 0-ch, so with added data they are statistically equivalent (p = 0.07 on the test set with 10k UM).",
"More importantly, Lematus 20-ch begins to outperform the context-free models with as few as 1k UM + Wikipedia examples, and the difference increases with more examples, eventually reaching over 4% better on the test set than the next best model (Lematus 0-ch) when 10k UM + Wikipedia examples are used (p < 0.001) This indicates that the system can learn useful contextual cues even from unambiguous training examples.",
"Finally, Figure 2 gives a breakdown of Lematus 20-ch dev set accuracy for individual languages, showing that data augmentation helps consistently, although results suggest diminishing returns.",
"Data Augmentation in Medium Resource Setting To examine the extent to which augmented data can help in the medium resource setting of 10k continuous tokens of UDT used in previous work, we follow Bergmanis and Goldwater (2018) and train Lematus 20-ch models for all ten languages using the first 10k tokens of UDT and compare them with models trained on 10k tokens of UDT augmented with 10k UM types.",
"To provide a better comparison of our results, we report both the type and the token level development set accuracy.",
"First of all, Table 4 shows that training on 10k continuous tokens of UDT yields a token level accuracy that is about 8% higher than when using the 1k types of UDT augmented with 10k UM types-the best-performing data augmentation systems (see Table 3 ).",
"Again, we believe this performance gap is due to the issues with the biased/incomplete data noted above.",
"For example, we analyzed errors that were unique to the model trained on the Latvian augmented data and found that 41% of the errors were due to wrongly lemmatized words other than nouns, verbs, and adjectives-the three POSs with available inflection tables in UM.",
"For instance, improperly lemmatized pronouns amounted to 14% of the errors on the Latvian dev set.",
"Table 4 also shows that UM examples with Wikipedia contexts benefit lemmatization not only in the low but also the medium resource setting, yielding statistically significant type and token level accuracy gains over models trained on 10k UDT continuous tokens alone (for both Unseen and All p < 0.001).",
"Conclusion We proposed a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive lemmatization model.",
"The proposed method uses Wikipedia sentences to provide contextualized examples for unambiguous inflection-lemma pairs from UniMorph tables.",
"These examples are noisy and biased, but nevertheless they improve lemmatization accuracy on all ten languages both in low (1k) and medium (10k) resource settings.",
"In particular, we showed that context is helpful, both overall and especially on unseen words-the first work we know of to demonstrate improvements from context in a very low-resource setting.",
"A Lematus Training Lematus is implemented using the Nematus machine translation toolkit 11 .",
"We use default training parameters of Lematus as specified by Bergmanis and Goldwater (2018) except for early stopping with patience (Prechelt, 1998) which we increase to 20.",
"Similar to Bergmanis and Goldwater (2018) we use the first epochs as a burn-in period, after which we validate the current model by its lemmatization exact match accuracy on the first 3k instances of development set and save this model if it performs better than the previous best model.",
"We choose a burn-in period of 20 and validation interval of 5 epochs for models that we train on datasets up to 2k instances and a burn-in period of 10 and validation interval of 2 epochs for others.",
"As we work with considerably smaller datasets than Bergmanis and Goldwater (2018) we reduce the effective model size and increase the rate of convergence by tying the input embeddings of the encoder, the decoder and the softmax output embeddings (Press and Wolf, 2017).",
"B Data Preparation Wikipedia database dumps contain XML structured articles that are formatted using the wikitext markup language.",
"To obtain wordforms in their sentence context we 1) use WikiExtractor 12 to extract plain text from Wikipedia database dumps, followed by scripts from Moses statistical machine translation system 13 (Koehn et al., 2007) to 2) split text into sentences (split-sentences.perl), and 3) extract separate tokens (tokenizer.perl).",
"Finally, we shuffle the extracted sentences to encourage homogeneous type distribution across the entire text.",
"Table 3 .",
"C Result Breakdown by Language"
]
} | {
"paper_header_number": [
"1",
"2.1",
"3",
"5"
],
"paper_header_content": [
"Introduction",
"Data Augmentation",
"Experimental Setup",
"Conclusion"
]
} | GEM-SciDuet-train-4#paper-965#slide-13 | Afix ambiguity wuger | Lemma depends on context:
A if wuger is adjective then lemma could be wug
B if wuger is noun then lemma could be wuger | Lemma depends on context:
A if wuger is adjective then lemma could be wug
B if wuger is noun then lemma could be wuger | [] |
GEM-SciDuet-train-4#paper-965#slide-14 | 965 | Data Augmentation for Context-Sensitive Neural Lemmatization Using Inflection Tables and Raw Text | Lemmatization aims to reduce the sparse data problem by relating the inflected forms of a word to its dictionary form. Using context can help, both for unseen and ambiguous words. Yet most context-sensitive approaches require full lemma-annotated sentences for training, which may be scarce or unavailable in lowresource languages. In addition (as shown here), in a low-resource setting, a lemmatizer can learn more from n labeled examples of distinct words (types) than from n (contiguous) labeled tokens, since the latter contain far fewer distinct types. To combine the efficiency of type-based learning with the benefits of context, we propose a way to train a context-sensitive lemmatizer with little or no labeled corpus data, using inflection tables from the UniMorph project and raw text examples from Wikipedia that provide sentence contexts for the unambiguous UniMorph examples. Despite these being unambiguous examples, the model successfully generalizes from them, leading to improved results (both overall, and especially on unseen words) in comparison to a baseline that does not use context. Method Lematus (Bergmanis and Goldwater, 2018) is a neural sequence-to-sequence model with attention 2 Garrette et al. (2013) found the same for POS tagging. 3 Code and data: https://bitbucket.org/ tomsbergmanis/data_augumentation_um_wiki 4 | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97
],
"paper_content_text": [
"Introduction Many lemmatizers work on isolated wordforms (Wicentowski, 2002; Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"Lemmatizing in context can improve accuracy on ambiguous and unseen words (Bergmanis and Goldwater, 2018) , but most systems for contextsensitive lemmatization must train on complete sentences labeled with POS and/or morphological tags as well as lemmas, and have only been tested with 20k-300k training tokens (Chrupała et al., 2008; Müller et al., 2015; Chakrabarty et al., 2017) .",
"1 1 The smallest of these corpora contains 20k tokens of Bengali annotated only with lemmas, which Chakrabarty et al.",
"(2017) reported took around two person months to create.",
"Intuitively, though, sentence-annotated data is inefficient for training a lemmatizer, especially in low-resource settings.",
"Training on (say) 1000 word types will provide far more information about a language's morphology than training on 1000 contiguous tokens, where fewer types are represented.",
"As noted above, sentence data can help with ambiguous and unseen words, but we show here that when data is scarce, this effect is small relative to the benefit of seeing more word types.",
"2 Motivated by this result, we propose a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive model.",
"3 We use Lematus (Bergmanis and Goldwater, 2018), a state-of-theart lemmatizer that learns from lemma-annotated words in their N -character contexts.",
"No predictions about surrounding words are used, so fully annotated training sentences are not needed.",
"We exploit this fact by combining two sources of training data: 1k lemma-annotated types (with contexts) from the Universal Dependency Treebank (UDT) v2.2 4 (Nivre et al., 2017) , plus examples obtained by finding unambiguous word-lemma pairs in inflection tables from the Universal Morphology (UM) project 5 and collecting sentence contexts for them from Wikipedia.",
"Although these examples are noisy and biased, we show that they improve lemmatization accuracy in experiments on 10 languages, and that the use of context helps, both overall and especially on unseen words.",
"inspired by the re-inflection model of Kann and Schütze (2016) , which won the 2016 SIGMOR-PHON shared task (Cotterell et al., 2016) .",
"It is built using the Nematus machine translation toolkit, 6 which uses the architecture of Sennrich et al.",
"(2017) : a 2-layer bidirectional GRU encoder and a 2-layer decoder with a conditional GRU (Sennrich et al., 2017) in the first layer and a GRU in the second layer.",
"Lematus takes as input a character sequence representing the wordform in its N -character context, and outputs the characters of the lemma.",
"Special input symbols are used to represent the left and right boundary of the target wordform (<lc>, <rc>) and other word boundaries (<s>).",
"For example, if N = 15, the system trained on Latvian would be expected to produce the characters of the lemma ceļš (meaning road) given input such as: s a k a <s> p aš v a l dī b u <lc> c e ļ u <rc> u n <s> i e l u <s> r eǵ i s t r When N = 0 (Lematus 0-ch), no context is used, making Lematus 0-ch comparable to other systems that do not model context (Dreyer et al., 2008; Rastogi et al., 2016; Makarov and Clematide, 2018b,a) .",
"In our experiments we use both Lematus 0-ch and Lematus 20-ch (20 characters of context), which was the best-performing system reported by Bergmanis and Goldwater (2018).",
"Data Augmentation Our data augmentation method uses UM inflection tables and creates additional training examples by finding Wikipedia sentences that use the inflected wordforms in context, pairing them with their lemma as shown in the inflection table.",
"However, we cannot use all the words in the tables because some of them are ambiguous: for example, Figure 1 shows that the form ceļi could be lemmatized either as ceļš or celis.",
"SG PL SG PL NOM ceļš ceļi celis ceļi GEN ceļa ceļu ceļa ceļu DAT ceļam ceļiem celim ceļiem ACC ceļu ceļus celi ceļus INS ceļu ceļiem celi ceļiem LOC ceļā ceļos celī ceļos VOC ceļ ceļi celi ceļi There are several other issues with this method that could potentially limit its usefulness.",
"First, the UM tables only include verbs, nouns and adjectives, whereas we test the system on UDT data, which includes all parts of speech.",
"Second, by excluding ambiguous forms, we may be restricting the added examples to a non-representative subset of the potential inflections, or the system may simply ignore the context because it isn't needed for these examples.",
"Finally, there are some annotation differences between UM and UDT.",
"7 Despite all of these issues, however, we show below that the added examples and their contexts do actually help.",
"Experimental Setup Baselines and Training Parameters We use four baselines: (1) Lemming 8 (Müller et al., 2015) is a context-sensitive system that uses log-linear models to jointly tag and lemmatize the data, and is trained on sentences annotated with both lemmas and POS tags.",
"(2) The hard monotonic attention model (HMAM) 9 (Makarov and Clematide, 2018b) is a neural sequence-tosequence model with a hard attention mechanism that advances through the sequence monotonically.",
"It is trained on word-lemma pairs (without context) 7 Recent efforts to unify the two resources have mostly focused on validating dataset schema (McCarthy et al., 2018) , leaving conflicts in word lemmas unresolved.",
"We estimated (by counting types that are unambiguous in each dataset but have different lemmas across them) that annotation inconsistencies affect up to 1% of types in the languages we used.",
"8 http://cistern.cis.lmu.de/lemming 9 https://github.com/ZurichNLP/ coling2018-neural-transition-basedmorphology with character-level alignments learned in a preprocessing step using an alignment model, and it has proved to be competitive in low resource scenarios.",
"(3) Our naive Baseline outputs the most frequent lemma (or one lemma at random from the options that are equally frequent) for words observed in training.",
"For unseen words it outputs the wordform itself.",
"(4) We also try a baseline data augmentation approach (AE Aug Baseline) inspired by Bergmanis et al.",
"(2017) and Kann and Schütze (2017) , who showed that adding training examples where the network simply learns to auto-encode corpus words can improve morphological inflection results in low-resource settings.",
"The AE Aug Baseline is a variant of Lematus 0-ch which augments the UDT lemmatization examples by auto-encoding the inflected forms of the UM examples (i.e., it just treats them as corpus words).",
"Comparing AE Aug Baseline to Lematus 0-ch augmented with UM lemma-inflection examples tells us whether using the UM lemma information helps more than simply auto-encoding more inflected examples.",
"To train the models we use the default settings for Lemming and the suggested lemmatization parameters for HMAM.",
"We mainly follow the hyperparameters used by Bergmanis and Goldwater (2018) for Lematus; details are in Appendix B.",
"Languages and Training Data We conduct preliminary experiments on five development languages: Estonian, Finnish, Latvian, Polish, and Russian.",
"In our final experiments we also add Bulgarian, Czech, Romanian, Swedish and Turkish.",
"We vary the amount and type of training data (types vs. tokens, UDT only, UM only, or UDT plus up to 10k UM examples), as described in Section 4.",
"To obtain N UM-based training examples, we select the first N unambiguous UM types (with their sentence contexts) from shuffled Wikipedia sentences.",
"For experiments with j > 1 examples per type, we first find all UM types with at least j sentence contexts in Wikipedia and then choose the N distinct types and their j contexts uniformly at random.",
"Evaluation To evaluate models' ability to lemmatize wordforms in their sentence context we follow Bergmanis and Goldwater (2018) and use the full UDT development and test sets.",
"Unlike Bergmanis and Goldwater (2018) who reported token level lemmatization exact match accuracy, we report type-level micro averaged lemmatization ex- act match accuracy.",
"This measure better reflects improvements on unseen words, which tend to be rare but are more important (since a most-frequentlemma baseline does very well on seen words, as shown by Bergmanis and Goldwater (2018) ).",
"We separately report performance on unseen and ambiguous tokens.",
"For a fair comparison across scenarios with different training sets, we count as unseen only words that are not ambiguous and are absent from all training sets/scenarios introduced in Section 4.",
"Due to the small training sets, between 70-90% of dev set types are classed as unseen in each language.",
"We define a type as ambiguous if the empirical entropy over its lemmas is greater than 0.1 in the full original UDT training splits.",
"10 According to this measure, only 1.2-5.3% of dev set types are classed as ambiguous in each language.",
"Significance Testing All systems are trained and tested on ten languages.",
"To test for statistically significant differences between the results of two systems we use a Monte Carlo method: for each set of results (i.e.",
"a set of 10 numerical values) we generate 10000 random samples, where each sample swaps the results of the two systems for each language with a probability of 0.5.",
"We then obtain a p-value as the proportion of samples for which the difference on average was at least as large as the difference observed in our experiments.",
"1k tokens vs. first 1k distinct types of the UDT training sets.",
"Table 2 shows that if only 1k examples are available, using types is clearly better for all systems.",
"Although Lematus does relatively poorly on the token data, it benefits the most from switching to types, putting it on par with HMAM and suggesting is it likely to benefit more from additional type data.",
"Lemming requires token-based data, but does worse than HMAM (a context-free method) in the token-based setting, and we also see no benefit from context in comparing Lematus 20-ch vs Lematus 0-ch.",
"So overall, in this very low-resource scenario with no data augmentation, context does not appear to help.",
"Using UM + Wikipedia Only We now try training only on UM + Wikipedia examples, rather than examples from UDT.",
"We use 1k, 2k or 5k unambiguous types from UM with a single example context from Wikipedia for each.",
"With 5k types we also try adding more example contexts (2, 3, or 5 examples for each type).",
"Figure 1 presents the results (for unseen words only).",
"As with the UDT experiments, there is little difference between Lematus 20-ch and Lematus 0ch in the smallest data setting.",
"However, when the number of training types increases to 5k, the benefits of context begin to show, with Lematus 20-ch yielding a 1.6% statistically significant (p < 0.001) improvement over Lematus 0-ch.",
"The results for increasing the number of examples per type are numerically higher than the one-example case, but the differences are not statistically significant.",
"It is worth noting that the accuracy even with 5k UM types is considerably lower than the accuracy of the model trained on only 1k UDT types (see Table 2 ).",
"We believe this discrepancy is due to the issues of biased/incomplete data noted above.",
"types with contexts from Wikipedia.",
"Table 3 summarizes the results, showing that despite the lower quality of the UM + Wikipedia examples, using them improves results of all systems, and more so with more examples.",
"Improvements are especially strong for unseen types, which constitute more than 70% of types in the dev set.",
"Furthermore, the benefit of the additional UM examples is above and beyond the effect of auto-encoding (AE Aug Baseline) for all systems in all data scenarios.",
"Considering the two context-free models, HMAM does better on the un-augmented 1k UDT data, but (as predicted by our results above) it benefits less from data augmentation than does Lematus 0-ch, so with added data they are statistically equivalent (p = 0.07 on the test set with 10k UM).",
"More importantly, Lematus 20-ch begins to outperform the context-free models with as few as 1k UM + Wikipedia examples, and the difference increases with more examples, eventually reaching over 4% better on the test set than the next best model (Lematus 0-ch) when 10k UM + Wikipedia examples are used (p < 0.001) This indicates that the system can learn useful contextual cues even from unambiguous training examples.",
"Finally, Figure 2 gives a breakdown of Lematus 20-ch dev set accuracy for individual languages, showing that data augmentation helps consistently, although results suggest diminishing returns.",
"Data Augmentation in Medium Resource Setting To examine the extent to which augmented data can help in the medium resource setting of 10k continuous tokens of UDT used in previous work, we follow Bergmanis and Goldwater (2018) and train Lematus 20-ch models for all ten languages using the first 10k tokens of UDT and compare them with models trained on 10k tokens of UDT augmented with 10k UM types.",
"To provide a better comparison of our results, we report both the type and the token level development set accuracy.",
"First of all, Table 4 shows that training on 10k continuous tokens of UDT yields a token level accuracy that is about 8% higher than when using the 1k types of UDT augmented with 10k UM types-the best-performing data augmentation systems (see Table 3 ).",
"Again, we believe this performance gap is due to the issues with the biased/incomplete data noted above.",
"For example, we analyzed errors that were unique to the model trained on the Latvian augmented data and found that 41% of the errors were due to wrongly lemmatized words other than nouns, verbs, and adjectives-the three POSs with available inflection tables in UM.",
"For instance, improperly lemmatized pronouns amounted to 14% of the errors on the Latvian dev set.",
"Table 4 also shows that UM examples with Wikipedia contexts benefit lemmatization not only in the low but also the medium resource setting, yielding statistically significant type and token level accuracy gains over models trained on 10k UDT continuous tokens alone (for both Unseen and All p < 0.001).",
"Conclusion We proposed a training data augmentation method that combines the efficiency of type-based learning and the expressive power of a context-sensitive lemmatization model.",
"The proposed method uses Wikipedia sentences to provide contextualized examples for unambiguous inflection-lemma pairs from UniMorph tables.",
"These examples are noisy and biased, but nevertheless they improve lemmatization accuracy on all ten languages both in low (1k) and medium (10k) resource settings.",
"In particular, we showed that context is helpful, both overall and especially on unseen words-the first work we know of to demonstrate improvements from context in a very low-resource setting.",
"A Lematus Training Lematus is implemented using the Nematus machine translation toolkit 11 .",
"We use default training parameters of Lematus as specified by Bergmanis and Goldwater (2018) except for early stopping with patience (Prechelt, 1998) which we increase to 20.",
"Similar to Bergmanis and Goldwater (2018) we use the first epochs as a burn-in period, after which we validate the current model by its lemmatization exact match accuracy on the first 3k instances of development set and save this model if it performs better than the previous best model.",
"We choose a burn-in period of 20 and validation interval of 5 epochs for models that we train on datasets up to 2k instances and a burn-in period of 10 and validation interval of 2 epochs for others.",
"As we work with considerably smaller datasets than Bergmanis and Goldwater (2018) we reduce the effective model size and increase the rate of convergence by tying the input embeddings of the encoder, the decoder and the softmax output embeddings (Press and Wolf, 2017).",
"B Data Preparation Wikipedia database dumps contain XML structured articles that are formatted using the wikitext markup language.",
"To obtain wordforms in their sentence context we 1) use WikiExtractor 12 to extract plain text from Wikipedia database dumps, followed by scripts from Moses statistical machine translation system 13 (Koehn et al., 2007) to 2) split text into sentences (split-sentences.perl), and 3) extract separate tokens (tokenizer.perl).",
"Finally, we shuffle the extracted sentences to encourage homogeneous type distribution across the entire text.",
"Table 3 .",
"C Result Breakdown by Language"
]
} | {
"paper_header_number": [
"1",
"2.1",
"3",
"5"
],
"paper_header_content": [
"Introduction",
"Data Augmentation",
"Experimental Setup",
"Conclusion"
]
} | GEM-SciDuet-train-4#paper-965#slide-14 | Takeaways conclusions | Despite biased data and divergent lemmatization standards
Type based data augmentation helps
Even without the ambiguous types that enforce the use of context
Model use context to disambiguate affixes of unseen words | Despite biased data and divergent lemmatization standards
Type based data augmentation helps
Even without the ambiguous types that enforce the use of context
Model use context to disambiguate affixes of unseen words | [] |
GEM-SciDuet-train-5#paper-966#slide-0 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-0 | What is Automated Essay Scoring AES | Computer produces summative assessment for evaluation
Aim: reduce human workload
AES has been put into practical use by ETS from 1999 | Computer produces summative assessment for evaluation
Aim: reduce human workload
AES has been put into practical use by ETS from 1999 | [] |
GEM-SciDuet-train-5#paper-966#slide-1 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-1 | Prompt specific and Independent AES | Most existing AES approaches are prompt-specific
Require human labels for each prompt to train
Can achieve satisfying human-machine agreement
Prompt-independent AES remains a challenge
Only non-target human labels are available | Most existing AES approaches are prompt-specific
Require human labels for each prompt to train
Can achieve satisfying human-machine agreement
Prompt-independent AES remains a challenge
Only non-target human labels are available | [] |
GEM-SciDuet-train-5#paper-966#slide-2 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-2 | Challenges in Prompt independent AES | Source Prompts Target Prompt
Learn essays Predict target
Previous approaches learn on source prompts
Domain adaption [Phandi et al. EMNLP 2015] Cross-domain learning [Dong & Zhang, EMNLP
Achieved Avg. QWK = 0.6395 at best with up to 100 labeled target essays
Off-topic: essays written for source prompts are mostly irrelevant | Source Prompts Target Prompt
Learn essays Predict target
Previous approaches learn on source prompts
Domain adaption [Phandi et al. EMNLP 2015] Cross-domain learning [Dong & Zhang, EMNLP
Achieved Avg. QWK = 0.6395 at best with up to 100 labeled target essays
Off-topic: essays written for source prompts are mostly irrelevant | [] |
GEM-SciDuet-train-5#paper-966#slide-3 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-3 | TDNN A Two stage Deep Neural Network for Prompt | Based on the idea of transductive transfer learning
Learn on target essays
Utilize the content of target essays to rate | Based on the idea of transductive transfer learning
Learn on target essays
Utilize the content of target essays to rate | [] |
GEM-SciDuet-train-5#paper-966#slide-4 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-4 | The Two stage Architecture | Prompt-independent stage: train a shallow model to create pseudo labels on the target prompt
Prompt-dependent stage: learn an end-to-end model to predict essay ratings for the target prompts | Prompt-independent stage: train a shallow model to create pseudo labels on the target prompt
Prompt-dependent stage: learn an end-to-end model to predict essay ratings for the target prompts | [] |
GEM-SciDuet-train-5#paper-966#slide-5 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-5 | Prompt independent stage | Train a robust prompt-independent AES model
Learning algorithm: RankSVM for AES
Select confident essays written for the target prompt
Predicted ratings in as negative examples
Predicted ratings in as positive examples
Converted to 0/1 labels
Common sense: 8 is good, <5 is bad | Train a robust prompt-independent AES model
Learning algorithm: RankSVM for AES
Select confident essays written for the target prompt
Predicted ratings in as negative examples
Predicted ratings in as positive examples
Converted to 0/1 labels
Common sense: 8 is good, <5 is bad | [] |
GEM-SciDuet-train-5#paper-966#slide-6 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-6 | Prompt dependent stage | Train a hybrid deep model for a prompt-
An end-to-end neural network with three parts | Train a hybrid deep model for a prompt-
An end-to-end neural network with three parts | [] |
GEM-SciDuet-train-5#paper-966#slide-7 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-7 | Architecture of the hybrid deep model | Multi-layer structure: Words (phrases) - Sentences Essay | Multi-layer structure: Words (phrases) - Sentences Essay | [] |
GEM-SciDuet-train-5#paper-966#slide-8 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-8 | Model Training | Training loss: MSE on 0/1 pseudo labels
Validation metric: Kappa on 30% non-target essays
Select the model that can best rate | Training loss: MSE on 0/1 pseudo labels
Validation metric: Kappa on 30% non-target essays
Select the model that can best rate | [] |
GEM-SciDuet-train-5#paper-966#slide-9 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-9 | Dataset and Metrics | We use the standard ASAP corpus
8 prompts with >10K essays in total
Prompt-independent AES: 7 prompts are used for training, 1 for testing
Report on common human-machine agreement metrics
Pearsons correlation coefficient (PCC)
Spearmans correlation coefficient (SCC)
Quadratic weighted Kappa (QWK) | We use the standard ASAP corpus
8 prompts with >10K essays in total
Prompt-independent AES: 7 prompts are used for training, 1 for testing
Report on common human-machine agreement metrics
Pearsons correlation coefficient (PCC)
Spearmans correlation coefficient (SCC)
Quadratic weighted Kappa (QWK) | [] |
GEM-SciDuet-train-5#paper-966#slide-10 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-10 | Baselines | RankSVM based on prompt-independent handcrafted
Also used in the prompt-independent stage in TDNN
Two LSTM layer + linear layer
CNN + LSTM + linear layer | RankSVM based on prompt-independent handcrafted
Also used in the prompt-independent stage in TDNN
Two LSTM layer + linear layer
CNN + LSTM + linear layer | [] |
GEM-SciDuet-train-5#paper-966#slide-11 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-11 | RankSVM is the most robust baseline | High variance of DNN models performance on all 8 prompts
Possibly caused by learning on non-target prompts RankSVM appears to be the most stable baseline Justifies the use of RankSVM in the first stage of TDNN | High variance of DNN models performance on all 8 prompts
Possibly caused by learning on non-target prompts RankSVM appears to be the most stable baseline Justifies the use of RankSVM in the first stage of TDNN | [] |
GEM-SciDuet-train-5#paper-966#slide-12 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-12 | Comparison to the best baseline | TDNN outperforms the best baseline on 7 out of 8 prompts Performance improvements gained by learning on the target prompt | TDNN outperforms the best baseline on 7 out of 8 prompts Performance improvements gained by learning on the target prompt | [] |
GEM-SciDuet-train-5#paper-966#slide-13 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-13 | Average performance on 8 prompts | Method QWK PCC SCC | Method QWK PCC SCC | [] |
GEM-SciDuet-train-5#paper-966#slide-14 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-14 | Sanity Check Relative Precision | How the quality of pseudo examples affects the performance of
The sanctity of the selected essays, namely, the number of positive
(negative) essays that are better (worse) than all negative (positive)
Such relative precision is at least 80% and mostly beyond 90% on different prompts
TDNN can at least learn
from correct 0/1 labels | How the quality of pseudo examples affects the performance of
The sanctity of the selected essays, namely, the number of positive
(negative) essays that are better (worse) than all negative (positive)
Such relative precision is at least 80% and mostly beyond 90% on different prompts
TDNN can at least learn
from correct 0/1 labels | [] |
GEM-SciDuet-train-5#paper-966#slide-15 | 966 | TDNN: A Two-stage Deep Neural Network for Prompt-independent Automated Essay Scoring | Existing automated essay scoring (AES) models rely on rated essays for the target prompt as training data. Despite their successes in prompt-dependent AES, how to effectively predict essay ratings under a prompt-independent setting remains a challenge, where the rated essays for the target prompt are not available. To close this gap, a two-stage deep neural network (TDNN) is proposed. In particular, in the first stage, using the rated essays for nontarget prompts as the training data, a shallow model is learned to select essays with an extreme quality for the target prompt, serving as pseudo training data; in the second stage, an end-to-end hybrid deep model is proposed to learn a prompt-dependent rating model consuming the pseudo training data from the first step. Evaluation of the proposed TDNN on the standard ASAP dataset demonstrates a promising improvement for the prompt-independent AES task. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119,
120,
121,
122,
123,
124,
125,
126,
127,
128,
129,
130,
131,
132,
133,
134,
135,
136,
137,
138,
139,
140,
141,
142,
143,
144,
145,
146,
147,
148,
149,
150,
151,
152,
153,
154,
155,
156,
157,
158,
159,
160,
161,
162,
163,
164,
165,
166,
167
],
"paper_content_text": [
"Introduction Automated essay scoring (AES) utilizes natural language processing and machine learning techniques to automatically rate essays written for a target prompt (Dikli, 2006) .",
"Currently, the AES systems have been widely used in large-scale English writing tests, e.g.",
"Graduate Record Examination (GRE), to reduce the human efforts in the writing assessments (Attali and Burstein, 2006) .",
"Existing AES approaches are promptdependent, where, given a target prompt, rated essays for this particular prompt are required for training (Dikli, 2006; Williamson, 2009; Foltz et al., 1999) .",
"While the established models are effective (Chen and He, 2013; Taghipour and Ng, 2016; Alikaniotis et al., 2016; Cummins et al., 2016; , we argue that the models for prompt-independent AES are also desirable to allow for better feasibility and flexibility of AES systems especially when the rated essays for a target prompt are difficult to obtain or even unaccessible.",
"For example, in a writing test within a small class, students are asked to write essays for a target prompt without any rated examples, where the prompt-dependent methods are unlikely to provide effective AES due to the lack of training data.",
"Prompt-independent AES, however, has drawn little attention in the literature, where there only exists unrated essays written for the target prompt, as well as the rated essays for several non-target prompts.",
"We argue that it is not straightforward, if possible, to apply the established promptdependent AES methods for the mentioned prompt-independent scenario.",
"On one hand, essays for different prompts may differ a lot in the uses of vocabulary, the structure, and the grammatic characteristics; on the other hand, however, established prompt-dependent AES models are designed to learn from these prompt-specific features, including the on/off-topic degree, the tfidf weights of topical terms (Attali and Burstein, 2006; Dikli, 2006) , and the n-gram features extracted from word semantic embeddings (Dong and Zhang, 2016; Alikaniotis et al., 2016) .",
"Consequently, the prompt-dependent models can hardly learn generalized rules from rated essays for nontarget prompts, and are not suitable for the promptindependent AES.",
"Being aware of this difficulty, to this end, a twostage deep neural network, coined as TDNN, is proposed to tackle the prompt-independent AES problem.",
"In particular, to mitigate the lack of the prompt-dependent labeled data, at the first stage, a shallow model is trained on a number of rated essays for several non-target prompts; given a target prompt and a set of essays to rate, the trained model is employed to generate pseudo training data by selecting essays with the extreme quality.",
"At the second stage, a novel end-to-end hybrid deep neural network learns prompt-dependent features from these selected training data, by considering semantic, part-of-speech, and syntactic features.",
"The contributions in this paper are threefold: 1) a two-stage learning framework is proposed to bridge the gap between the target and non-target prompts, by only consuming rated essays for nontarget prompts as training data; 2) a novel deep model is proposed to learn from pseudo labels by considering semantic, part-of-speech, and syntactic features; and most importantly, 3) to the best of our knowledge, the proposed TDNN is actually the first approach dedicated to addressing the prompt-independent AES.",
"Evaluation on the standard ASAP dataset demonstrates the effectiveness of the proposed method.",
"The rest of this paper is organized as follows.",
"In Section 2, we describe our novel TDNN model, including the two-stage framework and the proposed deep model.",
"Following that, we describe the setup of our empirical study in Section 3, thereafter present the results and provide analyzes in Section 4.",
"Section 5 recaps existing literature and put our work in context, before drawing final conclusions in Section 6.",
"Two-stage Deep Neural Network for AES In this section, the proposed two-stage deep neural network (TDNN) for prompt-independent AES is described.",
"To accurately rate an essay, on one hand, we need to consider its pertinence to the given prompt; on the other hand, the organization, the analyzes, as well as the uses of the vocabulary are all crucial for the assessment.",
"Henceforth, both prompt-dependent and -independent factors should be considered, but the latter ones actually do not require prompt-dependent training data.",
"Accordingly, in the proposed framework, a supervised ranking model is first trained to learn from prompt-independent data, hoping to roughly assess essays without considering the prompt; subsequently, given the test dataset, namely, a set of essays for a target prompt, a subset of essays are selected as positive and negative training data based on the prediction of the trained model from the first stage; ultimately, a novel deep model is proposed to learn both prompt-dependent and -independent factors on this selected subset.",
"As indicated in Figure 1 , the proposed framework includes two stages.",
"Prompt-independent stage.",
"Only the promptindependent factors are considered to train a shallow model, aiming to recognize the essays with the extreme quality in the test dataset, where the rated essays for non-target prompts are used for training.",
"Intuitively, one could recognize essays with the highest and the lowest scores correctly by solely examining their quality of writing, e.g., the number of typos, without even understanding them, and the prompt-independent features such as the number of grammatic and spelling errors should be sufficient to fulfill this screening procedure.",
"Accordingly, a supervised model trained solely on prompt-independent features is employed to identify the essays with the highest and lowest scores in a given set of essays for the target prompt, which are used as the positive and negative training data in the follow-up prompt-dependent learning phase.",
"Overview Prompt-dependent stage.",
"Intuitively, most essays are with a quality in between the extremes, requiring a good understanding of their meaning to make an accurate assessment, e.g., whether the examples from the essay are convincing or whether the analyzes are insightful, making the consideration of prompt-dependent features crucial.",
"To achieve that, a model is trained to learn from the comparison between essays with the highest and lowest scores for the target prompt according to the predictions from the first step.",
"Akin to the settings in transductive transfer learning (Pan and Yang, 2010), given essays for a particular prompt, quite a few confident essays at two extremes are selected and are used to train another model for a fine-grained content-based prompt-dependent assessment.",
"To enable this, a powerful deep model is proposed to consider the content of the essays from different perspectives using semantic, part-of-speech (POS) and syntactic network.",
"After being trained with the selected essays, the deep model is expected to memorize the properties of a good essay in response to the target prompt, thereafter accurately assessing all essays for it.",
"In Section 2.2, building blocks for the selection of the training data and the proposed deep model are described in details.",
"Building Blocks Select confident essays as training data.",
"The identification of the extremes is relatively simple, where a RankSVM (Joachims, 2002) is trained on essays for different non-target prompts, avoiding the risks of over-fitting some particular prompts.",
"A set of established prompt-independent features are employed, which are listed in Table 2 .",
"Given a prompt and a set of essays for evaluation, to begin with, the trained RankSVM is used to assign prediction scores to individual prompt-essay pairs, which are uniformly transformed into a 10point scale.",
"Thereafter, the essays with predicted scores in [0, 4] and [8, 10] are selected as negative and positive examples respectively, serving as the bad and good templates for training in the next stage.",
"Intuitively, an essay with a score beyond eight out of a 10-point scale is considered good, while the one receiving less than or equal to four, is considered to be with a poor quality.",
"A hybrid deep model for fine-grained assessment.",
"To enable a prompt-dependent assessment, a model is desired to comprehensively capture the ways in which a prompt is described or discussed in an essay.",
"In this paper, semantic meaning, part-of-speech (POS), and the syntactic taggings of the token sequence from an essay are considered, grasping the quality of an essay for a target prompt.",
"The model architecture is summarized in Figure 2 .",
"Intuitively, the model learns the semantic meaning of an essay by encoding it in terms of a sequence of word embeddings, denoted as − → e sem , hoping to understand what the essay is about; in addition, the part-of-speech information is encoded as a sequence of POS tag-gings, coined as − → e pos ; ultimately, the structural connections between different components in an essay (e.g., terms or phrases) are further captured via syntactic network, leading to − → e synt , where the model learns the organization of the essay.",
"Akin to (Li et al., 2015) and (Zhou and Xu, 2015) , bi-LSTM is employed as a basic component to encode a sequence.",
"Three features are separately captured using the stacked bi-LSTM layers as building blocks to encode different embeddings, whose outputs are subsequently concatenated and fed into several dense layers, generating the ultimate rating.",
"In the following, the architecture of the model is described in details.",
"-Semantic embedding.",
"Akin to the existing works (Alikaniotis et al., 2016; Taghipour and Ng, 2016) , semantic word embeddings, namely, the pre-trained 50-dimension GloVe (Pennington et al., 2014) , are employed.",
"On top of the word embeddings, two bi-LSTM layers are stacked, namely, the essay layer is constructed on top of the sentence layer, ending up with the semantic representation of the whole essay, which is denoted as − → e sem in Figure 2 .",
"-Part-Of-Speech (POS) embeddings for individual terms are first generated by the Stanford Tagger (Toutanova et al., 2003) , where 36 different POS tags present.",
"Accordingly, individual words are embedded with 36-dimensional one-hot representation, and is transformed to a 50-dimensional vector through a lookup layer.",
"After that, two bi-LSTM layers are stacked, leading to − → e pos .",
"Take Figure 3 for example, given a sentence \"Attention please, here is an example.",
"\", it is first converted into a POS sequence using the tagger, namely, VB, VBP, RB, VBZ, DT, NN; thereafter it is further mapped to vector space through one-hot embedding and a lookup layer.",
"-Syntactic embedding aims at encoding an essay in terms of the syntactic relationships among different syntactic components, by encoding an essay recursively.",
"The Stanford Parser (Socher et al., 2013) is employed to label the syntactic structure of words and phrases in sentences, accounting for 59 different types in total.",
"Similar to (Tai et al., 2015) , we opt for three stacked bi-LSTM, aiming at encoding individual phrases, sentences, and ultimately the whole essay in sequence.",
"In particular, according to the hierarchical structure from a parsing tree, the phrase-level bi-LSTM first encodes different phrases by consuming syntactic Figure 2 ) from a lookup table of individual syntactic units in the tree; thereafter, the encoded dense layers in individual sentences are further consumed by a sentence-level bi-LSTM, ending up with sentence-level syntactic representations, which are ultimately combined by the essay-level bi-LSTM, resulting in − → e synt .",
"For example, the parsed tree for a sentence \"Attention please, here is an example.\"",
"is displayed in Figure 3 .",
"To start with, the sentence is parsed into ((NP VP)(NP VP NP)), and the dense embeddings are fetched from a lookup table for all tokens, namely, NP and VP; thereafter, the phraselevel bi-LSTM encodes (NP VP) and (NP VP N-P) separately, which are further consumed by the sentence-level bi-LSTM.",
"Afterward, essay-level bi-LSTM further combines the representations of different sentences into − → e synt .",
"(ROOT (S (S (NP (VB Attention)) (VP (VBP please))) (, ,) (NP (RB here)) (VP (VBZ is) (NP (DT an) (NN example))) (.",
".)))",
"Figure 3: An example of the context-free phrase structure grammar tree.",
"-Combination.",
"A feed-forward network linearly transforms the concatenated representations of an essay from the mentioned three perspectives into a scalar, which is further normalized into [0, 1] with a sigmoid function.",
"Objective and Training Objective.",
"Mean square error (MSE) is optimized, which is widely used as a loss function in regression tasks.",
"Given N pairs of a target prompt p i and an essay e i , MSE measures the average value of square error between the normalized gold standard rating r * (p i , e i ) and the predicted rating r(p i , e i ) assigned by the AES model, as summarized in Equation 1.",
"1 N N ∑ i=1 ( r(p i , e i ) − r * (p i , e i ) ) 2 (1) Optimization.",
"Adam (Kingma and Ba, 2014) is employed to minimize the loss over the training data.",
"The initial learning rate η is set to 0.01 and the gradient is clipped between [−10, 10] during training.",
"In addition, dropout (Srivastava et al., 2014) is introduced for regularization with a dropout rate of 0.5, and 64 samples are used in each batch with batch normalization (Ioffe and Szegedy, 2015) .",
"30% of the training data are reserved for validation.",
"In addition, early stopping (Yao et al., 2007) is employed according to the validation loss, namely, the training is terminated if no decrease of the loss is observed for ten consecutive epochs.",
"Once training is finished, Prompt #Essays Avg Length Score Range 1 1783 350 2-12 2 1800 350 1-6 3 1726 150 0-3 4 1772 150 0-3 5 1805 150 0-4 6 1800 150 0-4 7 1569 250 0-30 8 723 650 0-60 Table 1 : Statistics for the ASAP dataset.",
"akin to , the model with the best quadratic weighted kappa on the validation set is selected.",
"3 Experimental Setup Dataset.",
"The Automated Student Assessment Prize (ASAP) dataset has been widely used for AES (Alikaniotis et al., 2016; Chen and He, 2013; , and is also employed as the prime evaluation instrument herein.",
"In total, AS-AP consists of eight sets of essays, each of which associates to one prompt, and is originally written by students between Grade 7 and Grade 10.",
"As summarized in Table 1 , essays from different sets differ in their rating criteria, length, as well as the rating distribution 1 .",
"Cross-validation.",
"To fully employ the rated data, a prompt-wise eight-fold cross validation on the ASAP is used for evaluation.",
"In each fold, essays corresponding to a prompt is reserved for testing, and the remaining essays are used as training data.",
"Evaluation metric.",
"The model outputs are first uniformly re-scaled into [0, 10], mirroring the range of ratings in practice.",
"Thereafter, akin to (Yannakoudakis et al., 2011; Chen and He, 2013; Alikaniotis et al., 2016) , we report our results primarily based on the quadratic weighted Kappa (QWK), examining the agreement between the predicted ratings and the ground truth.",
"Pearson correlation coefficient (PCC) and Spearman rankorder correlation coefficient (SCC) are also reported.",
"The correlations obtained from individual folds, as well as the average over all eight folds, are reported as the ultimate results.",
"Competing models.",
"Since the promptindependent AES is of interests in this work, the existing AES models are adapted for prompt-independent rating prediction, serving as baselines.",
"This is due to the facts that the 1 Details of this dataset can be found at https://www.",
"kaggle.com/c/asap-aes.",
"No.",
"Feature 1 Mean & variance of word length in characters 2 Mean & variance of sentence length in words 3 Essay length in characters and words 4 Number of prepositions and commas 5 Number of unique words in an essay 6 Mean number of clauses per sentence 7 Mean length of clauses 8 Maximum number of clauses of a sentence in an essay 9 Number of spelling errors 10 Average depth of the parser tree of each sentence in an essay 11 Average depth of each leaf node in the parser tree of each sentence prompt-dependent and -independent models differ a lot in terms of problem settings and model designs, especially in their requirements for the training data, where the latter ones release the prompt-dependent requirements and thereby are accessible to more data.",
"-RankSVM, using handcrafted features for AES (Yannakoudakis et al., 2011; Chen et al., 2014) , is trained on a set of pre-defined promptindependent features as listed in Table 2 , where the features are standardized beforehand to remove the mean and variance.",
"The RankSVM is also used for the prompt-independent stage in our proposed TDNN model.",
"In particular, the linear kernel RankSVM 2 is employed, where C is set to 5 according to our pilot experiments.",
"-2L-LSTM.",
"Two-layer bi-LSTM with GloVe for AES (Alikaniotis et al., 2016) is employed as another baseline.",
"Regularized word embeddings are dropped to avoid over-fitting the prompt-specific features.",
"-CNN-LSTM.",
"This model (Taghipour and Ng, 2016 ) employs a convolutional (CNN) layer over one-hot representations of words, followed by an LSTM layer to encode word sequences in a given essay.",
"A linear layer with sigmoid activation function is then employed to predict the essay rating.",
"-CNN-LSTM-ATT.",
"This model ) employs a CNN layer to encode word sequences into sentences, followed by an LSTM layer to generate the essay representation.",
"An attention mechanism is added to model the influence of individual sentences on the final essay representation.",
"For the proposed TDNN model, as introduced in Section 2.2, different variants of TDNN are examined by using one or multiple components out of the semantic, POS and the syntactic networks.",
"The combinations being considered are listed in the following.",
"In particular, the dimensions of POS tags and syntactic network are fixed to 50, whereas the sizes of the hidden units in LSTM, as well as the output units of the linear layers are tuned by grid search.",
"-TDNN(Sem) only includes the semantic building block, which is similar to the two-layer LSTM neural network from (Alikaniotis et al., 2016) but without regularizing the word embeddings; -TDNN(Sem+POS) employs the semantic and the POS building blocks; -TDNN(Sem+Synt) uses the semantic and the syntactic network building blocks; -TDNN(POS+Synt) includes the POS and the syntactic network building blocks; -TDNN(ALL) employs all three building blocks.",
"The use of POS or syntactic network alone is not presented for brevity given the facts that they perform no better than TDNN(POS+Synt) in our pilot experiments.",
"Source code of the TDNN model is publicly available to enable further comparison 3 .",
"Results and Analyzes In this section, the evaluation results for different competing methods are compared and analyzed in terms of their agreements with the manual ratings using three correlation metrics, namely, QWK, PCC and SCC, where the best results for each prompt is highlighted in bold in Table 3 .",
"It can be seen that, for seven out of all eight prompts, the proposed TDNN variants outperform the baselines by a margin in terms of QWK, and the TDNN variant with semantic and syntactic features, namely, TDNN(Sem+Synt), consistently performs the best among different competing methods.",
"More precisely, as indicated in the bottom right corner in Table 3 , on average, TDNN(Sem+Synt) outperforms the baselines by at least 25.52% under QWK, by 10.28% under PCC, and by 15.66% under SCC, demonstrating that the proposed model not only correlates better with the manual ratings in terms of QWK, but also linearly (PCC) and monotonically (SCC) correlates better with the manual ratings.",
"As for the four baselines, note that, the relatively underperformed deep models suffer from larger variances of performance under different prompts, e.g., for prompts two and eight, 2L-LSTM's QWK is lower than 0.3.",
"This actually confirms our choice of RankSVM for the first stage in TDNN, since a more complicated model (like 2L-LSTM) may end up with learning prompt-dependent signals, making it unsuitable for the prompt-independent rating prediction.",
"As a comparison, RankSVM performs more stable among different prompts.",
"As for the different TDNN variants, it turns out that the joint uses of syntactic network with semantic or POS features can lead to better performances.",
"This indicates that, when learning the prompt-dependent signals, apart from the widelyused semantic features, POS features and the sentence structure taggings (syntactic network) are also essential in learning the structure and the arrangement of an essay in response to a particular prompt, thereby being able to improve the results.",
"It is also worth mentioning, however, when using all three features, the TDNN actually performs worse than when only using (any) two features.",
"One possible explanation is that the uses of all three features result in a more complicated model, which over-fits the training data.",
"In addition, recall that the prompt-independent RankSVM model from the first stage enables the proposed TDNN in learning prompt-dependent information without manual ratings for the target prompt.",
"Therefore, one would like to understand how good the trained RankSVM is in feeding training data for the model in the second stage.",
"In particular, the precision, recall and F-score (P/R/F) of the essays selected by RanknSVM, namely, the negative ones rated between [0, 4], and the positive ones rated between [8, 10] , are displayed in Figure 4 .",
"It can be seen that the P/R/F scores of both positive and negative classes differ a lot among different prompts.",
"Moreover, it turns out that the P/R/F scores do not necessarily correlate with the performance of the TDNN model.",
"Take TDNN(Sem+Synt), the best TDNN variant, as an example: as indicated in Table 4 , the performance and the P/R/F scores of the pseudo examples are only weakly correlated in most cases.",
"To gain a better understanding in how the quality of pseudo examples affects the performance of TDNN, the sanctity of the selected essays are examined.",
"In Figure 5 , the relative precision of 7616 .7492 .7366 .7993 .7960 .6752 .6903 .7434 TDNN(POS+Synt) .7561 .7591 .7440 .7332 .7983 .7866 .6593 .6759 .7354 TDNN(All) .",
"7527 .7609 .7251 .7302 .7974 .7794 .6557 .6874 .7350 Method Prompt 7 Prompt 8 Average RankSVM .",
"5858 .6436 .6429 .4075 .5889 .6087 .5462 .6072 .5976 2L-LSTM .6690 .7637 .7607 .2486 .5137 .4979 .4687 .6548 .6214 CNN-LSTM .6609 .6849 .6865 .3812 .4666 .3872 .5362 .6569 .6139 CNN-LSTM-ATT .6002 .6314 .6223 .4468 .5358 .4536 .5057 .6535 .6368 TDNN(Sem) .",
"5482 .6957 .6902 .5003 .6083 .6545 .5875 .6779 .6795 TDNN(Sem+POS) .6239 .7111 .7243 .5519 .6219 .6614 .6582 .7103 .7130 TDNN(Sem+Synt) .6587 .7201 .7380 .5741 .6324 .6713 .6856 .7244 .7365 TDNN(POS+Synt) .6464 .7172 .7349 .5631 .6281 .6698 .6784 .7189 .7322 TDNN(All) .",
"6396 .7114 .7300 .5622 .6267 .6631 .6682 .7176 .7258 the selected positive and negative training data by RankSVM are displayed for all eight prompts in terms of their concordance with the manual ratings, by computing the number of positive (negative) essays that are better (worse) than all negative (positive) essays.",
"It can be seen that, such relative precision is at least 80% and mostly beyond 90% on different prompts, indicating that the overlap of the selected positive and negative essays are fairly small, guaranteeing that the deep model in the second stage at least learns from correct labels, which are crucial for the success of our TDNN model.",
"Beyond that, we further investigate the class balance of the selected training data from the first stage, which could also influence the ultimate results.",
"The number of selected positive and negative essays are reported in Table 5 , where for prompts three and eight the training data suffers from serious imbalanced problem, which may explain their lower performance (namely, the two lowest QWKs among different prompts).",
"On one hand, this is actually determined by real distribution of ratings for a particular prompt, e.g., how many essays are with an extreme quality for a given prompt in the target data.",
"On the other hand, a fine-grained tuning of the RankSVM (e.g., tuning C + and C − for positive and negative exam-ples separately) may partially resolve the problem, which is left for the future work.",
"Related Work Classical regression and classification algorithms are widely used for learning the rating model based on a variety of text features including lexical, syntactic, discourse and semantic features (Larkey, 1998; Rudner, 2002; Attali and Burstein, 2006; Mcnamara et al., 2015; Phandi et al., 2015) .",
"There are also approaches that see AES as a preference ranking problem by applying learning to ranking algorithms to learn the rating model.",
"Results show improvement of learning to rank approaches over classical regression and classification algorithms (Chen et al., 2014; Yannakoudakis et al., 2011) .",
"In addition, Chen & He propose to incorporate the evaluation metric into the loss function of listwise learning to rank for AES (Chen and He, 2013) .",
"Recently, there have been efforts in developing AES approaches based on deep neural networks (DNN), for which feature engineering is not required.",
"Taghipour & Ng explore a variety of neural network model architectures based on recurrent neural networks which can effectively encode the information required for essay scoring and learn the complex connections in the data through the non-linear neural layers (Taghipour and Ng, 2016) .",
"Alikaniotis et al.",
"introduce a neural network model to learn the extent to which specific words contribute to the text's score, which is embedded in the word representations.",
"Then a two-layer bi-directional Long-Short Term Memory networks (bi-LSTM) is used to learn the meaning of texts, and finally the essay score is predicted through a mutli-layer feed-forward network (Alikaniotis et al., 2016) .",
"Dong & Zhang employ a hierarchical convolutional neural network (CN-N) model, with a lower layer representing sentence structure and an upper layer representing essay structure based on sentence representations, to learn features automatically (Dong and Zhang, 2016) .",
"This model is later improved by employing attention layers.",
"Specifically, the model learns text representation with LSTMs which can model the coherence and co-reference among sequences of words and sentences, and uses attention pooling to capture more relevant words and sentences that contribute to the final quality of essays .",
"Song et al.",
"propose a deep model for identifying discourse modes in an essay (Song et al., 2017) .",
"While the literature has shown satisfactory performance of prompt-dependent AES, how to achieve effective essay scoring in a promptindependent setting remains to be explored.",
"Chen & He studied the usefulness of promptindependent text features and achieved a humanmachine rating agreement slightly lower than the use of all text features (Chen and He, 2013) for prompt-dependent essay scoring prediction.",
"A constrained multi-task pairwise preference learning approach was proposed in (Cummins et al., 2016) to combine essays from multiple prompts for training.",
"However, as shown by (Dong and Zhang, 2016; Zesch et al., 2015; Phandi et al., 2015) , straightforward applications of existing AES methods for prompt-independent AES lead to a poor performance.",
"Conclusions & Future Work This study aims at addressing the promptindependent automated essay scoring (AES), where no rated essay for the target prompt is available.",
"As demonstrated in the experiments, two kinds of established prompt-dependent AES models, namely, RankSVM for AES (Yannakoudakis et al., 2011; Chen et al., 2014) and the deep models for AES (Alikaniotis et al., 2016; Taghipour and Ng, 2016; , fail to provide satisfactory performances, justifying our arguments in Section 1 that the application of estab-lished prompt-dependent AES models on promptindependent AES is not straightforward.",
"Therefore, a two-stage TDNN learning framework was proposed to utilize the prompt-independent features to generate pseudo training data for the target prompt, on which a hybrid deep neural network model is proposed to learn a rating model consuming semantic, part-of-speech, and syntactic signals.",
"Through the experiments on the ASAP dataset, the proposed TDNN model outperforms the baselines, and leads to promising improvement in the human-machine agreement.",
"Given that our approach in this paper is similar to the methods for transductive transfer learning (Pan and Yang, 2010), we argue that the proposed TDNN could be further improved by migrating the non-target training data to the target prompt (Busto and Gall, 2017) .",
"Further study of the uses of transfer learning algorithms on promptindependent AES needs to be undertaken."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"2.3",
"4",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Two-stage Deep Neural Network for AES",
"Overview",
"Building Blocks",
"Objective and Training",
"Results and Analyzes",
"Related Work",
"Conclusions & Future Work"
]
} | GEM-SciDuet-train-5#paper-966#slide-15 | Conclusions | It is beneficial to learn an AES model on the target prompt
Syntactic features are useful addition to the widely used Word2Vec embeddings
Sanity check: small overlap between pos/neg examples
Prompt-independent AES remains an open problem
TDNN can achieve 0.68 at best | It is beneficial to learn an AES model on the target prompt
Syntactic features are useful addition to the widely used Word2Vec embeddings
Sanity check: small overlap between pos/neg examples
Prompt-independent AES remains an open problem
TDNN can achieve 0.68 at best | [] |
GEM-SciDuet-train-6#paper-970#slide-0 | 970 | Lancaster A at SemEval-2017 Task 5: Evaluation metrics matter: predicting sentiment from financial news headlines | This paper describes our participation in Task 5 track 2 of SemEval 2017 to predict the sentiment of financial news headlines for a specific company on a continuous scale between -1 and 1. We tackled the problem using a number of approaches, utilising a Support Vector Regression (SVR) and a Bidirectional Long Short-Term Memory (BLSTM). We found an improvement of 4-6% using the LSTM model over the SVR and came fourth in the track. We report a number of different evaluations using a finance specific word embedding model and reflect on the effects of using different evaluation metrics. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95
],
"paper_content_text": [
"Introduction The objective of Task 5 Track 2 of SemEval (2017) was to predict the sentiment of news headlines with respect to companies mentioned within the headlines.",
"This task can be seen as a financespecific aspect-based sentiment task (Nasukawa and Yi, 2003) .",
"The main motivations of this task is to find specific features and learning algorithms that will perform better for this domain as aspect based sentiment analysis tasks have been conducted before at SemEval (Pontiki et al., 2014) .",
"Domain specific terminology is expected to play a key part in this task, as reporters, investors and analysts in the financial domain will use a specific set of terminology when discussing financial performance.",
"Potentially, this may also vary across different financial domains and industry sectors.",
"Therefore, we took an exploratory approach and investigated how various features and learning algorithms perform differently, specifically SVR and BLSTMs.",
"We found that BLSTMs outperform an SVR without having any knowledge of the company that the sentiment is with respect to.",
"For replicability purposes, with this paper we are releasing our source code 1 and the finance specific BLSTM word embedding model 2 .",
"Related Work There is a growing amount of research being carried out related to sentiment analysis within the financial domain.",
"This work ranges from domainspecific lexicons (Loughran and McDonald, 2011) and lexicon creation (Moore et al., 2016) to stock market prediction models (Peng and Jiang, 2016; Kazemian et al., 2016) .",
"Peng and Jiang (2016) used a multi layer neural network to predict the stock market and found that incorporating textual features from financial news can improve the accuracy of prediction.",
"Kazemian et al.",
"(2016) showed the importance of tuning sentiment analysis to the task of stock market prediction.",
"However, much of the previous work was based on numerical financial stock market data rather than on aspect level financial textual data.",
"In aspect based sentiment analysis, there have been many different techniques used to predict the polarity of an aspect as shown in SemEval-2016 task 5 (Pontiki et al., 2014 ).",
"The winning system (Brun et al., 2016 ) used many different linguistic features and an ensemble model, and the runner up (Kumar et al., 2016) used uni-grams, bi-grams and sentiment lexicons as features for a Support Vector Machine (SVM).",
"Deep learning methods have also been applied to aspect polarity prediction.",
"Ruder et al.",
"(2016) created a hierarchical BLSTM with a sentence level BLSTM inputting into a review level BLSTM thus allowing them to take into account inter-and intra-sentence context.",
"They used only word embeddings making their system less dependent on extensive feature engineering or manual feature creation.",
"This system outperformed all others on certain languages on the SemEval-2016 task 5 dataset (Pontiki et al., 2014) and on other languages performed close to the best systems.",
"Wang et al.",
"(2016) also created an LSTM based model using word embeddings but instead of a hierarchical model it was a one layered LSTM with attention which puts more emphasis on learning the sentiment of words specific to a given aspect.",
"Data The training data published by the organisers for this track was a set of headline sentences from financial news articles where each sentence was tagged with the company name (which we treat as the aspect) and the polarity of the sentence with respect to the company.",
"There is the possibility that the same sentence occurs more than once if there is more than one company mentioned.",
"The polarity was a real value between -1 (negative sentiment) and 1 (positive sentiment).",
"We additionally trained a word2vec (Mikolov et al., 2013) word embedding model 3 on a set of 189,206 financial articles containing 161,877,425 tokens, that were manually downloaded from Factiva 4 .",
"The articles stem from a range of sources including the Financial Times and relate to companies from the United States only.",
"We trained the model on domain specific data as it has been shown many times that the financial domain can contain very different language.",
"System description Even though we have outlined this task as an aspect based sentiment task, this is instantiated in only one of the features in the SVR.",
"The following two subsections describe the two approaches, first SVR and then BLSTM.",
"Key implementation details are exposed here in the paper, but we have released the source code and word embedding models to aid replicability and further experimentation.",
"SVR The system was created using ScitKit learn (Pedregosa et al., 2011) linear Support Vector Regression model (Drucker et al., 1997) .",
"We exper-imented with the following different features and parameter settings: Tokenisation For comparison purposes, we tested whether or not a simple whitespace tokeniser can perform just as well as a full tokeniser, and in this case we used Unitok 5 .",
"N-grams We compared word-level uni-grams and bi-grams separately and in combination.",
"SVR parameters We tested different penalty parameters C and different epsilon parameters of the SVR.",
"Word Replacements We tested replacements to see if generalising words by inserting special tokens would help to reduce the sparsity problem.",
"We placed the word replacements into three separate groups: 1.",
"Company -When a company was mentioned in the input headline from the list of companies in the training data marked up as aspects, it was replaced by a company special token.",
"2.",
"Positive -When a positive word was mentioned in the input headline from a list of positive words (which was created using the N most similar words based on cosine distance) to 'excellent' using the pre-trained word2vec model.",
"3.",
"Negative -The same as the positive group however the word used was 'poor' instead of 'excellent'.",
"In the positive and negative groups, we chose the words 'excellent' and 'poor' following Turney (2002) to group the terms together under nondomain specific sentiment words.",
"Target aspect In order to incorporated the company as an aspect, we employed a boolean vector to represent the sentiment of the sentence.",
"This was done in order to see if the system could better differentiate the sentiment when the sentence was the same but the company was different.",
"BLSTM We created two different Bidirectional (Graves and Schmidhuber, 2005 ) Long Short-Term Memory (Hochreiter and Schmidhuber, 1997) using the Python Keras library (Chollet, 2015) with tensor flow backend (Abadi et al., 2016) .",
"We choose an LSTM model as it solves the vanishing gradients problem of Recurrent Neural Networks.",
"We used a bidirectional model as it allows us to capture information that came before and after instead of just before, thereby allowing us to capture more relevant context within the model.",
"Practically, a BLSTM is two LSTMs one going forward through the tokens the other in reverse order and in our models concatenating the resulting output vectors together at each time step.",
"The BLSTM models take as input a headline sentence of size L tokens 6 where L is the length of the longest sentence in the training texts.",
"Each word is converted into a 300 dimension vector using the word2vec model trained over the financial text 7 .",
"Any text that is not recognised by the word2vec model is represented as a vector of zeros; this is also used to pad out the sentence if it is shorter than L. Both BLSTM models have the following similar properties: 1.",
"Gradient clipping value of 5 -This was to help with the exploding gradients problem.",
"2.",
"Minimised the Mean Square Error (MSE) loss using RMSprop with a mini batch size of 32.",
"The output activation function is linear.",
"The main difference between the two models is the use of drop out and when they stop training over the data (epoch).",
"Both models architectures can be seen in figure 1.",
"Standard LSTM (SLSTM) The BLSTMs do contain drop out in both the input and between the connections of 0.2 each.",
"Finally the epoch is fixed at 25.",
"Early LSTM (ELSTM) As can be seen from figure 1, the drop out of 0.5 only happens between the layers and not the 6 Tokenised by Unitok 7 See the following link for detailed implementation details https://github.com/apmoore1/semeval# finance-word2vec-model connections as in the SLSTM.",
"Also the epoch is not fixed, it uses early stopping with a patience of 10.",
"We expect that this model can generalise better than the standard one due to the higher drop out and that the epoch is based on early stopping which relies on a validation set to know when to stop training.",
"Results We first present our findings on the best performing parameters and features for the SVRs.",
"These were determined by cross validation (CV) scores on the provided training data set using cosine similarity as the evaluation metric.",
"8 We found that using uni-grams and bi-grams performs best and using only bi-grams to be the worst.",
"Using the Unitok tokeniser always performed better than simple whitespace tokenisation.",
"The binary presence of tokens over frequency did not alter performance.",
"The C parameter was tested for three values; 0.01, 0.1 and 1.",
"We found very little difference between 0.1 and 1, but 0.01 produced much poorer results.",
"The eplison parameter was tested for 0.001, 0.01 and 0.1 the performance did not differ much but the lower the higher the performance but the more likely to overfit.",
"Using word replacements was effective for all three types (company, positive and negative) but using a value N=10 performed best for both positive and negative words.",
"Using target aspects also improved results.",
"Therefore, the best SVR model comprised of: Unitok tokenisation, uni-and bi-grams, word representation, C=0.1, eplison=0.01, company, positive, and negative word replacements and target aspects.",
"The main evaluation over the test data is based on the best performing SVR and the two BLSTM models once trained on all of the training data.",
"The result table 1 shows three columns based on the three evaluation metrics that the organisers have used.",
"Metric 1 is the original metric, weighted cosine similarity (the metric used to evaluate the final version of the results, where we were ranked 5th; metric provided on the task website 9 ).",
"This was then changed after the evaluation deadline to equation 1 10 (which we term metric 2; this is what the first version of the results were actually based on, where we were ranked 4th), which then changed by the organisers to their equation as presented in Cortis et al.",
"(2017) (which we term metric 3 and what the second version of the results were based on, where we were ranked 5th).",
"Model Metric 1 As you can see from the results table 1, the difference between the metrics is quite substantial.",
"This is due to the system's optimisation being based on metric 1 rather than 2.",
"Metric 2 is a classification metric for sentences with one aspect as it penalises values that are of opposite sign (giving -1 score) and rewards values with the same sign (giving +1 score).",
"Our systems are not optimised for this because it would predict scores of -0.01 and true value of 0.01 as very close (within vector of other results) with low error whereas metric 2 would give this the highest error rating of -1 as they are not the same sign.",
"Metric 3 is more similar to metric 1 as shown by the results, however the crucial difference is that again if you get opposite signs it will penalise more.",
"We analysed the top 50 errors based on Mean Absolute Error (MAE) in the test dataset specifically to examine the number of sentences containing more than one aspect.",
"Our investigation shows that no one system is better at predicting the sentiment of sentences that have more than one aspect (i.e.",
"company) within them.",
"Within those top 50 errors we found that the BLSTM systems do not know which parts of the sentence are associated to the company the sentiment is with respect to.",
"Also they do not know the strength/existence of certain sentiment words.",
"Conclusion and Future Work In this short paper, we have described our implemented solutions to SemEval Task 5 track 2, utilising both SVR and BLSTM approaches.",
"Our results show an improvement of around 5% when using LSTM models relative to SVR.",
"We have shown that this task can be partially represented as an aspect based sentiment task on a domain specific problem.",
"In general, our approaches acted as sentence level classifiers as they take no target company into consideration.",
"As our results show, the choice of evaluation metric makes a great deal of difference to system training and testing.",
"Future work will be to implement aspect specific information into an LSTM model as it has been shown to be useful in other work (Wang et al., 2016) ."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"4",
"4.1",
"4.1.1",
"4.1.2",
"4.1.3",
"4.1.4",
"4.1.5",
"4.2",
"3.",
"4.2.1",
"4.2.2",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Data",
"System description",
"SVR",
"Tokenisation",
"N-grams",
"SVR parameters",
"Word Replacements",
"Target aspect",
"BLSTM",
"The output activation function is linear.",
"Standard LSTM (SLSTM)",
"Early LSTM (ELSTM)",
"Results",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-6#paper-970#slide-0 | The task | Why AstraZeneca plc Dixons Carphone PLC Are Red-Hot Growth
Training data: 1142 samples, 960 headlines/sentences.
Testing data: 491 samples, 461 headlines/sentences. | Why AstraZeneca plc Dixons Carphone PLC Are Red-Hot Growth
Training data: 1142 samples, 960 headlines/sentences.
Testing data: 491 samples, 461 headlines/sentences. | [] |
GEM-SciDuet-train-6#paper-970#slide-1 | 970 | Lancaster A at SemEval-2017 Task 5: Evaluation metrics matter: predicting sentiment from financial news headlines | This paper describes our participation in Task 5 track 2 of SemEval 2017 to predict the sentiment of financial news headlines for a specific company on a continuous scale between -1 and 1. We tackled the problem using a number of approaches, utilising a Support Vector Regression (SVR) and a Bidirectional Long Short-Term Memory (BLSTM). We found an improvement of 4-6% using the LSTM model over the SVR and came fourth in the track. We report a number of different evaluations using a finance specific word embedding model and reflect on the effects of using different evaluation metrics. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95
],
"paper_content_text": [
"Introduction The objective of Task 5 Track 2 of SemEval (2017) was to predict the sentiment of news headlines with respect to companies mentioned within the headlines.",
"This task can be seen as a financespecific aspect-based sentiment task (Nasukawa and Yi, 2003) .",
"The main motivations of this task is to find specific features and learning algorithms that will perform better for this domain as aspect based sentiment analysis tasks have been conducted before at SemEval (Pontiki et al., 2014) .",
"Domain specific terminology is expected to play a key part in this task, as reporters, investors and analysts in the financial domain will use a specific set of terminology when discussing financial performance.",
"Potentially, this may also vary across different financial domains and industry sectors.",
"Therefore, we took an exploratory approach and investigated how various features and learning algorithms perform differently, specifically SVR and BLSTMs.",
"We found that BLSTMs outperform an SVR without having any knowledge of the company that the sentiment is with respect to.",
"For replicability purposes, with this paper we are releasing our source code 1 and the finance specific BLSTM word embedding model 2 .",
"Related Work There is a growing amount of research being carried out related to sentiment analysis within the financial domain.",
"This work ranges from domainspecific lexicons (Loughran and McDonald, 2011) and lexicon creation (Moore et al., 2016) to stock market prediction models (Peng and Jiang, 2016; Kazemian et al., 2016) .",
"Peng and Jiang (2016) used a multi layer neural network to predict the stock market and found that incorporating textual features from financial news can improve the accuracy of prediction.",
"Kazemian et al.",
"(2016) showed the importance of tuning sentiment analysis to the task of stock market prediction.",
"However, much of the previous work was based on numerical financial stock market data rather than on aspect level financial textual data.",
"In aspect based sentiment analysis, there have been many different techniques used to predict the polarity of an aspect as shown in SemEval-2016 task 5 (Pontiki et al., 2014 ).",
"The winning system (Brun et al., 2016 ) used many different linguistic features and an ensemble model, and the runner up (Kumar et al., 2016) used uni-grams, bi-grams and sentiment lexicons as features for a Support Vector Machine (SVM).",
"Deep learning methods have also been applied to aspect polarity prediction.",
"Ruder et al.",
"(2016) created a hierarchical BLSTM with a sentence level BLSTM inputting into a review level BLSTM thus allowing them to take into account inter-and intra-sentence context.",
"They used only word embeddings making their system less dependent on extensive feature engineering or manual feature creation.",
"This system outperformed all others on certain languages on the SemEval-2016 task 5 dataset (Pontiki et al., 2014) and on other languages performed close to the best systems.",
"Wang et al.",
"(2016) also created an LSTM based model using word embeddings but instead of a hierarchical model it was a one layered LSTM with attention which puts more emphasis on learning the sentiment of words specific to a given aspect.",
"Data The training data published by the organisers for this track was a set of headline sentences from financial news articles where each sentence was tagged with the company name (which we treat as the aspect) and the polarity of the sentence with respect to the company.",
"There is the possibility that the same sentence occurs more than once if there is more than one company mentioned.",
"The polarity was a real value between -1 (negative sentiment) and 1 (positive sentiment).",
"We additionally trained a word2vec (Mikolov et al., 2013) word embedding model 3 on a set of 189,206 financial articles containing 161,877,425 tokens, that were manually downloaded from Factiva 4 .",
"The articles stem from a range of sources including the Financial Times and relate to companies from the United States only.",
"We trained the model on domain specific data as it has been shown many times that the financial domain can contain very different language.",
"System description Even though we have outlined this task as an aspect based sentiment task, this is instantiated in only one of the features in the SVR.",
"The following two subsections describe the two approaches, first SVR and then BLSTM.",
"Key implementation details are exposed here in the paper, but we have released the source code and word embedding models to aid replicability and further experimentation.",
"SVR The system was created using ScitKit learn (Pedregosa et al., 2011) linear Support Vector Regression model (Drucker et al., 1997) .",
"We exper-imented with the following different features and parameter settings: Tokenisation For comparison purposes, we tested whether or not a simple whitespace tokeniser can perform just as well as a full tokeniser, and in this case we used Unitok 5 .",
"N-grams We compared word-level uni-grams and bi-grams separately and in combination.",
"SVR parameters We tested different penalty parameters C and different epsilon parameters of the SVR.",
"Word Replacements We tested replacements to see if generalising words by inserting special tokens would help to reduce the sparsity problem.",
"We placed the word replacements into three separate groups: 1.",
"Company -When a company was mentioned in the input headline from the list of companies in the training data marked up as aspects, it was replaced by a company special token.",
"2.",
"Positive -When a positive word was mentioned in the input headline from a list of positive words (which was created using the N most similar words based on cosine distance) to 'excellent' using the pre-trained word2vec model.",
"3.",
"Negative -The same as the positive group however the word used was 'poor' instead of 'excellent'.",
"In the positive and negative groups, we chose the words 'excellent' and 'poor' following Turney (2002) to group the terms together under nondomain specific sentiment words.",
"Target aspect In order to incorporated the company as an aspect, we employed a boolean vector to represent the sentiment of the sentence.",
"This was done in order to see if the system could better differentiate the sentiment when the sentence was the same but the company was different.",
"BLSTM We created two different Bidirectional (Graves and Schmidhuber, 2005 ) Long Short-Term Memory (Hochreiter and Schmidhuber, 1997) using the Python Keras library (Chollet, 2015) with tensor flow backend (Abadi et al., 2016) .",
"We choose an LSTM model as it solves the vanishing gradients problem of Recurrent Neural Networks.",
"We used a bidirectional model as it allows us to capture information that came before and after instead of just before, thereby allowing us to capture more relevant context within the model.",
"Practically, a BLSTM is two LSTMs one going forward through the tokens the other in reverse order and in our models concatenating the resulting output vectors together at each time step.",
"The BLSTM models take as input a headline sentence of size L tokens 6 where L is the length of the longest sentence in the training texts.",
"Each word is converted into a 300 dimension vector using the word2vec model trained over the financial text 7 .",
"Any text that is not recognised by the word2vec model is represented as a vector of zeros; this is also used to pad out the sentence if it is shorter than L. Both BLSTM models have the following similar properties: 1.",
"Gradient clipping value of 5 -This was to help with the exploding gradients problem.",
"2.",
"Minimised the Mean Square Error (MSE) loss using RMSprop with a mini batch size of 32.",
"The output activation function is linear.",
"The main difference between the two models is the use of drop out and when they stop training over the data (epoch).",
"Both models architectures can be seen in figure 1.",
"Standard LSTM (SLSTM) The BLSTMs do contain drop out in both the input and between the connections of 0.2 each.",
"Finally the epoch is fixed at 25.",
"Early LSTM (ELSTM) As can be seen from figure 1, the drop out of 0.5 only happens between the layers and not the 6 Tokenised by Unitok 7 See the following link for detailed implementation details https://github.com/apmoore1/semeval# finance-word2vec-model connections as in the SLSTM.",
"Also the epoch is not fixed, it uses early stopping with a patience of 10.",
"We expect that this model can generalise better than the standard one due to the higher drop out and that the epoch is based on early stopping which relies on a validation set to know when to stop training.",
"Results We first present our findings on the best performing parameters and features for the SVRs.",
"These were determined by cross validation (CV) scores on the provided training data set using cosine similarity as the evaluation metric.",
"8 We found that using uni-grams and bi-grams performs best and using only bi-grams to be the worst.",
"Using the Unitok tokeniser always performed better than simple whitespace tokenisation.",
"The binary presence of tokens over frequency did not alter performance.",
"The C parameter was tested for three values; 0.01, 0.1 and 1.",
"We found very little difference between 0.1 and 1, but 0.01 produced much poorer results.",
"The eplison parameter was tested for 0.001, 0.01 and 0.1 the performance did not differ much but the lower the higher the performance but the more likely to overfit.",
"Using word replacements was effective for all three types (company, positive and negative) but using a value N=10 performed best for both positive and negative words.",
"Using target aspects also improved results.",
"Therefore, the best SVR model comprised of: Unitok tokenisation, uni-and bi-grams, word representation, C=0.1, eplison=0.01, company, positive, and negative word replacements and target aspects.",
"The main evaluation over the test data is based on the best performing SVR and the two BLSTM models once trained on all of the training data.",
"The result table 1 shows three columns based on the three evaluation metrics that the organisers have used.",
"Metric 1 is the original metric, weighted cosine similarity (the metric used to evaluate the final version of the results, where we were ranked 5th; metric provided on the task website 9 ).",
"This was then changed after the evaluation deadline to equation 1 10 (which we term metric 2; this is what the first version of the results were actually based on, where we were ranked 4th), which then changed by the organisers to their equation as presented in Cortis et al.",
"(2017) (which we term metric 3 and what the second version of the results were based on, where we were ranked 5th).",
"Model Metric 1 As you can see from the results table 1, the difference between the metrics is quite substantial.",
"This is due to the system's optimisation being based on metric 1 rather than 2.",
"Metric 2 is a classification metric for sentences with one aspect as it penalises values that are of opposite sign (giving -1 score) and rewards values with the same sign (giving +1 score).",
"Our systems are not optimised for this because it would predict scores of -0.01 and true value of 0.01 as very close (within vector of other results) with low error whereas metric 2 would give this the highest error rating of -1 as they are not the same sign.",
"Metric 3 is more similar to metric 1 as shown by the results, however the crucial difference is that again if you get opposite signs it will penalise more.",
"We analysed the top 50 errors based on Mean Absolute Error (MAE) in the test dataset specifically to examine the number of sentences containing more than one aspect.",
"Our investigation shows that no one system is better at predicting the sentiment of sentences that have more than one aspect (i.e.",
"company) within them.",
"Within those top 50 errors we found that the BLSTM systems do not know which parts of the sentence are associated to the company the sentiment is with respect to.",
"Also they do not know the strength/existence of certain sentiment words.",
"Conclusion and Future Work In this short paper, we have described our implemented solutions to SemEval Task 5 track 2, utilising both SVR and BLSTM approaches.",
"Our results show an improvement of around 5% when using LSTM models relative to SVR.",
"We have shown that this task can be partially represented as an aspect based sentiment task on a domain specific problem.",
"In general, our approaches acted as sentence level classifiers as they take no target company into consideration.",
"As our results show, the choice of evaluation metric makes a great deal of difference to system training and testing.",
"Future work will be to implement aspect specific information into an LSTM model as it has been shown to be useful in other work (Wang et al., 2016) ."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"4",
"4.1",
"4.1.1",
"4.1.2",
"4.1.3",
"4.1.4",
"4.1.5",
"4.2",
"3.",
"4.2.1",
"4.2.2",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Data",
"System description",
"SVR",
"Tokenisation",
"N-grams",
"SVR parameters",
"Word Replacements",
"Target aspect",
"BLSTM",
"The output activation function is linear.",
"Standard LSTM (SLSTM)",
"Early LSTM (ELSTM)",
"Results",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-6#paper-970#slide-1 | Models | 1. Support Vector Regression (SVR) [1]
2. Bi-directional Long Short-Term Memory BLSTM [2][3] | 1. Support Vector Regression (SVR) [1]
2. Bi-directional Long Short-Term Memory BLSTM [2][3] | [] |
GEM-SciDuet-train-6#paper-970#slide-2 | 970 | Lancaster A at SemEval-2017 Task 5: Evaluation metrics matter: predicting sentiment from financial news headlines | This paper describes our participation in Task 5 track 2 of SemEval 2017 to predict the sentiment of financial news headlines for a specific company on a continuous scale between -1 and 1. We tackled the problem using a number of approaches, utilising a Support Vector Regression (SVR) and a Bidirectional Long Short-Term Memory (BLSTM). We found an improvement of 4-6% using the LSTM model over the SVR and came fourth in the track. We report a number of different evaluations using a finance specific word embedding model and reflect on the effects of using different evaluation metrics. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95
],
"paper_content_text": [
"Introduction The objective of Task 5 Track 2 of SemEval (2017) was to predict the sentiment of news headlines with respect to companies mentioned within the headlines.",
"This task can be seen as a financespecific aspect-based sentiment task (Nasukawa and Yi, 2003) .",
"The main motivations of this task is to find specific features and learning algorithms that will perform better for this domain as aspect based sentiment analysis tasks have been conducted before at SemEval (Pontiki et al., 2014) .",
"Domain specific terminology is expected to play a key part in this task, as reporters, investors and analysts in the financial domain will use a specific set of terminology when discussing financial performance.",
"Potentially, this may also vary across different financial domains and industry sectors.",
"Therefore, we took an exploratory approach and investigated how various features and learning algorithms perform differently, specifically SVR and BLSTMs.",
"We found that BLSTMs outperform an SVR without having any knowledge of the company that the sentiment is with respect to.",
"For replicability purposes, with this paper we are releasing our source code 1 and the finance specific BLSTM word embedding model 2 .",
"Related Work There is a growing amount of research being carried out related to sentiment analysis within the financial domain.",
"This work ranges from domainspecific lexicons (Loughran and McDonald, 2011) and lexicon creation (Moore et al., 2016) to stock market prediction models (Peng and Jiang, 2016; Kazemian et al., 2016) .",
"Peng and Jiang (2016) used a multi layer neural network to predict the stock market and found that incorporating textual features from financial news can improve the accuracy of prediction.",
"Kazemian et al.",
"(2016) showed the importance of tuning sentiment analysis to the task of stock market prediction.",
"However, much of the previous work was based on numerical financial stock market data rather than on aspect level financial textual data.",
"In aspect based sentiment analysis, there have been many different techniques used to predict the polarity of an aspect as shown in SemEval-2016 task 5 (Pontiki et al., 2014 ).",
"The winning system (Brun et al., 2016 ) used many different linguistic features and an ensemble model, and the runner up (Kumar et al., 2016) used uni-grams, bi-grams and sentiment lexicons as features for a Support Vector Machine (SVM).",
"Deep learning methods have also been applied to aspect polarity prediction.",
"Ruder et al.",
"(2016) created a hierarchical BLSTM with a sentence level BLSTM inputting into a review level BLSTM thus allowing them to take into account inter-and intra-sentence context.",
"They used only word embeddings making their system less dependent on extensive feature engineering or manual feature creation.",
"This system outperformed all others on certain languages on the SemEval-2016 task 5 dataset (Pontiki et al., 2014) and on other languages performed close to the best systems.",
"Wang et al.",
"(2016) also created an LSTM based model using word embeddings but instead of a hierarchical model it was a one layered LSTM with attention which puts more emphasis on learning the sentiment of words specific to a given aspect.",
"Data The training data published by the organisers for this track was a set of headline sentences from financial news articles where each sentence was tagged with the company name (which we treat as the aspect) and the polarity of the sentence with respect to the company.",
"There is the possibility that the same sentence occurs more than once if there is more than one company mentioned.",
"The polarity was a real value between -1 (negative sentiment) and 1 (positive sentiment).",
"We additionally trained a word2vec (Mikolov et al., 2013) word embedding model 3 on a set of 189,206 financial articles containing 161,877,425 tokens, that were manually downloaded from Factiva 4 .",
"The articles stem from a range of sources including the Financial Times and relate to companies from the United States only.",
"We trained the model on domain specific data as it has been shown many times that the financial domain can contain very different language.",
"System description Even though we have outlined this task as an aspect based sentiment task, this is instantiated in only one of the features in the SVR.",
"The following two subsections describe the two approaches, first SVR and then BLSTM.",
"Key implementation details are exposed here in the paper, but we have released the source code and word embedding models to aid replicability and further experimentation.",
"SVR The system was created using ScitKit learn (Pedregosa et al., 2011) linear Support Vector Regression model (Drucker et al., 1997) .",
"We exper-imented with the following different features and parameter settings: Tokenisation For comparison purposes, we tested whether or not a simple whitespace tokeniser can perform just as well as a full tokeniser, and in this case we used Unitok 5 .",
"N-grams We compared word-level uni-grams and bi-grams separately and in combination.",
"SVR parameters We tested different penalty parameters C and different epsilon parameters of the SVR.",
"Word Replacements We tested replacements to see if generalising words by inserting special tokens would help to reduce the sparsity problem.",
"We placed the word replacements into three separate groups: 1.",
"Company -When a company was mentioned in the input headline from the list of companies in the training data marked up as aspects, it was replaced by a company special token.",
"2.",
"Positive -When a positive word was mentioned in the input headline from a list of positive words (which was created using the N most similar words based on cosine distance) to 'excellent' using the pre-trained word2vec model.",
"3.",
"Negative -The same as the positive group however the word used was 'poor' instead of 'excellent'.",
"In the positive and negative groups, we chose the words 'excellent' and 'poor' following Turney (2002) to group the terms together under nondomain specific sentiment words.",
"Target aspect In order to incorporated the company as an aspect, we employed a boolean vector to represent the sentiment of the sentence.",
"This was done in order to see if the system could better differentiate the sentiment when the sentence was the same but the company was different.",
"BLSTM We created two different Bidirectional (Graves and Schmidhuber, 2005 ) Long Short-Term Memory (Hochreiter and Schmidhuber, 1997) using the Python Keras library (Chollet, 2015) with tensor flow backend (Abadi et al., 2016) .",
"We choose an LSTM model as it solves the vanishing gradients problem of Recurrent Neural Networks.",
"We used a bidirectional model as it allows us to capture information that came before and after instead of just before, thereby allowing us to capture more relevant context within the model.",
"Practically, a BLSTM is two LSTMs one going forward through the tokens the other in reverse order and in our models concatenating the resulting output vectors together at each time step.",
"The BLSTM models take as input a headline sentence of size L tokens 6 where L is the length of the longest sentence in the training texts.",
"Each word is converted into a 300 dimension vector using the word2vec model trained over the financial text 7 .",
"Any text that is not recognised by the word2vec model is represented as a vector of zeros; this is also used to pad out the sentence if it is shorter than L. Both BLSTM models have the following similar properties: 1.",
"Gradient clipping value of 5 -This was to help with the exploding gradients problem.",
"2.",
"Minimised the Mean Square Error (MSE) loss using RMSprop with a mini batch size of 32.",
"The output activation function is linear.",
"The main difference between the two models is the use of drop out and when they stop training over the data (epoch).",
"Both models architectures can be seen in figure 1.",
"Standard LSTM (SLSTM) The BLSTMs do contain drop out in both the input and between the connections of 0.2 each.",
"Finally the epoch is fixed at 25.",
"Early LSTM (ELSTM) As can be seen from figure 1, the drop out of 0.5 only happens between the layers and not the 6 Tokenised by Unitok 7 See the following link for detailed implementation details https://github.com/apmoore1/semeval# finance-word2vec-model connections as in the SLSTM.",
"Also the epoch is not fixed, it uses early stopping with a patience of 10.",
"We expect that this model can generalise better than the standard one due to the higher drop out and that the epoch is based on early stopping which relies on a validation set to know when to stop training.",
"Results We first present our findings on the best performing parameters and features for the SVRs.",
"These were determined by cross validation (CV) scores on the provided training data set using cosine similarity as the evaluation metric.",
"8 We found that using uni-grams and bi-grams performs best and using only bi-grams to be the worst.",
"Using the Unitok tokeniser always performed better than simple whitespace tokenisation.",
"The binary presence of tokens over frequency did not alter performance.",
"The C parameter was tested for three values; 0.01, 0.1 and 1.",
"We found very little difference between 0.1 and 1, but 0.01 produced much poorer results.",
"The eplison parameter was tested for 0.001, 0.01 and 0.1 the performance did not differ much but the lower the higher the performance but the more likely to overfit.",
"Using word replacements was effective for all three types (company, positive and negative) but using a value N=10 performed best for both positive and negative words.",
"Using target aspects also improved results.",
"Therefore, the best SVR model comprised of: Unitok tokenisation, uni-and bi-grams, word representation, C=0.1, eplison=0.01, company, positive, and negative word replacements and target aspects.",
"The main evaluation over the test data is based on the best performing SVR and the two BLSTM models once trained on all of the training data.",
"The result table 1 shows three columns based on the three evaluation metrics that the organisers have used.",
"Metric 1 is the original metric, weighted cosine similarity (the metric used to evaluate the final version of the results, where we were ranked 5th; metric provided on the task website 9 ).",
"This was then changed after the evaluation deadline to equation 1 10 (which we term metric 2; this is what the first version of the results were actually based on, where we were ranked 4th), which then changed by the organisers to their equation as presented in Cortis et al.",
"(2017) (which we term metric 3 and what the second version of the results were based on, where we were ranked 5th).",
"Model Metric 1 As you can see from the results table 1, the difference between the metrics is quite substantial.",
"This is due to the system's optimisation being based on metric 1 rather than 2.",
"Metric 2 is a classification metric for sentences with one aspect as it penalises values that are of opposite sign (giving -1 score) and rewards values with the same sign (giving +1 score).",
"Our systems are not optimised for this because it would predict scores of -0.01 and true value of 0.01 as very close (within vector of other results) with low error whereas metric 2 would give this the highest error rating of -1 as they are not the same sign.",
"Metric 3 is more similar to metric 1 as shown by the results, however the crucial difference is that again if you get opposite signs it will penalise more.",
"We analysed the top 50 errors based on Mean Absolute Error (MAE) in the test dataset specifically to examine the number of sentences containing more than one aspect.",
"Our investigation shows that no one system is better at predicting the sentiment of sentences that have more than one aspect (i.e.",
"company) within them.",
"Within those top 50 errors we found that the BLSTM systems do not know which parts of the sentence are associated to the company the sentiment is with respect to.",
"Also they do not know the strength/existence of certain sentiment words.",
"Conclusion and Future Work In this short paper, we have described our implemented solutions to SemEval Task 5 track 2, utilising both SVR and BLSTM approaches.",
"Our results show an improvement of around 5% when using LSTM models relative to SVR.",
"We have shown that this task can be partially represented as an aspect based sentiment task on a domain specific problem.",
"In general, our approaches acted as sentence level classifiers as they take no target company into consideration.",
"As our results show, the choice of evaluation metric makes a great deal of difference to system training and testing.",
"Future work will be to implement aspect specific information into an LSTM model as it has been shown to be useful in other work (Wang et al., 2016) ."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"4",
"4.1",
"4.1.1",
"4.1.2",
"4.1.3",
"4.1.4",
"4.1.5",
"4.2",
"3.",
"4.2.1",
"4.2.2",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Data",
"System description",
"SVR",
"Tokenisation",
"N-grams",
"SVR parameters",
"Word Replacements",
"Target aspect",
"BLSTM",
"The output activation function is linear.",
"Standard LSTM (SLSTM)",
"Early LSTM (ELSTM)",
"Results",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-6#paper-970#slide-2 | Pre Processing and Additional data used | Used 189, 206 financial articles (e.g. Financial Times) that were
manually downloaded from Factiva1 to create a Word2Vec model [5]2.
These were created using Gensim3. | Used 189, 206 financial articles (e.g. Financial Times) that were
manually downloaded from Factiva1 to create a Word2Vec model [5]2.
These were created using Gensim3. | [] |
GEM-SciDuet-train-6#paper-970#slide-3 | 970 | Lancaster A at SemEval-2017 Task 5: Evaluation metrics matter: predicting sentiment from financial news headlines | This paper describes our participation in Task 5 track 2 of SemEval 2017 to predict the sentiment of financial news headlines for a specific company on a continuous scale between -1 and 1. We tackled the problem using a number of approaches, utilising a Support Vector Regression (SVR) and a Bidirectional Long Short-Term Memory (BLSTM). We found an improvement of 4-6% using the LSTM model over the SVR and came fourth in the track. We report a number of different evaluations using a finance specific word embedding model and reflect on the effects of using different evaluation metrics. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95
],
"paper_content_text": [
"Introduction The objective of Task 5 Track 2 of SemEval (2017) was to predict the sentiment of news headlines with respect to companies mentioned within the headlines.",
"This task can be seen as a financespecific aspect-based sentiment task (Nasukawa and Yi, 2003) .",
"The main motivations of this task is to find specific features and learning algorithms that will perform better for this domain as aspect based sentiment analysis tasks have been conducted before at SemEval (Pontiki et al., 2014) .",
"Domain specific terminology is expected to play a key part in this task, as reporters, investors and analysts in the financial domain will use a specific set of terminology when discussing financial performance.",
"Potentially, this may also vary across different financial domains and industry sectors.",
"Therefore, we took an exploratory approach and investigated how various features and learning algorithms perform differently, specifically SVR and BLSTMs.",
"We found that BLSTMs outperform an SVR without having any knowledge of the company that the sentiment is with respect to.",
"For replicability purposes, with this paper we are releasing our source code 1 and the finance specific BLSTM word embedding model 2 .",
"Related Work There is a growing amount of research being carried out related to sentiment analysis within the financial domain.",
"This work ranges from domainspecific lexicons (Loughran and McDonald, 2011) and lexicon creation (Moore et al., 2016) to stock market prediction models (Peng and Jiang, 2016; Kazemian et al., 2016) .",
"Peng and Jiang (2016) used a multi layer neural network to predict the stock market and found that incorporating textual features from financial news can improve the accuracy of prediction.",
"Kazemian et al.",
"(2016) showed the importance of tuning sentiment analysis to the task of stock market prediction.",
"However, much of the previous work was based on numerical financial stock market data rather than on aspect level financial textual data.",
"In aspect based sentiment analysis, there have been many different techniques used to predict the polarity of an aspect as shown in SemEval-2016 task 5 (Pontiki et al., 2014 ).",
"The winning system (Brun et al., 2016 ) used many different linguistic features and an ensemble model, and the runner up (Kumar et al., 2016) used uni-grams, bi-grams and sentiment lexicons as features for a Support Vector Machine (SVM).",
"Deep learning methods have also been applied to aspect polarity prediction.",
"Ruder et al.",
"(2016) created a hierarchical BLSTM with a sentence level BLSTM inputting into a review level BLSTM thus allowing them to take into account inter-and intra-sentence context.",
"They used only word embeddings making their system less dependent on extensive feature engineering or manual feature creation.",
"This system outperformed all others on certain languages on the SemEval-2016 task 5 dataset (Pontiki et al., 2014) and on other languages performed close to the best systems.",
"Wang et al.",
"(2016) also created an LSTM based model using word embeddings but instead of a hierarchical model it was a one layered LSTM with attention which puts more emphasis on learning the sentiment of words specific to a given aspect.",
"Data The training data published by the organisers for this track was a set of headline sentences from financial news articles where each sentence was tagged with the company name (which we treat as the aspect) and the polarity of the sentence with respect to the company.",
"There is the possibility that the same sentence occurs more than once if there is more than one company mentioned.",
"The polarity was a real value between -1 (negative sentiment) and 1 (positive sentiment).",
"We additionally trained a word2vec (Mikolov et al., 2013) word embedding model 3 on a set of 189,206 financial articles containing 161,877,425 tokens, that were manually downloaded from Factiva 4 .",
"The articles stem from a range of sources including the Financial Times and relate to companies from the United States only.",
"We trained the model on domain specific data as it has been shown many times that the financial domain can contain very different language.",
"System description Even though we have outlined this task as an aspect based sentiment task, this is instantiated in only one of the features in the SVR.",
"The following two subsections describe the two approaches, first SVR and then BLSTM.",
"Key implementation details are exposed here in the paper, but we have released the source code and word embedding models to aid replicability and further experimentation.",
"SVR The system was created using ScitKit learn (Pedregosa et al., 2011) linear Support Vector Regression model (Drucker et al., 1997) .",
"We exper-imented with the following different features and parameter settings: Tokenisation For comparison purposes, we tested whether or not a simple whitespace tokeniser can perform just as well as a full tokeniser, and in this case we used Unitok 5 .",
"N-grams We compared word-level uni-grams and bi-grams separately and in combination.",
"SVR parameters We tested different penalty parameters C and different epsilon parameters of the SVR.",
"Word Replacements We tested replacements to see if generalising words by inserting special tokens would help to reduce the sparsity problem.",
"We placed the word replacements into three separate groups: 1.",
"Company -When a company was mentioned in the input headline from the list of companies in the training data marked up as aspects, it was replaced by a company special token.",
"2.",
"Positive -When a positive word was mentioned in the input headline from a list of positive words (which was created using the N most similar words based on cosine distance) to 'excellent' using the pre-trained word2vec model.",
"3.",
"Negative -The same as the positive group however the word used was 'poor' instead of 'excellent'.",
"In the positive and negative groups, we chose the words 'excellent' and 'poor' following Turney (2002) to group the terms together under nondomain specific sentiment words.",
"Target aspect In order to incorporated the company as an aspect, we employed a boolean vector to represent the sentiment of the sentence.",
"This was done in order to see if the system could better differentiate the sentiment when the sentence was the same but the company was different.",
"BLSTM We created two different Bidirectional (Graves and Schmidhuber, 2005 ) Long Short-Term Memory (Hochreiter and Schmidhuber, 1997) using the Python Keras library (Chollet, 2015) with tensor flow backend (Abadi et al., 2016) .",
"We choose an LSTM model as it solves the vanishing gradients problem of Recurrent Neural Networks.",
"We used a bidirectional model as it allows us to capture information that came before and after instead of just before, thereby allowing us to capture more relevant context within the model.",
"Practically, a BLSTM is two LSTMs one going forward through the tokens the other in reverse order and in our models concatenating the resulting output vectors together at each time step.",
"The BLSTM models take as input a headline sentence of size L tokens 6 where L is the length of the longest sentence in the training texts.",
"Each word is converted into a 300 dimension vector using the word2vec model trained over the financial text 7 .",
"Any text that is not recognised by the word2vec model is represented as a vector of zeros; this is also used to pad out the sentence if it is shorter than L. Both BLSTM models have the following similar properties: 1.",
"Gradient clipping value of 5 -This was to help with the exploding gradients problem.",
"2.",
"Minimised the Mean Square Error (MSE) loss using RMSprop with a mini batch size of 32.",
"The output activation function is linear.",
"The main difference between the two models is the use of drop out and when they stop training over the data (epoch).",
"Both models architectures can be seen in figure 1.",
"Standard LSTM (SLSTM) The BLSTMs do contain drop out in both the input and between the connections of 0.2 each.",
"Finally the epoch is fixed at 25.",
"Early LSTM (ELSTM) As can be seen from figure 1, the drop out of 0.5 only happens between the layers and not the 6 Tokenised by Unitok 7 See the following link for detailed implementation details https://github.com/apmoore1/semeval# finance-word2vec-model connections as in the SLSTM.",
"Also the epoch is not fixed, it uses early stopping with a patience of 10.",
"We expect that this model can generalise better than the standard one due to the higher drop out and that the epoch is based on early stopping which relies on a validation set to know when to stop training.",
"Results We first present our findings on the best performing parameters and features for the SVRs.",
"These were determined by cross validation (CV) scores on the provided training data set using cosine similarity as the evaluation metric.",
"8 We found that using uni-grams and bi-grams performs best and using only bi-grams to be the worst.",
"Using the Unitok tokeniser always performed better than simple whitespace tokenisation.",
"The binary presence of tokens over frequency did not alter performance.",
"The C parameter was tested for three values; 0.01, 0.1 and 1.",
"We found very little difference between 0.1 and 1, but 0.01 produced much poorer results.",
"The eplison parameter was tested for 0.001, 0.01 and 0.1 the performance did not differ much but the lower the higher the performance but the more likely to overfit.",
"Using word replacements was effective for all three types (company, positive and negative) but using a value N=10 performed best for both positive and negative words.",
"Using target aspects also improved results.",
"Therefore, the best SVR model comprised of: Unitok tokenisation, uni-and bi-grams, word representation, C=0.1, eplison=0.01, company, positive, and negative word replacements and target aspects.",
"The main evaluation over the test data is based on the best performing SVR and the two BLSTM models once trained on all of the training data.",
"The result table 1 shows three columns based on the three evaluation metrics that the organisers have used.",
"Metric 1 is the original metric, weighted cosine similarity (the metric used to evaluate the final version of the results, where we were ranked 5th; metric provided on the task website 9 ).",
"This was then changed after the evaluation deadline to equation 1 10 (which we term metric 2; this is what the first version of the results were actually based on, where we were ranked 4th), which then changed by the organisers to their equation as presented in Cortis et al.",
"(2017) (which we term metric 3 and what the second version of the results were based on, where we were ranked 5th).",
"Model Metric 1 As you can see from the results table 1, the difference between the metrics is quite substantial.",
"This is due to the system's optimisation being based on metric 1 rather than 2.",
"Metric 2 is a classification metric for sentences with one aspect as it penalises values that are of opposite sign (giving -1 score) and rewards values with the same sign (giving +1 score).",
"Our systems are not optimised for this because it would predict scores of -0.01 and true value of 0.01 as very close (within vector of other results) with low error whereas metric 2 would give this the highest error rating of -1 as they are not the same sign.",
"Metric 3 is more similar to metric 1 as shown by the results, however the crucial difference is that again if you get opposite signs it will penalise more.",
"We analysed the top 50 errors based on Mean Absolute Error (MAE) in the test dataset specifically to examine the number of sentences containing more than one aspect.",
"Our investigation shows that no one system is better at predicting the sentiment of sentences that have more than one aspect (i.e.",
"company) within them.",
"Within those top 50 errors we found that the BLSTM systems do not know which parts of the sentence are associated to the company the sentiment is with respect to.",
"Also they do not know the strength/existence of certain sentiment words.",
"Conclusion and Future Work In this short paper, we have described our implemented solutions to SemEval Task 5 track 2, utilising both SVR and BLSTM approaches.",
"Our results show an improvement of around 5% when using LSTM models relative to SVR.",
"We have shown that this task can be partially represented as an aspect based sentiment task on a domain specific problem.",
"In general, our approaches acted as sentence level classifiers as they take no target company into consideration.",
"As our results show, the choice of evaluation metric makes a great deal of difference to system training and testing.",
"Future work will be to implement aspect specific information into an LSTM model as it has been shown to be useful in other work (Wang et al., 2016) ."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"4",
"4.1",
"4.1.1",
"4.1.2",
"4.1.3",
"4.1.4",
"4.1.5",
"4.2",
"3.",
"4.2.1",
"4.2.2",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Data",
"System description",
"SVR",
"Tokenisation",
"N-grams",
"SVR parameters",
"Word Replacements",
"Target aspect",
"BLSTM",
"The output activation function is linear.",
"Standard LSTM (SLSTM)",
"Early LSTM (ELSTM)",
"Results",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-6#paper-970#slide-3 | Support Vector Regression SVR 1 | Features and settings that we changed
1. Tokenisation - Whitespace or Unitok4
2. N-grams - uni-grams, bi-grams and both.
3. SVR settings - penalty parameter C and epsilon parameter. | Features and settings that we changed
1. Tokenisation - Whitespace or Unitok4
2. N-grams - uni-grams, bi-grams and both.
3. SVR settings - penalty parameter C and epsilon parameter. | [] |
GEM-SciDuet-train-6#paper-970#slide-4 | 970 | Lancaster A at SemEval-2017 Task 5: Evaluation metrics matter: predicting sentiment from financial news headlines | This paper describes our participation in Task 5 track 2 of SemEval 2017 to predict the sentiment of financial news headlines for a specific company on a continuous scale between -1 and 1. We tackled the problem using a number of approaches, utilising a Support Vector Regression (SVR) and a Bidirectional Long Short-Term Memory (BLSTM). We found an improvement of 4-6% using the LSTM model over the SVR and came fourth in the track. We report a number of different evaluations using a finance specific word embedding model and reflect on the effects of using different evaluation metrics. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95
],
"paper_content_text": [
"Introduction The objective of Task 5 Track 2 of SemEval (2017) was to predict the sentiment of news headlines with respect to companies mentioned within the headlines.",
"This task can be seen as a financespecific aspect-based sentiment task (Nasukawa and Yi, 2003) .",
"The main motivations of this task is to find specific features and learning algorithms that will perform better for this domain as aspect based sentiment analysis tasks have been conducted before at SemEval (Pontiki et al., 2014) .",
"Domain specific terminology is expected to play a key part in this task, as reporters, investors and analysts in the financial domain will use a specific set of terminology when discussing financial performance.",
"Potentially, this may also vary across different financial domains and industry sectors.",
"Therefore, we took an exploratory approach and investigated how various features and learning algorithms perform differently, specifically SVR and BLSTMs.",
"We found that BLSTMs outperform an SVR without having any knowledge of the company that the sentiment is with respect to.",
"For replicability purposes, with this paper we are releasing our source code 1 and the finance specific BLSTM word embedding model 2 .",
"Related Work There is a growing amount of research being carried out related to sentiment analysis within the financial domain.",
"This work ranges from domainspecific lexicons (Loughran and McDonald, 2011) and lexicon creation (Moore et al., 2016) to stock market prediction models (Peng and Jiang, 2016; Kazemian et al., 2016) .",
"Peng and Jiang (2016) used a multi layer neural network to predict the stock market and found that incorporating textual features from financial news can improve the accuracy of prediction.",
"Kazemian et al.",
"(2016) showed the importance of tuning sentiment analysis to the task of stock market prediction.",
"However, much of the previous work was based on numerical financial stock market data rather than on aspect level financial textual data.",
"In aspect based sentiment analysis, there have been many different techniques used to predict the polarity of an aspect as shown in SemEval-2016 task 5 (Pontiki et al., 2014 ).",
"The winning system (Brun et al., 2016 ) used many different linguistic features and an ensemble model, and the runner up (Kumar et al., 2016) used uni-grams, bi-grams and sentiment lexicons as features for a Support Vector Machine (SVM).",
"Deep learning methods have also been applied to aspect polarity prediction.",
"Ruder et al.",
"(2016) created a hierarchical BLSTM with a sentence level BLSTM inputting into a review level BLSTM thus allowing them to take into account inter-and intra-sentence context.",
"They used only word embeddings making their system less dependent on extensive feature engineering or manual feature creation.",
"This system outperformed all others on certain languages on the SemEval-2016 task 5 dataset (Pontiki et al., 2014) and on other languages performed close to the best systems.",
"Wang et al.",
"(2016) also created an LSTM based model using word embeddings but instead of a hierarchical model it was a one layered LSTM with attention which puts more emphasis on learning the sentiment of words specific to a given aspect.",
"Data The training data published by the organisers for this track was a set of headline sentences from financial news articles where each sentence was tagged with the company name (which we treat as the aspect) and the polarity of the sentence with respect to the company.",
"There is the possibility that the same sentence occurs more than once if there is more than one company mentioned.",
"The polarity was a real value between -1 (negative sentiment) and 1 (positive sentiment).",
"We additionally trained a word2vec (Mikolov et al., 2013) word embedding model 3 on a set of 189,206 financial articles containing 161,877,425 tokens, that were manually downloaded from Factiva 4 .",
"The articles stem from a range of sources including the Financial Times and relate to companies from the United States only.",
"We trained the model on domain specific data as it has been shown many times that the financial domain can contain very different language.",
"System description Even though we have outlined this task as an aspect based sentiment task, this is instantiated in only one of the features in the SVR.",
"The following two subsections describe the two approaches, first SVR and then BLSTM.",
"Key implementation details are exposed here in the paper, but we have released the source code and word embedding models to aid replicability and further experimentation.",
"SVR The system was created using ScitKit learn (Pedregosa et al., 2011) linear Support Vector Regression model (Drucker et al., 1997) .",
"We exper-imented with the following different features and parameter settings: Tokenisation For comparison purposes, we tested whether or not a simple whitespace tokeniser can perform just as well as a full tokeniser, and in this case we used Unitok 5 .",
"N-grams We compared word-level uni-grams and bi-grams separately and in combination.",
"SVR parameters We tested different penalty parameters C and different epsilon parameters of the SVR.",
"Word Replacements We tested replacements to see if generalising words by inserting special tokens would help to reduce the sparsity problem.",
"We placed the word replacements into three separate groups: 1.",
"Company -When a company was mentioned in the input headline from the list of companies in the training data marked up as aspects, it was replaced by a company special token.",
"2.",
"Positive -When a positive word was mentioned in the input headline from a list of positive words (which was created using the N most similar words based on cosine distance) to 'excellent' using the pre-trained word2vec model.",
"3.",
"Negative -The same as the positive group however the word used was 'poor' instead of 'excellent'.",
"In the positive and negative groups, we chose the words 'excellent' and 'poor' following Turney (2002) to group the terms together under nondomain specific sentiment words.",
"Target aspect In order to incorporated the company as an aspect, we employed a boolean vector to represent the sentiment of the sentence.",
"This was done in order to see if the system could better differentiate the sentiment when the sentence was the same but the company was different.",
"BLSTM We created two different Bidirectional (Graves and Schmidhuber, 2005 ) Long Short-Term Memory (Hochreiter and Schmidhuber, 1997) using the Python Keras library (Chollet, 2015) with tensor flow backend (Abadi et al., 2016) .",
"We choose an LSTM model as it solves the vanishing gradients problem of Recurrent Neural Networks.",
"We used a bidirectional model as it allows us to capture information that came before and after instead of just before, thereby allowing us to capture more relevant context within the model.",
"Practically, a BLSTM is two LSTMs one going forward through the tokens the other in reverse order and in our models concatenating the resulting output vectors together at each time step.",
"The BLSTM models take as input a headline sentence of size L tokens 6 where L is the length of the longest sentence in the training texts.",
"Each word is converted into a 300 dimension vector using the word2vec model trained over the financial text 7 .",
"Any text that is not recognised by the word2vec model is represented as a vector of zeros; this is also used to pad out the sentence if it is shorter than L. Both BLSTM models have the following similar properties: 1.",
"Gradient clipping value of 5 -This was to help with the exploding gradients problem.",
"2.",
"Minimised the Mean Square Error (MSE) loss using RMSprop with a mini batch size of 32.",
"The output activation function is linear.",
"The main difference between the two models is the use of drop out and when they stop training over the data (epoch).",
"Both models architectures can be seen in figure 1.",
"Standard LSTM (SLSTM) The BLSTMs do contain drop out in both the input and between the connections of 0.2 each.",
"Finally the epoch is fixed at 25.",
"Early LSTM (ELSTM) As can be seen from figure 1, the drop out of 0.5 only happens between the layers and not the 6 Tokenised by Unitok 7 See the following link for detailed implementation details https://github.com/apmoore1/semeval# finance-word2vec-model connections as in the SLSTM.",
"Also the epoch is not fixed, it uses early stopping with a patience of 10.",
"We expect that this model can generalise better than the standard one due to the higher drop out and that the epoch is based on early stopping which relies on a validation set to know when to stop training.",
"Results We first present our findings on the best performing parameters and features for the SVRs.",
"These were determined by cross validation (CV) scores on the provided training data set using cosine similarity as the evaluation metric.",
"8 We found that using uni-grams and bi-grams performs best and using only bi-grams to be the worst.",
"Using the Unitok tokeniser always performed better than simple whitespace tokenisation.",
"The binary presence of tokens over frequency did not alter performance.",
"The C parameter was tested for three values; 0.01, 0.1 and 1.",
"We found very little difference between 0.1 and 1, but 0.01 produced much poorer results.",
"The eplison parameter was tested for 0.001, 0.01 and 0.1 the performance did not differ much but the lower the higher the performance but the more likely to overfit.",
"Using word replacements was effective for all three types (company, positive and negative) but using a value N=10 performed best for both positive and negative words.",
"Using target aspects also improved results.",
"Therefore, the best SVR model comprised of: Unitok tokenisation, uni-and bi-grams, word representation, C=0.1, eplison=0.01, company, positive, and negative word replacements and target aspects.",
"The main evaluation over the test data is based on the best performing SVR and the two BLSTM models once trained on all of the training data.",
"The result table 1 shows three columns based on the three evaluation metrics that the organisers have used.",
"Metric 1 is the original metric, weighted cosine similarity (the metric used to evaluate the final version of the results, where we were ranked 5th; metric provided on the task website 9 ).",
"This was then changed after the evaluation deadline to equation 1 10 (which we term metric 2; this is what the first version of the results were actually based on, where we were ranked 4th), which then changed by the organisers to their equation as presented in Cortis et al.",
"(2017) (which we term metric 3 and what the second version of the results were based on, where we were ranked 5th).",
"Model Metric 1 As you can see from the results table 1, the difference between the metrics is quite substantial.",
"This is due to the system's optimisation being based on metric 1 rather than 2.",
"Metric 2 is a classification metric for sentences with one aspect as it penalises values that are of opposite sign (giving -1 score) and rewards values with the same sign (giving +1 score).",
"Our systems are not optimised for this because it would predict scores of -0.01 and true value of 0.01 as very close (within vector of other results) with low error whereas metric 2 would give this the highest error rating of -1 as they are not the same sign.",
"Metric 3 is more similar to metric 1 as shown by the results, however the crucial difference is that again if you get opposite signs it will penalise more.",
"We analysed the top 50 errors based on Mean Absolute Error (MAE) in the test dataset specifically to examine the number of sentences containing more than one aspect.",
"Our investigation shows that no one system is better at predicting the sentiment of sentences that have more than one aspect (i.e.",
"company) within them.",
"Within those top 50 errors we found that the BLSTM systems do not know which parts of the sentence are associated to the company the sentiment is with respect to.",
"Also they do not know the strength/existence of certain sentiment words.",
"Conclusion and Future Work In this short paper, we have described our implemented solutions to SemEval Task 5 track 2, utilising both SVR and BLSTM approaches.",
"Our results show an improvement of around 5% when using LSTM models relative to SVR.",
"We have shown that this task can be partially represented as an aspect based sentiment task on a domain specific problem.",
"In general, our approaches acted as sentence level classifiers as they take no target company into consideration.",
"As our results show, the choice of evaluation metric makes a great deal of difference to system training and testing.",
"Future work will be to implement aspect specific information into an LSTM model as it has been shown to be useful in other work (Wang et al., 2016) ."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"4",
"4.1",
"4.1.1",
"4.1.2",
"4.1.3",
"4.1.4",
"4.1.5",
"4.2",
"3.",
"4.2.1",
"4.2.2",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Data",
"System description",
"SVR",
"Tokenisation",
"N-grams",
"SVR parameters",
"Word Replacements",
"Target aspect",
"BLSTM",
"The output activation function is linear.",
"Standard LSTM (SLSTM)",
"Early LSTM (ELSTM)",
"Results",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-6#paper-970#slide-4 | Word Replacements | AstraZeneca PLC had an improved performance where as Dixons
companyname had an posword performance where as companyname | AstraZeneca PLC had an improved performance where as Dixons
companyname had an posword performance where as companyname | [] |
GEM-SciDuet-train-6#paper-970#slide-5 | 970 | Lancaster A at SemEval-2017 Task 5: Evaluation metrics matter: predicting sentiment from financial news headlines | This paper describes our participation in Task 5 track 2 of SemEval 2017 to predict the sentiment of financial news headlines for a specific company on a continuous scale between -1 and 1. We tackled the problem using a number of approaches, utilising a Support Vector Regression (SVR) and a Bidirectional Long Short-Term Memory (BLSTM). We found an improvement of 4-6% using the LSTM model over the SVR and came fourth in the track. We report a number of different evaluations using a finance specific word embedding model and reflect on the effects of using different evaluation metrics. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95
],
"paper_content_text": [
"Introduction The objective of Task 5 Track 2 of SemEval (2017) was to predict the sentiment of news headlines with respect to companies mentioned within the headlines.",
"This task can be seen as a financespecific aspect-based sentiment task (Nasukawa and Yi, 2003) .",
"The main motivations of this task is to find specific features and learning algorithms that will perform better for this domain as aspect based sentiment analysis tasks have been conducted before at SemEval (Pontiki et al., 2014) .",
"Domain specific terminology is expected to play a key part in this task, as reporters, investors and analysts in the financial domain will use a specific set of terminology when discussing financial performance.",
"Potentially, this may also vary across different financial domains and industry sectors.",
"Therefore, we took an exploratory approach and investigated how various features and learning algorithms perform differently, specifically SVR and BLSTMs.",
"We found that BLSTMs outperform an SVR without having any knowledge of the company that the sentiment is with respect to.",
"For replicability purposes, with this paper we are releasing our source code 1 and the finance specific BLSTM word embedding model 2 .",
"Related Work There is a growing amount of research being carried out related to sentiment analysis within the financial domain.",
"This work ranges from domainspecific lexicons (Loughran and McDonald, 2011) and lexicon creation (Moore et al., 2016) to stock market prediction models (Peng and Jiang, 2016; Kazemian et al., 2016) .",
"Peng and Jiang (2016) used a multi layer neural network to predict the stock market and found that incorporating textual features from financial news can improve the accuracy of prediction.",
"Kazemian et al.",
"(2016) showed the importance of tuning sentiment analysis to the task of stock market prediction.",
"However, much of the previous work was based on numerical financial stock market data rather than on aspect level financial textual data.",
"In aspect based sentiment analysis, there have been many different techniques used to predict the polarity of an aspect as shown in SemEval-2016 task 5 (Pontiki et al., 2014 ).",
"The winning system (Brun et al., 2016 ) used many different linguistic features and an ensemble model, and the runner up (Kumar et al., 2016) used uni-grams, bi-grams and sentiment lexicons as features for a Support Vector Machine (SVM).",
"Deep learning methods have also been applied to aspect polarity prediction.",
"Ruder et al.",
"(2016) created a hierarchical BLSTM with a sentence level BLSTM inputting into a review level BLSTM thus allowing them to take into account inter-and intra-sentence context.",
"They used only word embeddings making their system less dependent on extensive feature engineering or manual feature creation.",
"This system outperformed all others on certain languages on the SemEval-2016 task 5 dataset (Pontiki et al., 2014) and on other languages performed close to the best systems.",
"Wang et al.",
"(2016) also created an LSTM based model using word embeddings but instead of a hierarchical model it was a one layered LSTM with attention which puts more emphasis on learning the sentiment of words specific to a given aspect.",
"Data The training data published by the organisers for this track was a set of headline sentences from financial news articles where each sentence was tagged with the company name (which we treat as the aspect) and the polarity of the sentence with respect to the company.",
"There is the possibility that the same sentence occurs more than once if there is more than one company mentioned.",
"The polarity was a real value between -1 (negative sentiment) and 1 (positive sentiment).",
"We additionally trained a word2vec (Mikolov et al., 2013) word embedding model 3 on a set of 189,206 financial articles containing 161,877,425 tokens, that were manually downloaded from Factiva 4 .",
"The articles stem from a range of sources including the Financial Times and relate to companies from the United States only.",
"We trained the model on domain specific data as it has been shown many times that the financial domain can contain very different language.",
"System description Even though we have outlined this task as an aspect based sentiment task, this is instantiated in only one of the features in the SVR.",
"The following two subsections describe the two approaches, first SVR and then BLSTM.",
"Key implementation details are exposed here in the paper, but we have released the source code and word embedding models to aid replicability and further experimentation.",
"SVR The system was created using ScitKit learn (Pedregosa et al., 2011) linear Support Vector Regression model (Drucker et al., 1997) .",
"We exper-imented with the following different features and parameter settings: Tokenisation For comparison purposes, we tested whether or not a simple whitespace tokeniser can perform just as well as a full tokeniser, and in this case we used Unitok 5 .",
"N-grams We compared word-level uni-grams and bi-grams separately and in combination.",
"SVR parameters We tested different penalty parameters C and different epsilon parameters of the SVR.",
"Word Replacements We tested replacements to see if generalising words by inserting special tokens would help to reduce the sparsity problem.",
"We placed the word replacements into three separate groups: 1.",
"Company -When a company was mentioned in the input headline from the list of companies in the training data marked up as aspects, it was replaced by a company special token.",
"2.",
"Positive -When a positive word was mentioned in the input headline from a list of positive words (which was created using the N most similar words based on cosine distance) to 'excellent' using the pre-trained word2vec model.",
"3.",
"Negative -The same as the positive group however the word used was 'poor' instead of 'excellent'.",
"In the positive and negative groups, we chose the words 'excellent' and 'poor' following Turney (2002) to group the terms together under nondomain specific sentiment words.",
"Target aspect In order to incorporated the company as an aspect, we employed a boolean vector to represent the sentiment of the sentence.",
"This was done in order to see if the system could better differentiate the sentiment when the sentence was the same but the company was different.",
"BLSTM We created two different Bidirectional (Graves and Schmidhuber, 2005 ) Long Short-Term Memory (Hochreiter and Schmidhuber, 1997) using the Python Keras library (Chollet, 2015) with tensor flow backend (Abadi et al., 2016) .",
"We choose an LSTM model as it solves the vanishing gradients problem of Recurrent Neural Networks.",
"We used a bidirectional model as it allows us to capture information that came before and after instead of just before, thereby allowing us to capture more relevant context within the model.",
"Practically, a BLSTM is two LSTMs one going forward through the tokens the other in reverse order and in our models concatenating the resulting output vectors together at each time step.",
"The BLSTM models take as input a headline sentence of size L tokens 6 where L is the length of the longest sentence in the training texts.",
"Each word is converted into a 300 dimension vector using the word2vec model trained over the financial text 7 .",
"Any text that is not recognised by the word2vec model is represented as a vector of zeros; this is also used to pad out the sentence if it is shorter than L. Both BLSTM models have the following similar properties: 1.",
"Gradient clipping value of 5 -This was to help with the exploding gradients problem.",
"2.",
"Minimised the Mean Square Error (MSE) loss using RMSprop with a mini batch size of 32.",
"The output activation function is linear.",
"The main difference between the two models is the use of drop out and when they stop training over the data (epoch).",
"Both models architectures can be seen in figure 1.",
"Standard LSTM (SLSTM) The BLSTMs do contain drop out in both the input and between the connections of 0.2 each.",
"Finally the epoch is fixed at 25.",
"Early LSTM (ELSTM) As can be seen from figure 1, the drop out of 0.5 only happens between the layers and not the 6 Tokenised by Unitok 7 See the following link for detailed implementation details https://github.com/apmoore1/semeval# finance-word2vec-model connections as in the SLSTM.",
"Also the epoch is not fixed, it uses early stopping with a patience of 10.",
"We expect that this model can generalise better than the standard one due to the higher drop out and that the epoch is based on early stopping which relies on a validation set to know when to stop training.",
"Results We first present our findings on the best performing parameters and features for the SVRs.",
"These were determined by cross validation (CV) scores on the provided training data set using cosine similarity as the evaluation metric.",
"8 We found that using uni-grams and bi-grams performs best and using only bi-grams to be the worst.",
"Using the Unitok tokeniser always performed better than simple whitespace tokenisation.",
"The binary presence of tokens over frequency did not alter performance.",
"The C parameter was tested for three values; 0.01, 0.1 and 1.",
"We found very little difference between 0.1 and 1, but 0.01 produced much poorer results.",
"The eplison parameter was tested for 0.001, 0.01 and 0.1 the performance did not differ much but the lower the higher the performance but the more likely to overfit.",
"Using word replacements was effective for all three types (company, positive and negative) but using a value N=10 performed best for both positive and negative words.",
"Using target aspects also improved results.",
"Therefore, the best SVR model comprised of: Unitok tokenisation, uni-and bi-grams, word representation, C=0.1, eplison=0.01, company, positive, and negative word replacements and target aspects.",
"The main evaluation over the test data is based on the best performing SVR and the two BLSTM models once trained on all of the training data.",
"The result table 1 shows three columns based on the three evaluation metrics that the organisers have used.",
"Metric 1 is the original metric, weighted cosine similarity (the metric used to evaluate the final version of the results, where we were ranked 5th; metric provided on the task website 9 ).",
"This was then changed after the evaluation deadline to equation 1 10 (which we term metric 2; this is what the first version of the results were actually based on, where we were ranked 4th), which then changed by the organisers to their equation as presented in Cortis et al.",
"(2017) (which we term metric 3 and what the second version of the results were based on, where we were ranked 5th).",
"Model Metric 1 As you can see from the results table 1, the difference between the metrics is quite substantial.",
"This is due to the system's optimisation being based on metric 1 rather than 2.",
"Metric 2 is a classification metric for sentences with one aspect as it penalises values that are of opposite sign (giving -1 score) and rewards values with the same sign (giving +1 score).",
"Our systems are not optimised for this because it would predict scores of -0.01 and true value of 0.01 as very close (within vector of other results) with low error whereas metric 2 would give this the highest error rating of -1 as they are not the same sign.",
"Metric 3 is more similar to metric 1 as shown by the results, however the crucial difference is that again if you get opposite signs it will penalise more.",
"We analysed the top 50 errors based on Mean Absolute Error (MAE) in the test dataset specifically to examine the number of sentences containing more than one aspect.",
"Our investigation shows that no one system is better at predicting the sentiment of sentences that have more than one aspect (i.e.",
"company) within them.",
"Within those top 50 errors we found that the BLSTM systems do not know which parts of the sentence are associated to the company the sentiment is with respect to.",
"Also they do not know the strength/existence of certain sentiment words.",
"Conclusion and Future Work In this short paper, we have described our implemented solutions to SemEval Task 5 track 2, utilising both SVR and BLSTM approaches.",
"Our results show an improvement of around 5% when using LSTM models relative to SVR.",
"We have shown that this task can be partially represented as an aspect based sentiment task on a domain specific problem.",
"In general, our approaches acted as sentence level classifiers as they take no target company into consideration.",
"As our results show, the choice of evaluation metric makes a great deal of difference to system training and testing.",
"Future work will be to implement aspect specific information into an LSTM model as it has been shown to be useful in other work (Wang et al., 2016) ."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"4",
"4.1",
"4.1.1",
"4.1.2",
"4.1.3",
"4.1.4",
"4.1.5",
"4.2",
"3.",
"4.2.1",
"4.2.2",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Data",
"System description",
"SVR",
"Tokenisation",
"N-grams",
"SVR parameters",
"Word Replacements",
"Target aspect",
"BLSTM",
"The output activation function is linear.",
"Standard LSTM (SLSTM)",
"Early LSTM (ELSTM)",
"Results",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-6#paper-970#slide-5 | Two BLSTM models | Drop out between layers
25 times trained over
Early stopping used to | Drop out between layers
25 times trained over
Early stopping used to | [] |
GEM-SciDuet-train-6#paper-970#slide-7 | 970 | Lancaster A at SemEval-2017 Task 5: Evaluation metrics matter: predicting sentiment from financial news headlines | This paper describes our participation in Task 5 track 2 of SemEval 2017 to predict the sentiment of financial news headlines for a specific company on a continuous scale between -1 and 1. We tackled the problem using a number of approaches, utilising a Support Vector Regression (SVR) and a Bidirectional Long Short-Term Memory (BLSTM). We found an improvement of 4-6% using the LSTM model over the SVR and came fourth in the track. We report a number of different evaluations using a finance specific word embedding model and reflect on the effects of using different evaluation metrics. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95
],
"paper_content_text": [
"Introduction The objective of Task 5 Track 2 of SemEval (2017) was to predict the sentiment of news headlines with respect to companies mentioned within the headlines.",
"This task can be seen as a financespecific aspect-based sentiment task (Nasukawa and Yi, 2003) .",
"The main motivations of this task is to find specific features and learning algorithms that will perform better for this domain as aspect based sentiment analysis tasks have been conducted before at SemEval (Pontiki et al., 2014) .",
"Domain specific terminology is expected to play a key part in this task, as reporters, investors and analysts in the financial domain will use a specific set of terminology when discussing financial performance.",
"Potentially, this may also vary across different financial domains and industry sectors.",
"Therefore, we took an exploratory approach and investigated how various features and learning algorithms perform differently, specifically SVR and BLSTMs.",
"We found that BLSTMs outperform an SVR without having any knowledge of the company that the sentiment is with respect to.",
"For replicability purposes, with this paper we are releasing our source code 1 and the finance specific BLSTM word embedding model 2 .",
"Related Work There is a growing amount of research being carried out related to sentiment analysis within the financial domain.",
"This work ranges from domainspecific lexicons (Loughran and McDonald, 2011) and lexicon creation (Moore et al., 2016) to stock market prediction models (Peng and Jiang, 2016; Kazemian et al., 2016) .",
"Peng and Jiang (2016) used a multi layer neural network to predict the stock market and found that incorporating textual features from financial news can improve the accuracy of prediction.",
"Kazemian et al.",
"(2016) showed the importance of tuning sentiment analysis to the task of stock market prediction.",
"However, much of the previous work was based on numerical financial stock market data rather than on aspect level financial textual data.",
"In aspect based sentiment analysis, there have been many different techniques used to predict the polarity of an aspect as shown in SemEval-2016 task 5 (Pontiki et al., 2014 ).",
"The winning system (Brun et al., 2016 ) used many different linguistic features and an ensemble model, and the runner up (Kumar et al., 2016) used uni-grams, bi-grams and sentiment lexicons as features for a Support Vector Machine (SVM).",
"Deep learning methods have also been applied to aspect polarity prediction.",
"Ruder et al.",
"(2016) created a hierarchical BLSTM with a sentence level BLSTM inputting into a review level BLSTM thus allowing them to take into account inter-and intra-sentence context.",
"They used only word embeddings making their system less dependent on extensive feature engineering or manual feature creation.",
"This system outperformed all others on certain languages on the SemEval-2016 task 5 dataset (Pontiki et al., 2014) and on other languages performed close to the best systems.",
"Wang et al.",
"(2016) also created an LSTM based model using word embeddings but instead of a hierarchical model it was a one layered LSTM with attention which puts more emphasis on learning the sentiment of words specific to a given aspect.",
"Data The training data published by the organisers for this track was a set of headline sentences from financial news articles where each sentence was tagged with the company name (which we treat as the aspect) and the polarity of the sentence with respect to the company.",
"There is the possibility that the same sentence occurs more than once if there is more than one company mentioned.",
"The polarity was a real value between -1 (negative sentiment) and 1 (positive sentiment).",
"We additionally trained a word2vec (Mikolov et al., 2013) word embedding model 3 on a set of 189,206 financial articles containing 161,877,425 tokens, that were manually downloaded from Factiva 4 .",
"The articles stem from a range of sources including the Financial Times and relate to companies from the United States only.",
"We trained the model on domain specific data as it has been shown many times that the financial domain can contain very different language.",
"System description Even though we have outlined this task as an aspect based sentiment task, this is instantiated in only one of the features in the SVR.",
"The following two subsections describe the two approaches, first SVR and then BLSTM.",
"Key implementation details are exposed here in the paper, but we have released the source code and word embedding models to aid replicability and further experimentation.",
"SVR The system was created using ScitKit learn (Pedregosa et al., 2011) linear Support Vector Regression model (Drucker et al., 1997) .",
"We exper-imented with the following different features and parameter settings: Tokenisation For comparison purposes, we tested whether or not a simple whitespace tokeniser can perform just as well as a full tokeniser, and in this case we used Unitok 5 .",
"N-grams We compared word-level uni-grams and bi-grams separately and in combination.",
"SVR parameters We tested different penalty parameters C and different epsilon parameters of the SVR.",
"Word Replacements We tested replacements to see if generalising words by inserting special tokens would help to reduce the sparsity problem.",
"We placed the word replacements into three separate groups: 1.",
"Company -When a company was mentioned in the input headline from the list of companies in the training data marked up as aspects, it was replaced by a company special token.",
"2.",
"Positive -When a positive word was mentioned in the input headline from a list of positive words (which was created using the N most similar words based on cosine distance) to 'excellent' using the pre-trained word2vec model.",
"3.",
"Negative -The same as the positive group however the word used was 'poor' instead of 'excellent'.",
"In the positive and negative groups, we chose the words 'excellent' and 'poor' following Turney (2002) to group the terms together under nondomain specific sentiment words.",
"Target aspect In order to incorporated the company as an aspect, we employed a boolean vector to represent the sentiment of the sentence.",
"This was done in order to see if the system could better differentiate the sentiment when the sentence was the same but the company was different.",
"BLSTM We created two different Bidirectional (Graves and Schmidhuber, 2005 ) Long Short-Term Memory (Hochreiter and Schmidhuber, 1997) using the Python Keras library (Chollet, 2015) with tensor flow backend (Abadi et al., 2016) .",
"We choose an LSTM model as it solves the vanishing gradients problem of Recurrent Neural Networks.",
"We used a bidirectional model as it allows us to capture information that came before and after instead of just before, thereby allowing us to capture more relevant context within the model.",
"Practically, a BLSTM is two LSTMs one going forward through the tokens the other in reverse order and in our models concatenating the resulting output vectors together at each time step.",
"The BLSTM models take as input a headline sentence of size L tokens 6 where L is the length of the longest sentence in the training texts.",
"Each word is converted into a 300 dimension vector using the word2vec model trained over the financial text 7 .",
"Any text that is not recognised by the word2vec model is represented as a vector of zeros; this is also used to pad out the sentence if it is shorter than L. Both BLSTM models have the following similar properties: 1.",
"Gradient clipping value of 5 -This was to help with the exploding gradients problem.",
"2.",
"Minimised the Mean Square Error (MSE) loss using RMSprop with a mini batch size of 32.",
"The output activation function is linear.",
"The main difference between the two models is the use of drop out and when they stop training over the data (epoch).",
"Both models architectures can be seen in figure 1.",
"Standard LSTM (SLSTM) The BLSTMs do contain drop out in both the input and between the connections of 0.2 each.",
"Finally the epoch is fixed at 25.",
"Early LSTM (ELSTM) As can be seen from figure 1, the drop out of 0.5 only happens between the layers and not the 6 Tokenised by Unitok 7 See the following link for detailed implementation details https://github.com/apmoore1/semeval# finance-word2vec-model connections as in the SLSTM.",
"Also the epoch is not fixed, it uses early stopping with a patience of 10.",
"We expect that this model can generalise better than the standard one due to the higher drop out and that the epoch is based on early stopping which relies on a validation set to know when to stop training.",
"Results We first present our findings on the best performing parameters and features for the SVRs.",
"These were determined by cross validation (CV) scores on the provided training data set using cosine similarity as the evaluation metric.",
"8 We found that using uni-grams and bi-grams performs best and using only bi-grams to be the worst.",
"Using the Unitok tokeniser always performed better than simple whitespace tokenisation.",
"The binary presence of tokens over frequency did not alter performance.",
"The C parameter was tested for three values; 0.01, 0.1 and 1.",
"We found very little difference between 0.1 and 1, but 0.01 produced much poorer results.",
"The eplison parameter was tested for 0.001, 0.01 and 0.1 the performance did not differ much but the lower the higher the performance but the more likely to overfit.",
"Using word replacements was effective for all three types (company, positive and negative) but using a value N=10 performed best for both positive and negative words.",
"Using target aspects also improved results.",
"Therefore, the best SVR model comprised of: Unitok tokenisation, uni-and bi-grams, word representation, C=0.1, eplison=0.01, company, positive, and negative word replacements and target aspects.",
"The main evaluation over the test data is based on the best performing SVR and the two BLSTM models once trained on all of the training data.",
"The result table 1 shows three columns based on the three evaluation metrics that the organisers have used.",
"Metric 1 is the original metric, weighted cosine similarity (the metric used to evaluate the final version of the results, where we were ranked 5th; metric provided on the task website 9 ).",
"This was then changed after the evaluation deadline to equation 1 10 (which we term metric 2; this is what the first version of the results were actually based on, where we were ranked 4th), which then changed by the organisers to their equation as presented in Cortis et al.",
"(2017) (which we term metric 3 and what the second version of the results were based on, where we were ranked 5th).",
"Model Metric 1 As you can see from the results table 1, the difference between the metrics is quite substantial.",
"This is due to the system's optimisation being based on metric 1 rather than 2.",
"Metric 2 is a classification metric for sentences with one aspect as it penalises values that are of opposite sign (giving -1 score) and rewards values with the same sign (giving +1 score).",
"Our systems are not optimised for this because it would predict scores of -0.01 and true value of 0.01 as very close (within vector of other results) with low error whereas metric 2 would give this the highest error rating of -1 as they are not the same sign.",
"Metric 3 is more similar to metric 1 as shown by the results, however the crucial difference is that again if you get opposite signs it will penalise more.",
"We analysed the top 50 errors based on Mean Absolute Error (MAE) in the test dataset specifically to examine the number of sentences containing more than one aspect.",
"Our investigation shows that no one system is better at predicting the sentiment of sentences that have more than one aspect (i.e.",
"company) within them.",
"Within those top 50 errors we found that the BLSTM systems do not know which parts of the sentence are associated to the company the sentiment is with respect to.",
"Also they do not know the strength/existence of certain sentiment words.",
"Conclusion and Future Work In this short paper, we have described our implemented solutions to SemEval Task 5 track 2, utilising both SVR and BLSTM approaches.",
"Our results show an improvement of around 5% when using LSTM models relative to SVR.",
"We have shown that this task can be partially represented as an aspect based sentiment task on a domain specific problem.",
"In general, our approaches acted as sentence level classifiers as they take no target company into consideration.",
"As our results show, the choice of evaluation metric makes a great deal of difference to system training and testing.",
"Future work will be to implement aspect specific information into an LSTM model as it has been shown to be useful in other work (Wang et al., 2016) ."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"4",
"4.1",
"4.1.1",
"4.1.2",
"4.1.3",
"4.1.4",
"4.1.5",
"4.2",
"3.",
"4.2.1",
"4.2.2",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Data",
"System description",
"SVR",
"Tokenisation",
"N-grams",
"SVR parameters",
"Word Replacements",
"Target aspect",
"BLSTM",
"The output activation function is linear.",
"Standard LSTM (SLSTM)",
"Early LSTM (ELSTM)",
"Results",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-6#paper-970#slide-7 | SVR best features | Using uni-grams and bi-grams to be the best. 2.4% improvement
Using a tokeniser always better. Affects bi-gram results the most.
1% improvement using Unitok5 over whitespace.
SVR parameter settings important 8% difference between using
Incorporating the target aspect increased performance. 0.3%
Using all word replacements. N=10 for POS and NEG words and
N=0 for company. 0.8% improvement using company and 0.2% for | Using uni-grams and bi-grams to be the best. 2.4% improvement
Using a tokeniser always better. Affects bi-gram results the most.
1% improvement using Unitok5 over whitespace.
SVR parameter settings important 8% difference between using
Incorporating the target aspect increased performance. 0.3%
Using all word replacements. N=10 for POS and NEG words and
N=0 for company. 0.8% improvement using company and 0.2% for | [] |
GEM-SciDuet-train-6#paper-970#slide-8 | 970 | Lancaster A at SemEval-2017 Task 5: Evaluation metrics matter: predicting sentiment from financial news headlines | This paper describes our participation in Task 5 track 2 of SemEval 2017 to predict the sentiment of financial news headlines for a specific company on a continuous scale between -1 and 1. We tackled the problem using a number of approaches, utilising a Support Vector Regression (SVR) and a Bidirectional Long Short-Term Memory (BLSTM). We found an improvement of 4-6% using the LSTM model over the SVR and came fourth in the track. We report a number of different evaluations using a finance specific word embedding model and reflect on the effects of using different evaluation metrics. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95
],
"paper_content_text": [
"Introduction The objective of Task 5 Track 2 of SemEval (2017) was to predict the sentiment of news headlines with respect to companies mentioned within the headlines.",
"This task can be seen as a financespecific aspect-based sentiment task (Nasukawa and Yi, 2003) .",
"The main motivations of this task is to find specific features and learning algorithms that will perform better for this domain as aspect based sentiment analysis tasks have been conducted before at SemEval (Pontiki et al., 2014) .",
"Domain specific terminology is expected to play a key part in this task, as reporters, investors and analysts in the financial domain will use a specific set of terminology when discussing financial performance.",
"Potentially, this may also vary across different financial domains and industry sectors.",
"Therefore, we took an exploratory approach and investigated how various features and learning algorithms perform differently, specifically SVR and BLSTMs.",
"We found that BLSTMs outperform an SVR without having any knowledge of the company that the sentiment is with respect to.",
"For replicability purposes, with this paper we are releasing our source code 1 and the finance specific BLSTM word embedding model 2 .",
"Related Work There is a growing amount of research being carried out related to sentiment analysis within the financial domain.",
"This work ranges from domainspecific lexicons (Loughran and McDonald, 2011) and lexicon creation (Moore et al., 2016) to stock market prediction models (Peng and Jiang, 2016; Kazemian et al., 2016) .",
"Peng and Jiang (2016) used a multi layer neural network to predict the stock market and found that incorporating textual features from financial news can improve the accuracy of prediction.",
"Kazemian et al.",
"(2016) showed the importance of tuning sentiment analysis to the task of stock market prediction.",
"However, much of the previous work was based on numerical financial stock market data rather than on aspect level financial textual data.",
"In aspect based sentiment analysis, there have been many different techniques used to predict the polarity of an aspect as shown in SemEval-2016 task 5 (Pontiki et al., 2014 ).",
"The winning system (Brun et al., 2016 ) used many different linguistic features and an ensemble model, and the runner up (Kumar et al., 2016) used uni-grams, bi-grams and sentiment lexicons as features for a Support Vector Machine (SVM).",
"Deep learning methods have also been applied to aspect polarity prediction.",
"Ruder et al.",
"(2016) created a hierarchical BLSTM with a sentence level BLSTM inputting into a review level BLSTM thus allowing them to take into account inter-and intra-sentence context.",
"They used only word embeddings making their system less dependent on extensive feature engineering or manual feature creation.",
"This system outperformed all others on certain languages on the SemEval-2016 task 5 dataset (Pontiki et al., 2014) and on other languages performed close to the best systems.",
"Wang et al.",
"(2016) also created an LSTM based model using word embeddings but instead of a hierarchical model it was a one layered LSTM with attention which puts more emphasis on learning the sentiment of words specific to a given aspect.",
"Data The training data published by the organisers for this track was a set of headline sentences from financial news articles where each sentence was tagged with the company name (which we treat as the aspect) and the polarity of the sentence with respect to the company.",
"There is the possibility that the same sentence occurs more than once if there is more than one company mentioned.",
"The polarity was a real value between -1 (negative sentiment) and 1 (positive sentiment).",
"We additionally trained a word2vec (Mikolov et al., 2013) word embedding model 3 on a set of 189,206 financial articles containing 161,877,425 tokens, that were manually downloaded from Factiva 4 .",
"The articles stem from a range of sources including the Financial Times and relate to companies from the United States only.",
"We trained the model on domain specific data as it has been shown many times that the financial domain can contain very different language.",
"System description Even though we have outlined this task as an aspect based sentiment task, this is instantiated in only one of the features in the SVR.",
"The following two subsections describe the two approaches, first SVR and then BLSTM.",
"Key implementation details are exposed here in the paper, but we have released the source code and word embedding models to aid replicability and further experimentation.",
"SVR The system was created using ScitKit learn (Pedregosa et al., 2011) linear Support Vector Regression model (Drucker et al., 1997) .",
"We exper-imented with the following different features and parameter settings: Tokenisation For comparison purposes, we tested whether or not a simple whitespace tokeniser can perform just as well as a full tokeniser, and in this case we used Unitok 5 .",
"N-grams We compared word-level uni-grams and bi-grams separately and in combination.",
"SVR parameters We tested different penalty parameters C and different epsilon parameters of the SVR.",
"Word Replacements We tested replacements to see if generalising words by inserting special tokens would help to reduce the sparsity problem.",
"We placed the word replacements into three separate groups: 1.",
"Company -When a company was mentioned in the input headline from the list of companies in the training data marked up as aspects, it was replaced by a company special token.",
"2.",
"Positive -When a positive word was mentioned in the input headline from a list of positive words (which was created using the N most similar words based on cosine distance) to 'excellent' using the pre-trained word2vec model.",
"3.",
"Negative -The same as the positive group however the word used was 'poor' instead of 'excellent'.",
"In the positive and negative groups, we chose the words 'excellent' and 'poor' following Turney (2002) to group the terms together under nondomain specific sentiment words.",
"Target aspect In order to incorporated the company as an aspect, we employed a boolean vector to represent the sentiment of the sentence.",
"This was done in order to see if the system could better differentiate the sentiment when the sentence was the same but the company was different.",
"BLSTM We created two different Bidirectional (Graves and Schmidhuber, 2005 ) Long Short-Term Memory (Hochreiter and Schmidhuber, 1997) using the Python Keras library (Chollet, 2015) with tensor flow backend (Abadi et al., 2016) .",
"We choose an LSTM model as it solves the vanishing gradients problem of Recurrent Neural Networks.",
"We used a bidirectional model as it allows us to capture information that came before and after instead of just before, thereby allowing us to capture more relevant context within the model.",
"Practically, a BLSTM is two LSTMs one going forward through the tokens the other in reverse order and in our models concatenating the resulting output vectors together at each time step.",
"The BLSTM models take as input a headline sentence of size L tokens 6 where L is the length of the longest sentence in the training texts.",
"Each word is converted into a 300 dimension vector using the word2vec model trained over the financial text 7 .",
"Any text that is not recognised by the word2vec model is represented as a vector of zeros; this is also used to pad out the sentence if it is shorter than L. Both BLSTM models have the following similar properties: 1.",
"Gradient clipping value of 5 -This was to help with the exploding gradients problem.",
"2.",
"Minimised the Mean Square Error (MSE) loss using RMSprop with a mini batch size of 32.",
"The output activation function is linear.",
"The main difference between the two models is the use of drop out and when they stop training over the data (epoch).",
"Both models architectures can be seen in figure 1.",
"Standard LSTM (SLSTM) The BLSTMs do contain drop out in both the input and between the connections of 0.2 each.",
"Finally the epoch is fixed at 25.",
"Early LSTM (ELSTM) As can be seen from figure 1, the drop out of 0.5 only happens between the layers and not the 6 Tokenised by Unitok 7 See the following link for detailed implementation details https://github.com/apmoore1/semeval# finance-word2vec-model connections as in the SLSTM.",
"Also the epoch is not fixed, it uses early stopping with a patience of 10.",
"We expect that this model can generalise better than the standard one due to the higher drop out and that the epoch is based on early stopping which relies on a validation set to know when to stop training.",
"Results We first present our findings on the best performing parameters and features for the SVRs.",
"These were determined by cross validation (CV) scores on the provided training data set using cosine similarity as the evaluation metric.",
"8 We found that using uni-grams and bi-grams performs best and using only bi-grams to be the worst.",
"Using the Unitok tokeniser always performed better than simple whitespace tokenisation.",
"The binary presence of tokens over frequency did not alter performance.",
"The C parameter was tested for three values; 0.01, 0.1 and 1.",
"We found very little difference between 0.1 and 1, but 0.01 produced much poorer results.",
"The eplison parameter was tested for 0.001, 0.01 and 0.1 the performance did not differ much but the lower the higher the performance but the more likely to overfit.",
"Using word replacements was effective for all three types (company, positive and negative) but using a value N=10 performed best for both positive and negative words.",
"Using target aspects also improved results.",
"Therefore, the best SVR model comprised of: Unitok tokenisation, uni-and bi-grams, word representation, C=0.1, eplison=0.01, company, positive, and negative word replacements and target aspects.",
"The main evaluation over the test data is based on the best performing SVR and the two BLSTM models once trained on all of the training data.",
"The result table 1 shows three columns based on the three evaluation metrics that the organisers have used.",
"Metric 1 is the original metric, weighted cosine similarity (the metric used to evaluate the final version of the results, where we were ranked 5th; metric provided on the task website 9 ).",
"This was then changed after the evaluation deadline to equation 1 10 (which we term metric 2; this is what the first version of the results were actually based on, where we were ranked 4th), which then changed by the organisers to their equation as presented in Cortis et al.",
"(2017) (which we term metric 3 and what the second version of the results were based on, where we were ranked 5th).",
"Model Metric 1 As you can see from the results table 1, the difference between the metrics is quite substantial.",
"This is due to the system's optimisation being based on metric 1 rather than 2.",
"Metric 2 is a classification metric for sentences with one aspect as it penalises values that are of opposite sign (giving -1 score) and rewards values with the same sign (giving +1 score).",
"Our systems are not optimised for this because it would predict scores of -0.01 and true value of 0.01 as very close (within vector of other results) with low error whereas metric 2 would give this the highest error rating of -1 as they are not the same sign.",
"Metric 3 is more similar to metric 1 as shown by the results, however the crucial difference is that again if you get opposite signs it will penalise more.",
"We analysed the top 50 errors based on Mean Absolute Error (MAE) in the test dataset specifically to examine the number of sentences containing more than one aspect.",
"Our investigation shows that no one system is better at predicting the sentiment of sentences that have more than one aspect (i.e.",
"company) within them.",
"Within those top 50 errors we found that the BLSTM systems do not know which parts of the sentence are associated to the company the sentiment is with respect to.",
"Also they do not know the strength/existence of certain sentiment words.",
"Conclusion and Future Work In this short paper, we have described our implemented solutions to SemEval Task 5 track 2, utilising both SVR and BLSTM approaches.",
"Our results show an improvement of around 5% when using LSTM models relative to SVR.",
"We have shown that this task can be partially represented as an aspect based sentiment task on a domain specific problem.",
"In general, our approaches acted as sentence level classifiers as they take no target company into consideration.",
"As our results show, the choice of evaluation metric makes a great deal of difference to system training and testing.",
"Future work will be to implement aspect specific information into an LSTM model as it has been shown to be useful in other work (Wang et al., 2016) ."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"4",
"4.1",
"4.1.1",
"4.1.2",
"4.1.3",
"4.1.4",
"4.1.5",
"4.2",
"3.",
"4.2.1",
"4.2.2",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Data",
"System description",
"SVR",
"Tokenisation",
"N-grams",
"SVR parameters",
"Word Replacements",
"Target aspect",
"BLSTM",
"The output activation function is linear.",
"Standard LSTM (SLSTM)",
"Early LSTM (ELSTM)",
"Results",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-6#paper-970#slide-8 | Results across the different metrics | Metric 1 was the final metric used. | Metric 1 was the final metric used. | [] |
GEM-SciDuet-train-6#paper-970#slide-9 | 970 | Lancaster A at SemEval-2017 Task 5: Evaluation metrics matter: predicting sentiment from financial news headlines | This paper describes our participation in Task 5 track 2 of SemEval 2017 to predict the sentiment of financial news headlines for a specific company on a continuous scale between -1 and 1. We tackled the problem using a number of approaches, utilising a Support Vector Regression (SVR) and a Bidirectional Long Short-Term Memory (BLSTM). We found an improvement of 4-6% using the LSTM model over the SVR and came fourth in the track. We report a number of different evaluations using a finance specific word embedding model and reflect on the effects of using different evaluation metrics. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95
],
"paper_content_text": [
"Introduction The objective of Task 5 Track 2 of SemEval (2017) was to predict the sentiment of news headlines with respect to companies mentioned within the headlines.",
"This task can be seen as a financespecific aspect-based sentiment task (Nasukawa and Yi, 2003) .",
"The main motivations of this task is to find specific features and learning algorithms that will perform better for this domain as aspect based sentiment analysis tasks have been conducted before at SemEval (Pontiki et al., 2014) .",
"Domain specific terminology is expected to play a key part in this task, as reporters, investors and analysts in the financial domain will use a specific set of terminology when discussing financial performance.",
"Potentially, this may also vary across different financial domains and industry sectors.",
"Therefore, we took an exploratory approach and investigated how various features and learning algorithms perform differently, specifically SVR and BLSTMs.",
"We found that BLSTMs outperform an SVR without having any knowledge of the company that the sentiment is with respect to.",
"For replicability purposes, with this paper we are releasing our source code 1 and the finance specific BLSTM word embedding model 2 .",
"Related Work There is a growing amount of research being carried out related to sentiment analysis within the financial domain.",
"This work ranges from domainspecific lexicons (Loughran and McDonald, 2011) and lexicon creation (Moore et al., 2016) to stock market prediction models (Peng and Jiang, 2016; Kazemian et al., 2016) .",
"Peng and Jiang (2016) used a multi layer neural network to predict the stock market and found that incorporating textual features from financial news can improve the accuracy of prediction.",
"Kazemian et al.",
"(2016) showed the importance of tuning sentiment analysis to the task of stock market prediction.",
"However, much of the previous work was based on numerical financial stock market data rather than on aspect level financial textual data.",
"In aspect based sentiment analysis, there have been many different techniques used to predict the polarity of an aspect as shown in SemEval-2016 task 5 (Pontiki et al., 2014 ).",
"The winning system (Brun et al., 2016 ) used many different linguistic features and an ensemble model, and the runner up (Kumar et al., 2016) used uni-grams, bi-grams and sentiment lexicons as features for a Support Vector Machine (SVM).",
"Deep learning methods have also been applied to aspect polarity prediction.",
"Ruder et al.",
"(2016) created a hierarchical BLSTM with a sentence level BLSTM inputting into a review level BLSTM thus allowing them to take into account inter-and intra-sentence context.",
"They used only word embeddings making their system less dependent on extensive feature engineering or manual feature creation.",
"This system outperformed all others on certain languages on the SemEval-2016 task 5 dataset (Pontiki et al., 2014) and on other languages performed close to the best systems.",
"Wang et al.",
"(2016) also created an LSTM based model using word embeddings but instead of a hierarchical model it was a one layered LSTM with attention which puts more emphasis on learning the sentiment of words specific to a given aspect.",
"Data The training data published by the organisers for this track was a set of headline sentences from financial news articles where each sentence was tagged with the company name (which we treat as the aspect) and the polarity of the sentence with respect to the company.",
"There is the possibility that the same sentence occurs more than once if there is more than one company mentioned.",
"The polarity was a real value between -1 (negative sentiment) and 1 (positive sentiment).",
"We additionally trained a word2vec (Mikolov et al., 2013) word embedding model 3 on a set of 189,206 financial articles containing 161,877,425 tokens, that were manually downloaded from Factiva 4 .",
"The articles stem from a range of sources including the Financial Times and relate to companies from the United States only.",
"We trained the model on domain specific data as it has been shown many times that the financial domain can contain very different language.",
"System description Even though we have outlined this task as an aspect based sentiment task, this is instantiated in only one of the features in the SVR.",
"The following two subsections describe the two approaches, first SVR and then BLSTM.",
"Key implementation details are exposed here in the paper, but we have released the source code and word embedding models to aid replicability and further experimentation.",
"SVR The system was created using ScitKit learn (Pedregosa et al., 2011) linear Support Vector Regression model (Drucker et al., 1997) .",
"We exper-imented with the following different features and parameter settings: Tokenisation For comparison purposes, we tested whether or not a simple whitespace tokeniser can perform just as well as a full tokeniser, and in this case we used Unitok 5 .",
"N-grams We compared word-level uni-grams and bi-grams separately and in combination.",
"SVR parameters We tested different penalty parameters C and different epsilon parameters of the SVR.",
"Word Replacements We tested replacements to see if generalising words by inserting special tokens would help to reduce the sparsity problem.",
"We placed the word replacements into three separate groups: 1.",
"Company -When a company was mentioned in the input headline from the list of companies in the training data marked up as aspects, it was replaced by a company special token.",
"2.",
"Positive -When a positive word was mentioned in the input headline from a list of positive words (which was created using the N most similar words based on cosine distance) to 'excellent' using the pre-trained word2vec model.",
"3.",
"Negative -The same as the positive group however the word used was 'poor' instead of 'excellent'.",
"In the positive and negative groups, we chose the words 'excellent' and 'poor' following Turney (2002) to group the terms together under nondomain specific sentiment words.",
"Target aspect In order to incorporated the company as an aspect, we employed a boolean vector to represent the sentiment of the sentence.",
"This was done in order to see if the system could better differentiate the sentiment when the sentence was the same but the company was different.",
"BLSTM We created two different Bidirectional (Graves and Schmidhuber, 2005 ) Long Short-Term Memory (Hochreiter and Schmidhuber, 1997) using the Python Keras library (Chollet, 2015) with tensor flow backend (Abadi et al., 2016) .",
"We choose an LSTM model as it solves the vanishing gradients problem of Recurrent Neural Networks.",
"We used a bidirectional model as it allows us to capture information that came before and after instead of just before, thereby allowing us to capture more relevant context within the model.",
"Practically, a BLSTM is two LSTMs one going forward through the tokens the other in reverse order and in our models concatenating the resulting output vectors together at each time step.",
"The BLSTM models take as input a headline sentence of size L tokens 6 where L is the length of the longest sentence in the training texts.",
"Each word is converted into a 300 dimension vector using the word2vec model trained over the financial text 7 .",
"Any text that is not recognised by the word2vec model is represented as a vector of zeros; this is also used to pad out the sentence if it is shorter than L. Both BLSTM models have the following similar properties: 1.",
"Gradient clipping value of 5 -This was to help with the exploding gradients problem.",
"2.",
"Minimised the Mean Square Error (MSE) loss using RMSprop with a mini batch size of 32.",
"The output activation function is linear.",
"The main difference between the two models is the use of drop out and when they stop training over the data (epoch).",
"Both models architectures can be seen in figure 1.",
"Standard LSTM (SLSTM) The BLSTMs do contain drop out in both the input and between the connections of 0.2 each.",
"Finally the epoch is fixed at 25.",
"Early LSTM (ELSTM) As can be seen from figure 1, the drop out of 0.5 only happens between the layers and not the 6 Tokenised by Unitok 7 See the following link for detailed implementation details https://github.com/apmoore1/semeval# finance-word2vec-model connections as in the SLSTM.",
"Also the epoch is not fixed, it uses early stopping with a patience of 10.",
"We expect that this model can generalise better than the standard one due to the higher drop out and that the epoch is based on early stopping which relies on a validation set to know when to stop training.",
"Results We first present our findings on the best performing parameters and features for the SVRs.",
"These were determined by cross validation (CV) scores on the provided training data set using cosine similarity as the evaluation metric.",
"8 We found that using uni-grams and bi-grams performs best and using only bi-grams to be the worst.",
"Using the Unitok tokeniser always performed better than simple whitespace tokenisation.",
"The binary presence of tokens over frequency did not alter performance.",
"The C parameter was tested for three values; 0.01, 0.1 and 1.",
"We found very little difference between 0.1 and 1, but 0.01 produced much poorer results.",
"The eplison parameter was tested for 0.001, 0.01 and 0.1 the performance did not differ much but the lower the higher the performance but the more likely to overfit.",
"Using word replacements was effective for all three types (company, positive and negative) but using a value N=10 performed best for both positive and negative words.",
"Using target aspects also improved results.",
"Therefore, the best SVR model comprised of: Unitok tokenisation, uni-and bi-grams, word representation, C=0.1, eplison=0.01, company, positive, and negative word replacements and target aspects.",
"The main evaluation over the test data is based on the best performing SVR and the two BLSTM models once trained on all of the training data.",
"The result table 1 shows three columns based on the three evaluation metrics that the organisers have used.",
"Metric 1 is the original metric, weighted cosine similarity (the metric used to evaluate the final version of the results, where we were ranked 5th; metric provided on the task website 9 ).",
"This was then changed after the evaluation deadline to equation 1 10 (which we term metric 2; this is what the first version of the results were actually based on, where we were ranked 4th), which then changed by the organisers to their equation as presented in Cortis et al.",
"(2017) (which we term metric 3 and what the second version of the results were based on, where we were ranked 5th).",
"Model Metric 1 As you can see from the results table 1, the difference between the metrics is quite substantial.",
"This is due to the system's optimisation being based on metric 1 rather than 2.",
"Metric 2 is a classification metric for sentences with one aspect as it penalises values that are of opposite sign (giving -1 score) and rewards values with the same sign (giving +1 score).",
"Our systems are not optimised for this because it would predict scores of -0.01 and true value of 0.01 as very close (within vector of other results) with low error whereas metric 2 would give this the highest error rating of -1 as they are not the same sign.",
"Metric 3 is more similar to metric 1 as shown by the results, however the crucial difference is that again if you get opposite signs it will penalise more.",
"We analysed the top 50 errors based on Mean Absolute Error (MAE) in the test dataset specifically to examine the number of sentences containing more than one aspect.",
"Our investigation shows that no one system is better at predicting the sentiment of sentences that have more than one aspect (i.e.",
"company) within them.",
"Within those top 50 errors we found that the BLSTM systems do not know which parts of the sentence are associated to the company the sentiment is with respect to.",
"Also they do not know the strength/existence of certain sentiment words.",
"Conclusion and Future Work In this short paper, we have described our implemented solutions to SemEval Task 5 track 2, utilising both SVR and BLSTM approaches.",
"Our results show an improvement of around 5% when using LSTM models relative to SVR.",
"We have shown that this task can be partially represented as an aspect based sentiment task on a domain specific problem.",
"In general, our approaches acted as sentence level classifiers as they take no target company into consideration.",
"As our results show, the choice of evaluation metric makes a great deal of difference to system training and testing.",
"Future work will be to implement aspect specific information into an LSTM model as it has been shown to be useful in other work (Wang et al., 2016) ."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"4",
"4.1",
"4.1.1",
"4.1.2",
"4.1.3",
"4.1.4",
"4.1.5",
"4.2",
"3.",
"4.2.1",
"4.2.2",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Data",
"System description",
"SVR",
"Tokenisation",
"N-grams",
"SVR parameters",
"Word Replacements",
"Target aspect",
"BLSTM",
"The output activation function is linear.",
"Standard LSTM (SLSTM)",
"Early LSTM (ELSTM)",
"Results",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-6#paper-970#slide-9 | Future Work | 1. Incorporate aspects into the BLSTMs shown to be useful by Wang
2. Improve BLSTMs by using an attention model Wang et al. [7].
3. Add known financial sentiment lexicon into the LSTM model [6]. | 1. Incorporate aspects into the BLSTMs shown to be useful by Wang
2. Improve BLSTMs by using an attention model Wang et al. [7].
3. Add known financial sentiment lexicon into the LSTM model [6]. | [] |
GEM-SciDuet-train-6#paper-970#slide-10 | 970 | Lancaster A at SemEval-2017 Task 5: Evaluation metrics matter: predicting sentiment from financial news headlines | This paper describes our participation in Task 5 track 2 of SemEval 2017 to predict the sentiment of financial news headlines for a specific company on a continuous scale between -1 and 1. We tackled the problem using a number of approaches, utilising a Support Vector Regression (SVR) and a Bidirectional Long Short-Term Memory (BLSTM). We found an improvement of 4-6% using the LSTM model over the SVR and came fourth in the track. We report a number of different evaluations using a finance specific word embedding model and reflect on the effects of using different evaluation metrics. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95
],
"paper_content_text": [
"Introduction The objective of Task 5 Track 2 of SemEval (2017) was to predict the sentiment of news headlines with respect to companies mentioned within the headlines.",
"This task can be seen as a financespecific aspect-based sentiment task (Nasukawa and Yi, 2003) .",
"The main motivations of this task is to find specific features and learning algorithms that will perform better for this domain as aspect based sentiment analysis tasks have been conducted before at SemEval (Pontiki et al., 2014) .",
"Domain specific terminology is expected to play a key part in this task, as reporters, investors and analysts in the financial domain will use a specific set of terminology when discussing financial performance.",
"Potentially, this may also vary across different financial domains and industry sectors.",
"Therefore, we took an exploratory approach and investigated how various features and learning algorithms perform differently, specifically SVR and BLSTMs.",
"We found that BLSTMs outperform an SVR without having any knowledge of the company that the sentiment is with respect to.",
"For replicability purposes, with this paper we are releasing our source code 1 and the finance specific BLSTM word embedding model 2 .",
"Related Work There is a growing amount of research being carried out related to sentiment analysis within the financial domain.",
"This work ranges from domainspecific lexicons (Loughran and McDonald, 2011) and lexicon creation (Moore et al., 2016) to stock market prediction models (Peng and Jiang, 2016; Kazemian et al., 2016) .",
"Peng and Jiang (2016) used a multi layer neural network to predict the stock market and found that incorporating textual features from financial news can improve the accuracy of prediction.",
"Kazemian et al.",
"(2016) showed the importance of tuning sentiment analysis to the task of stock market prediction.",
"However, much of the previous work was based on numerical financial stock market data rather than on aspect level financial textual data.",
"In aspect based sentiment analysis, there have been many different techniques used to predict the polarity of an aspect as shown in SemEval-2016 task 5 (Pontiki et al., 2014 ).",
"The winning system (Brun et al., 2016 ) used many different linguistic features and an ensemble model, and the runner up (Kumar et al., 2016) used uni-grams, bi-grams and sentiment lexicons as features for a Support Vector Machine (SVM).",
"Deep learning methods have also been applied to aspect polarity prediction.",
"Ruder et al.",
"(2016) created a hierarchical BLSTM with a sentence level BLSTM inputting into a review level BLSTM thus allowing them to take into account inter-and intra-sentence context.",
"They used only word embeddings making their system less dependent on extensive feature engineering or manual feature creation.",
"This system outperformed all others on certain languages on the SemEval-2016 task 5 dataset (Pontiki et al., 2014) and on other languages performed close to the best systems.",
"Wang et al.",
"(2016) also created an LSTM based model using word embeddings but instead of a hierarchical model it was a one layered LSTM with attention which puts more emphasis on learning the sentiment of words specific to a given aspect.",
"Data The training data published by the organisers for this track was a set of headline sentences from financial news articles where each sentence was tagged with the company name (which we treat as the aspect) and the polarity of the sentence with respect to the company.",
"There is the possibility that the same sentence occurs more than once if there is more than one company mentioned.",
"The polarity was a real value between -1 (negative sentiment) and 1 (positive sentiment).",
"We additionally trained a word2vec (Mikolov et al., 2013) word embedding model 3 on a set of 189,206 financial articles containing 161,877,425 tokens, that were manually downloaded from Factiva 4 .",
"The articles stem from a range of sources including the Financial Times and relate to companies from the United States only.",
"We trained the model on domain specific data as it has been shown many times that the financial domain can contain very different language.",
"System description Even though we have outlined this task as an aspect based sentiment task, this is instantiated in only one of the features in the SVR.",
"The following two subsections describe the two approaches, first SVR and then BLSTM.",
"Key implementation details are exposed here in the paper, but we have released the source code and word embedding models to aid replicability and further experimentation.",
"SVR The system was created using ScitKit learn (Pedregosa et al., 2011) linear Support Vector Regression model (Drucker et al., 1997) .",
"We exper-imented with the following different features and parameter settings: Tokenisation For comparison purposes, we tested whether or not a simple whitespace tokeniser can perform just as well as a full tokeniser, and in this case we used Unitok 5 .",
"N-grams We compared word-level uni-grams and bi-grams separately and in combination.",
"SVR parameters We tested different penalty parameters C and different epsilon parameters of the SVR.",
"Word Replacements We tested replacements to see if generalising words by inserting special tokens would help to reduce the sparsity problem.",
"We placed the word replacements into three separate groups: 1.",
"Company -When a company was mentioned in the input headline from the list of companies in the training data marked up as aspects, it was replaced by a company special token.",
"2.",
"Positive -When a positive word was mentioned in the input headline from a list of positive words (which was created using the N most similar words based on cosine distance) to 'excellent' using the pre-trained word2vec model.",
"3.",
"Negative -The same as the positive group however the word used was 'poor' instead of 'excellent'.",
"In the positive and negative groups, we chose the words 'excellent' and 'poor' following Turney (2002) to group the terms together under nondomain specific sentiment words.",
"Target aspect In order to incorporated the company as an aspect, we employed a boolean vector to represent the sentiment of the sentence.",
"This was done in order to see if the system could better differentiate the sentiment when the sentence was the same but the company was different.",
"BLSTM We created two different Bidirectional (Graves and Schmidhuber, 2005 ) Long Short-Term Memory (Hochreiter and Schmidhuber, 1997) using the Python Keras library (Chollet, 2015) with tensor flow backend (Abadi et al., 2016) .",
"We choose an LSTM model as it solves the vanishing gradients problem of Recurrent Neural Networks.",
"We used a bidirectional model as it allows us to capture information that came before and after instead of just before, thereby allowing us to capture more relevant context within the model.",
"Practically, a BLSTM is two LSTMs one going forward through the tokens the other in reverse order and in our models concatenating the resulting output vectors together at each time step.",
"The BLSTM models take as input a headline sentence of size L tokens 6 where L is the length of the longest sentence in the training texts.",
"Each word is converted into a 300 dimension vector using the word2vec model trained over the financial text 7 .",
"Any text that is not recognised by the word2vec model is represented as a vector of zeros; this is also used to pad out the sentence if it is shorter than L. Both BLSTM models have the following similar properties: 1.",
"Gradient clipping value of 5 -This was to help with the exploding gradients problem.",
"2.",
"Minimised the Mean Square Error (MSE) loss using RMSprop with a mini batch size of 32.",
"The output activation function is linear.",
"The main difference between the two models is the use of drop out and when they stop training over the data (epoch).",
"Both models architectures can be seen in figure 1.",
"Standard LSTM (SLSTM) The BLSTMs do contain drop out in both the input and between the connections of 0.2 each.",
"Finally the epoch is fixed at 25.",
"Early LSTM (ELSTM) As can be seen from figure 1, the drop out of 0.5 only happens between the layers and not the 6 Tokenised by Unitok 7 See the following link for detailed implementation details https://github.com/apmoore1/semeval# finance-word2vec-model connections as in the SLSTM.",
"Also the epoch is not fixed, it uses early stopping with a patience of 10.",
"We expect that this model can generalise better than the standard one due to the higher drop out and that the epoch is based on early stopping which relies on a validation set to know when to stop training.",
"Results We first present our findings on the best performing parameters and features for the SVRs.",
"These were determined by cross validation (CV) scores on the provided training data set using cosine similarity as the evaluation metric.",
"8 We found that using uni-grams and bi-grams performs best and using only bi-grams to be the worst.",
"Using the Unitok tokeniser always performed better than simple whitespace tokenisation.",
"The binary presence of tokens over frequency did not alter performance.",
"The C parameter was tested for three values; 0.01, 0.1 and 1.",
"We found very little difference between 0.1 and 1, but 0.01 produced much poorer results.",
"The eplison parameter was tested for 0.001, 0.01 and 0.1 the performance did not differ much but the lower the higher the performance but the more likely to overfit.",
"Using word replacements was effective for all three types (company, positive and negative) but using a value N=10 performed best for both positive and negative words.",
"Using target aspects also improved results.",
"Therefore, the best SVR model comprised of: Unitok tokenisation, uni-and bi-grams, word representation, C=0.1, eplison=0.01, company, positive, and negative word replacements and target aspects.",
"The main evaluation over the test data is based on the best performing SVR and the two BLSTM models once trained on all of the training data.",
"The result table 1 shows three columns based on the three evaluation metrics that the organisers have used.",
"Metric 1 is the original metric, weighted cosine similarity (the metric used to evaluate the final version of the results, where we were ranked 5th; metric provided on the task website 9 ).",
"This was then changed after the evaluation deadline to equation 1 10 (which we term metric 2; this is what the first version of the results were actually based on, where we were ranked 4th), which then changed by the organisers to their equation as presented in Cortis et al.",
"(2017) (which we term metric 3 and what the second version of the results were based on, where we were ranked 5th).",
"Model Metric 1 As you can see from the results table 1, the difference between the metrics is quite substantial.",
"This is due to the system's optimisation being based on metric 1 rather than 2.",
"Metric 2 is a classification metric for sentences with one aspect as it penalises values that are of opposite sign (giving -1 score) and rewards values with the same sign (giving +1 score).",
"Our systems are not optimised for this because it would predict scores of -0.01 and true value of 0.01 as very close (within vector of other results) with low error whereas metric 2 would give this the highest error rating of -1 as they are not the same sign.",
"Metric 3 is more similar to metric 1 as shown by the results, however the crucial difference is that again if you get opposite signs it will penalise more.",
"We analysed the top 50 errors based on Mean Absolute Error (MAE) in the test dataset specifically to examine the number of sentences containing more than one aspect.",
"Our investigation shows that no one system is better at predicting the sentiment of sentences that have more than one aspect (i.e.",
"company) within them.",
"Within those top 50 errors we found that the BLSTM systems do not know which parts of the sentence are associated to the company the sentiment is with respect to.",
"Also they do not know the strength/existence of certain sentiment words.",
"Conclusion and Future Work In this short paper, we have described our implemented solutions to SemEval Task 5 track 2, utilising both SVR and BLSTM approaches.",
"Our results show an improvement of around 5% when using LSTM models relative to SVR.",
"We have shown that this task can be partially represented as an aspect based sentiment task on a domain specific problem.",
"In general, our approaches acted as sentence level classifiers as they take no target company into consideration.",
"As our results show, the choice of evaluation metric makes a great deal of difference to system training and testing.",
"Future work will be to implement aspect specific information into an LSTM model as it has been shown to be useful in other work (Wang et al., 2016) ."
]
} | {
"paper_header_number": [
"1",
"2",
"3",
"4",
"4.1",
"4.1.1",
"4.1.2",
"4.1.3",
"4.1.4",
"4.1.5",
"4.2",
"3.",
"4.2.1",
"4.2.2",
"5",
"6"
],
"paper_header_content": [
"Introduction",
"Related Work",
"Data",
"System description",
"SVR",
"Tokenisation",
"N-grams",
"SVR parameters",
"Word Replacements",
"Target aspect",
"BLSTM",
"The output activation function is linear.",
"Standard LSTM (SLSTM)",
"Early LSTM (ELSTM)",
"Results",
"Conclusion and Future Work"
]
} | GEM-SciDuet-train-6#paper-970#slide-10 | Summary | 1. BLSTM outperform SVRs with minimal feature engineering.
2. The future is to incorporate more financial information into the | 1. BLSTM outperform SVRs with minimal feature engineering.
2. The future is to incorporate more financial information into the | [] |
GEM-SciDuet-train-7#paper-971#slide-0 | 971 | Exploring the leading authors and journals in major topics by citation sentences and topic modeling | Citation plays an important role in understanding the knowledge sharing among scholars. Citation sentences embed useful contents that signify the influence of cited authors on shared ideas, and express own opinion of citing authors to others' articles. The purpose of the study is to provide a new lens to analyze the topical relationship embedded in the citation sentences in an integrated manner. To this end, we extract citation sentences from full-text articles in the field of Oncology. In addition, we adopt Author-Journal-Topic (AJT) model to take both authors and journals into consideration of topic analysis. For the study, we collect the 6,360 full-text articles from PubMed Central and select the top 15 journals on Oncology. By applying AJT model, we identify what the major topics are shared among researchers in Oncology and which authors and journal lead the idea exchange in sub-disciplines of Oncology. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106
],
"paper_content_text": [
"Introduction As the size of data on the web continues to increase in an exponential manner, finding valuable meaning between data becomes of paramount importance in many research areas.",
"In the information science field, citations are challenging, pivotal materials to discover the relationship between academic documents because citations present the description of authors' ideas and the hidden relationship between authors and documents.",
"The earliest works focused mainly on classifying the citation behaviors and discovering the citation reasons with limited data such as the location of citation sentences and the number of references [1, 2] .",
"Since the mid-1990s, with the development of computer technology, citation content analysis was elaborated by applying data analysis techniques like text-mining or natural language processing.",
"Zhang et al.",
"[3] present citation analysis based on sematic and syntactic approaches.",
"Semantic-based citation analysis is performed by qualitative analysis to discover the citation motivation and citation classification.",
"On the other hand, syntactic-based citation analysis can be conducted by citation location and citation frequency, which reveals the hidden relation of authors by using meta-data of documents such as journal, venue of publication, affiliation of authors, etc.",
"Following their study, Ding et al.",
"[4] propose a theoretical methodology through content citation analysis.",
"However, these analyses are somewhat limited to the explicit context that primarily represents their own ideas and arguments.",
"The main goal of the paper is to discover the implicit topical relationships buried in citation sentences by utilizing the citation information from the author's perspective of sharing other authors' point of view.",
"Implicitness of the topical relationship is realized by using citation sentences as the input for the topic modeling technique.",
"In this study, a citation sentence indicates the sentence including citation expression consisting of year and author of the cited work.",
"In general, the citation sentence contains brief content of cited work and opinion that the author of citing work on the cited work.",
"We claim that citation sentences reveal interesting characteristics of scholarly communication such as influence, idea exchange, justification for citer's arguments, etc.",
"We assume that using citation sentences for topic analysis reveals aforementioned characteristics.",
"To explore such intellectual space created by citation sentences, we take both authors and journals into consideration of topic analysis.",
"To this end, we applied Author-Conference-Topic (ACT) model proposed by Tang et al.",
"[5] for our topic analysis in relation with both authors and journals, which is called Author-Journal-Topic (AJT) topic model.",
"ACT model is a probabilistic topic model for simultaneously extracting topics of papers, authors, and conferences.",
"There are a few studies to analyze content of citation sentences.",
"Most of previous studies focus on how the topic of document influences citation and vice versa [6, 7, 8] using Topic Modeling.",
"Kataria, Mitra, and Bhatia [8] adapt citation to Author-Topic model [9] with the assumption that the context surrounding the citation anchor could be used to get topical information about the cited authors.",
"These studies including Tang et al.",
"[10] 's ACT model are the examples of combining topic modelling methods and citation content analysis.",
"However, most previous studies used metadata of documents.",
"In this work, we focus on identifying the landscape of the oncology field from a perspective of citation.",
"By using citation sentences, our results can indicate which authors are actively cited and which journals lead a certain topic.",
"The rest of the paper is organized as follows: Section 2 describes the proposed approach.",
"Section 3 analyzes the topic modeling results.",
"Section 4 concludes the paper with the future work.",
"Methodology Main idea The basic assumption of the proposed approach is that citation sentences embed useful contents signifying the influence of cited authors on shared ideas of citing authors.",
"Citation sentences are also considered as an invisible intellectual place for idea exchanging since citations are effective means of supporting and expressing their own arguments by using other works.",
"In the similar vein, Di Marco and Mercer [11] claim that citation sentences play a major role in creating the relationship among relevant authors within the similar research fields.",
"With these assumptions, we are to explore the implicitness of topic relationships resided in citation sentences from the integrated perspective by incorporating the citing authors and journal titles into interpreting the topical relationships.",
"As shown in Figure 1 , we utilized various features including citing authors, citing sentences and journal titles for topic analysis.",
"Authors in Figure 1 mean the citing authors who write a paper and who cite other's work.",
"Citation sentences are the sentences written by the authors when they cite other's work in the paper, and journal titles are the journal names publishing the citing authors' paper.",
"By employing AJT model with these three parameters , we can discover which topics are the most salient ones referred to frequently by researchers and who are the leading authors sharing other authors' ideas in the research field and which journal leads such endeavor.",
"Fig.",
"1.",
"Three parameters for AJT model Data collection For this study, we compile the dataset on the field of Oncology from PubMed Central that provides the full-text in the biomedical field.",
"We select top 15 journals of Oncology by Thomson Reuter's JCR and journal's impact factor, and from these 15 journals, we are able to collect 6,360 full-text articles.",
"Figure 2 describes the workflow of our study.",
"As mentioned earlier, with the fulltext articles collected from PubMed Central, we extract the citation sentences.",
"Most citation sentences are kept in the following format: (author, year), (reference number) [reference number].",
"An example of such format is \"(<xref rid=\"bib00\" ref-type=\"bibr\">Author name, 2000</xref>)\".",
"We use the regular expression technique to parse and extract the citation sentences, when the tag <xref rid=>, </xref> appears on the sentences after parsing XML records with the Java-based SAX parser.",
"Method Fig.",
"2.",
"Workflow We also parse other metadata for AJT model such as the name of authors and journal titles.",
"The author tags, <surname> </surname> and <given-names></given-names> inside the <contrib-gourp></contrib-group>, denote the list of authors who wrote the paper.",
"For journal, we extract the titles when the journal tags, <journal-title> and </journal-title>, are included in the tag of <journal-meta> and </journal-meta>.",
"We also preprocess extracted sentences by removing both functional and general words and applying the Porter's stemming algorithm to improve the input for AJT Model.",
"AJT Model For our study, we apply ACT [10] model with several metadata such as citation sentences, journal titles and citing authors to develop AJT model.",
"Our AJT model utilizes journal titles and citation sentences instead of conference and abstract on documents.",
"The change of model is needed to analyze most influential topics in Oncology and to find leading authors who frequently mention the active topics and to detect the journals involved in such topics.",
"Figure 1 , Table 1) Like ACT model, AJT model assumes that each citing author is related to distribution over topics and each word in citation sentences is derived from a topic.",
"In the AJT model, the journal titles are related to each word.",
"To determine a word (ω_Si) in citation sentences (S), citing authors (x_Si) are consider for a word.",
"Each citing author is associated with a distributed topic.",
"A topic is generated from the citing author-topic distribution.",
"The words and journal titles are generated from a specific topic.",
"AJT model presents (1) 3 Results and Analyses For AJT model, we set the number of topics to 15 and finally select 8 topics as major topics.",
"Since we discovered that there are similar topics on our results, we calculated the similarity between 15 topics to select the most representative topics.",
"The topical similarities are measured by each word on topics and we calculated the similarities of two topics where each topic are represented in an array of a term vector.",
"Through this process, we chose 8 topics which have high topical similarities (over 0.5).",
"Each topic presents top 5 words from topic-word distribution, and 5 most related authors and journal titles are displayed along with each topic.",
"By performing several times on the pilot studies, we decided to choose top 5 words which are quite appropriate to describe each topics.",
"The results of AJT-based topic modeling is shown in Table 1 .",
"We label topic 1 \"breast cancer\" whose top words include breast, expression women, and growth.",
"Since the dataset is compiled with citation sentences, it implies that the topic \"breast cancer\" is a popular topic where researchers share and exchange ideas and facts related to breast cancer.",
"In relation to the topic \"breast cancer\", the active authors of breast cancer are Johnston Stephen RD, Colditz Graham A, and Sternlicht Mark D, and they share ideas with others on breast cancer from our results.",
"In terms of journals that provide a common place for idea sharing and communication, the journal \"Breast Cancer Research\" is the top journal of topic 1, and its impact factor is 5.49.",
"Authors such as Kurzrock Razelle, and Axelrod Haley in group 4 are the leading researchers sharing ideas on the topic \"targeted therapy.\"",
"The topic 4 is associated with the targeted therapy represented by words like mutations, treatments, therapy and disease.",
"The two most influential journals in topic 4 are \"Oncotarget\" and \"Journal of Thoracic Oncology\" whose impact factors are 6.36 and 5.28 respectively, which indicates that these two journals are the major journals encouraging authors to share ideas and collaborate with each other on cancer targeted therapy subject area.",
"Authors like Zitgel Laurence, Galluzzi Lorenzo, and Kroemer Guido in the author group 7 are the ones that actively share ideas about the topic \"Cancer Immunology.\"",
"Top concepts that are related to this topic are cell, immune, clinical and antitumor.",
"The top journal of the topic \"Cancer Immunology\" is Oncoinmmunology whose impact factor is 6.266.",
"Romagnani Paola and Salem Husein K in topic 8 \"Stem Cell\" are the authors that communicate and share ideas actively with each other in the given field, and the journal \"Stem Cells\" (impact factor: 6.523) is the leading journal.",
"We visualize topic keywords obtained from results of AJT-based topic model.",
"We construct the co-occurrence network and analyze which topic words play an important role in this domain.",
"Each node in the network represents a topic word, and an edge represents a co-occurrence frequency between keywords.",
"The size of nodes represents degree centrality and the color means network clusters obtained by using modularity algorithm.",
"This network consists of 100 nodes and 1,436 edges.",
"As shown in Figure 4 , each topic belongs to a specific community, but shares some important topic keywords.",
"Especially, the topic words positioned at the center is represented core-keywords in Oncology.",
"Figure 4 indicates that these words are the essential concepts of the Oncology domain.",
"Along with the results of AJT-based topic models, we can infer the major journals and authors develop their own research area based on these core-concepts.",
"Fig.",
"4.",
"Network of topic keywords The above results imply that the proposed approach identifies which topics are frequently shared, who facilitates to exchange ideas, and which journals provide a placeholder for it.",
"Identification of the triple relationship among authors, journals, and topics sheds new insight on understanding the well-discussed topics driven by the leading journals and authors that play a mediator role in the development of Oncology.",
"Conclusion One of the major research problems in bibliometrics is how to map out the intellectual structure of a research field.",
"The proposed approach tackles such research problem by utilizing citation sentences and AJT model.",
"By using citation sentences as the input for AJT model to find latent meaning, AJT model suggests a new way to detect leading authors and journals in sub-disciplines represented by discovered topics in a certain field.",
"Achieving this is not feasible by traditional frequency-based citation analysis.",
"One of the interesting observations is that the top-ranked journals in the discovered topics derived from AJT model are not ranked top in terms of JCR.",
"For example, the \"Oncotarget\" journal is the top-ranked journal in three topics in our analysis, but the ranking of the journal is 20 according to JCR.",
"Since we only report on preliminary results of our approach, we undertake in-depth analysis to investigate why this difference exists.",
"We also conduct various statistical tests on the results.",
"Based on the reported results in this paper, though, we claim that AJT can be used for discovering latent meaning associated citation sentences and the major players leading the field.",
"As a follow-up study, we will conduct a comparative study that compares the proposed approach with the general topic modeling technique such as LDA.",
"We also plan to investigate whether there is a different impact of using citation sentences and general meta-data such as abstract and title for topic analysis on facilitating idea sharing and scholarly communication.",
"In addition, we would like to consider the window size of citation sentences enriching citation context and to discover the authors' relationships among the neighboring citation sentences."
]
} | {
"paper_header_number": [
"1",
"2.1",
"2.2",
"2.4",
"4"
],
"paper_header_content": [
"Introduction",
"Main idea",
"Data collection",
"AJT Model",
"Conclusion"
]
} | GEM-SciDuet-train-7#paper-971#slide-0 | Exploring intellectual structures | Collaboration, Author co-citation analysis,
Journal Impact Factor, SJR
Document citation analysis, Co-word analysis,
Citation sentence: Containing brief content of cited work and opinion
that the author of citing work on the cited work
Topic Model: Adopting Author Conference Topic (ACT) model (Tang, Jin
Oncology: The recent surge in number of publications in this field. Stem
cells, one of the subfields of oncology, has been at the forefront of medicine
Tang, J., Jin, R., & Zhang, J. (2008, December). A topic modeling approach and its integration into the random walk framework for academic search. In Data Mining, 2008. ICDM'08. Eighth IEEE International Conference on (pp. 1055-1060). IEEE. | Collaboration, Author co-citation analysis,
Journal Impact Factor, SJR
Document citation analysis, Co-word analysis,
Citation sentence: Containing brief content of cited work and opinion
that the author of citing work on the cited work
Topic Model: Adopting Author Conference Topic (ACT) model (Tang, Jin
Oncology: The recent surge in number of publications in this field. Stem
cells, one of the subfields of oncology, has been at the forefront of medicine
Tang, J., Jin, R., & Zhang, J. (2008, December). A topic modeling approach and its integration into the random walk framework for academic search. In Data Mining, 2008. ICDM'08. Eighth IEEE International Conference on (pp. 1055-1060). IEEE. | [] |
GEM-SciDuet-train-7#paper-971#slide-1 | 971 | Exploring the leading authors and journals in major topics by citation sentences and topic modeling | Citation plays an important role in understanding the knowledge sharing among scholars. Citation sentences embed useful contents that signify the influence of cited authors on shared ideas, and express own opinion of citing authors to others' articles. The purpose of the study is to provide a new lens to analyze the topical relationship embedded in the citation sentences in an integrated manner. To this end, we extract citation sentences from full-text articles in the field of Oncology. In addition, we adopt Author-Journal-Topic (AJT) model to take both authors and journals into consideration of topic analysis. For the study, we collect the 6,360 full-text articles from PubMed Central and select the top 15 journals on Oncology. By applying AJT model, we identify what the major topics are shared among researchers in Oncology and which authors and journal lead the idea exchange in sub-disciplines of Oncology. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106
],
"paper_content_text": [
"Introduction As the size of data on the web continues to increase in an exponential manner, finding valuable meaning between data becomes of paramount importance in many research areas.",
"In the information science field, citations are challenging, pivotal materials to discover the relationship between academic documents because citations present the description of authors' ideas and the hidden relationship between authors and documents.",
"The earliest works focused mainly on classifying the citation behaviors and discovering the citation reasons with limited data such as the location of citation sentences and the number of references [1, 2] .",
"Since the mid-1990s, with the development of computer technology, citation content analysis was elaborated by applying data analysis techniques like text-mining or natural language processing.",
"Zhang et al.",
"[3] present citation analysis based on sematic and syntactic approaches.",
"Semantic-based citation analysis is performed by qualitative analysis to discover the citation motivation and citation classification.",
"On the other hand, syntactic-based citation analysis can be conducted by citation location and citation frequency, which reveals the hidden relation of authors by using meta-data of documents such as journal, venue of publication, affiliation of authors, etc.",
"Following their study, Ding et al.",
"[4] propose a theoretical methodology through content citation analysis.",
"However, these analyses are somewhat limited to the explicit context that primarily represents their own ideas and arguments.",
"The main goal of the paper is to discover the implicit topical relationships buried in citation sentences by utilizing the citation information from the author's perspective of sharing other authors' point of view.",
"Implicitness of the topical relationship is realized by using citation sentences as the input for the topic modeling technique.",
"In this study, a citation sentence indicates the sentence including citation expression consisting of year and author of the cited work.",
"In general, the citation sentence contains brief content of cited work and opinion that the author of citing work on the cited work.",
"We claim that citation sentences reveal interesting characteristics of scholarly communication such as influence, idea exchange, justification for citer's arguments, etc.",
"We assume that using citation sentences for topic analysis reveals aforementioned characteristics.",
"To explore such intellectual space created by citation sentences, we take both authors and journals into consideration of topic analysis.",
"To this end, we applied Author-Conference-Topic (ACT) model proposed by Tang et al.",
"[5] for our topic analysis in relation with both authors and journals, which is called Author-Journal-Topic (AJT) topic model.",
"ACT model is a probabilistic topic model for simultaneously extracting topics of papers, authors, and conferences.",
"There are a few studies to analyze content of citation sentences.",
"Most of previous studies focus on how the topic of document influences citation and vice versa [6, 7, 8] using Topic Modeling.",
"Kataria, Mitra, and Bhatia [8] adapt citation to Author-Topic model [9] with the assumption that the context surrounding the citation anchor could be used to get topical information about the cited authors.",
"These studies including Tang et al.",
"[10] 's ACT model are the examples of combining topic modelling methods and citation content analysis.",
"However, most previous studies used metadata of documents.",
"In this work, we focus on identifying the landscape of the oncology field from a perspective of citation.",
"By using citation sentences, our results can indicate which authors are actively cited and which journals lead a certain topic.",
"The rest of the paper is organized as follows: Section 2 describes the proposed approach.",
"Section 3 analyzes the topic modeling results.",
"Section 4 concludes the paper with the future work.",
"Methodology Main idea The basic assumption of the proposed approach is that citation sentences embed useful contents signifying the influence of cited authors on shared ideas of citing authors.",
"Citation sentences are also considered as an invisible intellectual place for idea exchanging since citations are effective means of supporting and expressing their own arguments by using other works.",
"In the similar vein, Di Marco and Mercer [11] claim that citation sentences play a major role in creating the relationship among relevant authors within the similar research fields.",
"With these assumptions, we are to explore the implicitness of topic relationships resided in citation sentences from the integrated perspective by incorporating the citing authors and journal titles into interpreting the topical relationships.",
"As shown in Figure 1 , we utilized various features including citing authors, citing sentences and journal titles for topic analysis.",
"Authors in Figure 1 mean the citing authors who write a paper and who cite other's work.",
"Citation sentences are the sentences written by the authors when they cite other's work in the paper, and journal titles are the journal names publishing the citing authors' paper.",
"By employing AJT model with these three parameters , we can discover which topics are the most salient ones referred to frequently by researchers and who are the leading authors sharing other authors' ideas in the research field and which journal leads such endeavor.",
"Fig.",
"1.",
"Three parameters for AJT model Data collection For this study, we compile the dataset on the field of Oncology from PubMed Central that provides the full-text in the biomedical field.",
"We select top 15 journals of Oncology by Thomson Reuter's JCR and journal's impact factor, and from these 15 journals, we are able to collect 6,360 full-text articles.",
"Figure 2 describes the workflow of our study.",
"As mentioned earlier, with the fulltext articles collected from PubMed Central, we extract the citation sentences.",
"Most citation sentences are kept in the following format: (author, year), (reference number) [reference number].",
"An example of such format is \"(<xref rid=\"bib00\" ref-type=\"bibr\">Author name, 2000</xref>)\".",
"We use the regular expression technique to parse and extract the citation sentences, when the tag <xref rid=>, </xref> appears on the sentences after parsing XML records with the Java-based SAX parser.",
"Method Fig.",
"2.",
"Workflow We also parse other metadata for AJT model such as the name of authors and journal titles.",
"The author tags, <surname> </surname> and <given-names></given-names> inside the <contrib-gourp></contrib-group>, denote the list of authors who wrote the paper.",
"For journal, we extract the titles when the journal tags, <journal-title> and </journal-title>, are included in the tag of <journal-meta> and </journal-meta>.",
"We also preprocess extracted sentences by removing both functional and general words and applying the Porter's stemming algorithm to improve the input for AJT Model.",
"AJT Model For our study, we apply ACT [10] model with several metadata such as citation sentences, journal titles and citing authors to develop AJT model.",
"Our AJT model utilizes journal titles and citation sentences instead of conference and abstract on documents.",
"The change of model is needed to analyze most influential topics in Oncology and to find leading authors who frequently mention the active topics and to detect the journals involved in such topics.",
"Figure 1 , Table 1) Like ACT model, AJT model assumes that each citing author is related to distribution over topics and each word in citation sentences is derived from a topic.",
"In the AJT model, the journal titles are related to each word.",
"To determine a word (ω_Si) in citation sentences (S), citing authors (x_Si) are consider for a word.",
"Each citing author is associated with a distributed topic.",
"A topic is generated from the citing author-topic distribution.",
"The words and journal titles are generated from a specific topic.",
"AJT model presents (1) 3 Results and Analyses For AJT model, we set the number of topics to 15 and finally select 8 topics as major topics.",
"Since we discovered that there are similar topics on our results, we calculated the similarity between 15 topics to select the most representative topics.",
"The topical similarities are measured by each word on topics and we calculated the similarities of two topics where each topic are represented in an array of a term vector.",
"Through this process, we chose 8 topics which have high topical similarities (over 0.5).",
"Each topic presents top 5 words from topic-word distribution, and 5 most related authors and journal titles are displayed along with each topic.",
"By performing several times on the pilot studies, we decided to choose top 5 words which are quite appropriate to describe each topics.",
"The results of AJT-based topic modeling is shown in Table 1 .",
"We label topic 1 \"breast cancer\" whose top words include breast, expression women, and growth.",
"Since the dataset is compiled with citation sentences, it implies that the topic \"breast cancer\" is a popular topic where researchers share and exchange ideas and facts related to breast cancer.",
"In relation to the topic \"breast cancer\", the active authors of breast cancer are Johnston Stephen RD, Colditz Graham A, and Sternlicht Mark D, and they share ideas with others on breast cancer from our results.",
"In terms of journals that provide a common place for idea sharing and communication, the journal \"Breast Cancer Research\" is the top journal of topic 1, and its impact factor is 5.49.",
"Authors such as Kurzrock Razelle, and Axelrod Haley in group 4 are the leading researchers sharing ideas on the topic \"targeted therapy.\"",
"The topic 4 is associated with the targeted therapy represented by words like mutations, treatments, therapy and disease.",
"The two most influential journals in topic 4 are \"Oncotarget\" and \"Journal of Thoracic Oncology\" whose impact factors are 6.36 and 5.28 respectively, which indicates that these two journals are the major journals encouraging authors to share ideas and collaborate with each other on cancer targeted therapy subject area.",
"Authors like Zitgel Laurence, Galluzzi Lorenzo, and Kroemer Guido in the author group 7 are the ones that actively share ideas about the topic \"Cancer Immunology.\"",
"Top concepts that are related to this topic are cell, immune, clinical and antitumor.",
"The top journal of the topic \"Cancer Immunology\" is Oncoinmmunology whose impact factor is 6.266.",
"Romagnani Paola and Salem Husein K in topic 8 \"Stem Cell\" are the authors that communicate and share ideas actively with each other in the given field, and the journal \"Stem Cells\" (impact factor: 6.523) is the leading journal.",
"We visualize topic keywords obtained from results of AJT-based topic model.",
"We construct the co-occurrence network and analyze which topic words play an important role in this domain.",
"Each node in the network represents a topic word, and an edge represents a co-occurrence frequency between keywords.",
"The size of nodes represents degree centrality and the color means network clusters obtained by using modularity algorithm.",
"This network consists of 100 nodes and 1,436 edges.",
"As shown in Figure 4 , each topic belongs to a specific community, but shares some important topic keywords.",
"Especially, the topic words positioned at the center is represented core-keywords in Oncology.",
"Figure 4 indicates that these words are the essential concepts of the Oncology domain.",
"Along with the results of AJT-based topic models, we can infer the major journals and authors develop their own research area based on these core-concepts.",
"Fig.",
"4.",
"Network of topic keywords The above results imply that the proposed approach identifies which topics are frequently shared, who facilitates to exchange ideas, and which journals provide a placeholder for it.",
"Identification of the triple relationship among authors, journals, and topics sheds new insight on understanding the well-discussed topics driven by the leading journals and authors that play a mediator role in the development of Oncology.",
"Conclusion One of the major research problems in bibliometrics is how to map out the intellectual structure of a research field.",
"The proposed approach tackles such research problem by utilizing citation sentences and AJT model.",
"By using citation sentences as the input for AJT model to find latent meaning, AJT model suggests a new way to detect leading authors and journals in sub-disciplines represented by discovered topics in a certain field.",
"Achieving this is not feasible by traditional frequency-based citation analysis.",
"One of the interesting observations is that the top-ranked journals in the discovered topics derived from AJT model are not ranked top in terms of JCR.",
"For example, the \"Oncotarget\" journal is the top-ranked journal in three topics in our analysis, but the ranking of the journal is 20 according to JCR.",
"Since we only report on preliminary results of our approach, we undertake in-depth analysis to investigate why this difference exists.",
"We also conduct various statistical tests on the results.",
"Based on the reported results in this paper, though, we claim that AJT can be used for discovering latent meaning associated citation sentences and the major players leading the field.",
"As a follow-up study, we will conduct a comparative study that compares the proposed approach with the general topic modeling technique such as LDA.",
"We also plan to investigate whether there is a different impact of using citation sentences and general meta-data such as abstract and title for topic analysis on facilitating idea sharing and scholarly communication.",
"In addition, we would like to consider the window size of citation sentences enriching citation context and to discover the authors' relationships among the neighboring citation sentences."
]
} | {
"paper_header_number": [
"1",
"2.1",
"2.2",
"2.4",
"4"
],
"paper_header_content": [
"Introduction",
"Main idea",
"Data collection",
"AJT Model",
"Conclusion"
]
} | GEM-SciDuet-train-7#paper-971#slide-1 | Citation Sentence | Embedding useful contents signifying the influence of cited authors on
Being considered as an invisible intellectual place for idea exchanging
Playing a role of supporting and expressing their own arguments by
Exploring the implicit topics resided in citation sentences | Embedding useful contents signifying the influence of cited authors on
Being considered as an invisible intellectual place for idea exchanging
Playing a role of supporting and expressing their own arguments by
Exploring the implicit topics resided in citation sentences | [] |
GEM-SciDuet-train-7#paper-971#slide-2 | 971 | Exploring the leading authors and journals in major topics by citation sentences and topic modeling | Citation plays an important role in understanding the knowledge sharing among scholars. Citation sentences embed useful contents that signify the influence of cited authors on shared ideas, and express own opinion of citing authors to others' articles. The purpose of the study is to provide a new lens to analyze the topical relationship embedded in the citation sentences in an integrated manner. To this end, we extract citation sentences from full-text articles in the field of Oncology. In addition, we adopt Author-Journal-Topic (AJT) model to take both authors and journals into consideration of topic analysis. For the study, we collect the 6,360 full-text articles from PubMed Central and select the top 15 journals on Oncology. By applying AJT model, we identify what the major topics are shared among researchers in Oncology and which authors and journal lead the idea exchange in sub-disciplines of Oncology. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106
],
"paper_content_text": [
"Introduction As the size of data on the web continues to increase in an exponential manner, finding valuable meaning between data becomes of paramount importance in many research areas.",
"In the information science field, citations are challenging, pivotal materials to discover the relationship between academic documents because citations present the description of authors' ideas and the hidden relationship between authors and documents.",
"The earliest works focused mainly on classifying the citation behaviors and discovering the citation reasons with limited data such as the location of citation sentences and the number of references [1, 2] .",
"Since the mid-1990s, with the development of computer technology, citation content analysis was elaborated by applying data analysis techniques like text-mining or natural language processing.",
"Zhang et al.",
"[3] present citation analysis based on sematic and syntactic approaches.",
"Semantic-based citation analysis is performed by qualitative analysis to discover the citation motivation and citation classification.",
"On the other hand, syntactic-based citation analysis can be conducted by citation location and citation frequency, which reveals the hidden relation of authors by using meta-data of documents such as journal, venue of publication, affiliation of authors, etc.",
"Following their study, Ding et al.",
"[4] propose a theoretical methodology through content citation analysis.",
"However, these analyses are somewhat limited to the explicit context that primarily represents their own ideas and arguments.",
"The main goal of the paper is to discover the implicit topical relationships buried in citation sentences by utilizing the citation information from the author's perspective of sharing other authors' point of view.",
"Implicitness of the topical relationship is realized by using citation sentences as the input for the topic modeling technique.",
"In this study, a citation sentence indicates the sentence including citation expression consisting of year and author of the cited work.",
"In general, the citation sentence contains brief content of cited work and opinion that the author of citing work on the cited work.",
"We claim that citation sentences reveal interesting characteristics of scholarly communication such as influence, idea exchange, justification for citer's arguments, etc.",
"We assume that using citation sentences for topic analysis reveals aforementioned characteristics.",
"To explore such intellectual space created by citation sentences, we take both authors and journals into consideration of topic analysis.",
"To this end, we applied Author-Conference-Topic (ACT) model proposed by Tang et al.",
"[5] for our topic analysis in relation with both authors and journals, which is called Author-Journal-Topic (AJT) topic model.",
"ACT model is a probabilistic topic model for simultaneously extracting topics of papers, authors, and conferences.",
"There are a few studies to analyze content of citation sentences.",
"Most of previous studies focus on how the topic of document influences citation and vice versa [6, 7, 8] using Topic Modeling.",
"Kataria, Mitra, and Bhatia [8] adapt citation to Author-Topic model [9] with the assumption that the context surrounding the citation anchor could be used to get topical information about the cited authors.",
"These studies including Tang et al.",
"[10] 's ACT model are the examples of combining topic modelling methods and citation content analysis.",
"However, most previous studies used metadata of documents.",
"In this work, we focus on identifying the landscape of the oncology field from a perspective of citation.",
"By using citation sentences, our results can indicate which authors are actively cited and which journals lead a certain topic.",
"The rest of the paper is organized as follows: Section 2 describes the proposed approach.",
"Section 3 analyzes the topic modeling results.",
"Section 4 concludes the paper with the future work.",
"Methodology Main idea The basic assumption of the proposed approach is that citation sentences embed useful contents signifying the influence of cited authors on shared ideas of citing authors.",
"Citation sentences are also considered as an invisible intellectual place for idea exchanging since citations are effective means of supporting and expressing their own arguments by using other works.",
"In the similar vein, Di Marco and Mercer [11] claim that citation sentences play a major role in creating the relationship among relevant authors within the similar research fields.",
"With these assumptions, we are to explore the implicitness of topic relationships resided in citation sentences from the integrated perspective by incorporating the citing authors and journal titles into interpreting the topical relationships.",
"As shown in Figure 1 , we utilized various features including citing authors, citing sentences and journal titles for topic analysis.",
"Authors in Figure 1 mean the citing authors who write a paper and who cite other's work.",
"Citation sentences are the sentences written by the authors when they cite other's work in the paper, and journal titles are the journal names publishing the citing authors' paper.",
"By employing AJT model with these three parameters , we can discover which topics are the most salient ones referred to frequently by researchers and who are the leading authors sharing other authors' ideas in the research field and which journal leads such endeavor.",
"Fig.",
"1.",
"Three parameters for AJT model Data collection For this study, we compile the dataset on the field of Oncology from PubMed Central that provides the full-text in the biomedical field.",
"We select top 15 journals of Oncology by Thomson Reuter's JCR and journal's impact factor, and from these 15 journals, we are able to collect 6,360 full-text articles.",
"Figure 2 describes the workflow of our study.",
"As mentioned earlier, with the fulltext articles collected from PubMed Central, we extract the citation sentences.",
"Most citation sentences are kept in the following format: (author, year), (reference number) [reference number].",
"An example of such format is \"(<xref rid=\"bib00\" ref-type=\"bibr\">Author name, 2000</xref>)\".",
"We use the regular expression technique to parse and extract the citation sentences, when the tag <xref rid=>, </xref> appears on the sentences after parsing XML records with the Java-based SAX parser.",
"Method Fig.",
"2.",
"Workflow We also parse other metadata for AJT model such as the name of authors and journal titles.",
"The author tags, <surname> </surname> and <given-names></given-names> inside the <contrib-gourp></contrib-group>, denote the list of authors who wrote the paper.",
"For journal, we extract the titles when the journal tags, <journal-title> and </journal-title>, are included in the tag of <journal-meta> and </journal-meta>.",
"We also preprocess extracted sentences by removing both functional and general words and applying the Porter's stemming algorithm to improve the input for AJT Model.",
"AJT Model For our study, we apply ACT [10] model with several metadata such as citation sentences, journal titles and citing authors to develop AJT model.",
"Our AJT model utilizes journal titles and citation sentences instead of conference and abstract on documents.",
"The change of model is needed to analyze most influential topics in Oncology and to find leading authors who frequently mention the active topics and to detect the journals involved in such topics.",
"Figure 1 , Table 1) Like ACT model, AJT model assumes that each citing author is related to distribution over topics and each word in citation sentences is derived from a topic.",
"In the AJT model, the journal titles are related to each word.",
"To determine a word (ω_Si) in citation sentences (S), citing authors (x_Si) are consider for a word.",
"Each citing author is associated with a distributed topic.",
"A topic is generated from the citing author-topic distribution.",
"The words and journal titles are generated from a specific topic.",
"AJT model presents (1) 3 Results and Analyses For AJT model, we set the number of topics to 15 and finally select 8 topics as major topics.",
"Since we discovered that there are similar topics on our results, we calculated the similarity between 15 topics to select the most representative topics.",
"The topical similarities are measured by each word on topics and we calculated the similarities of two topics where each topic are represented in an array of a term vector.",
"Through this process, we chose 8 topics which have high topical similarities (over 0.5).",
"Each topic presents top 5 words from topic-word distribution, and 5 most related authors and journal titles are displayed along with each topic.",
"By performing several times on the pilot studies, we decided to choose top 5 words which are quite appropriate to describe each topics.",
"The results of AJT-based topic modeling is shown in Table 1 .",
"We label topic 1 \"breast cancer\" whose top words include breast, expression women, and growth.",
"Since the dataset is compiled with citation sentences, it implies that the topic \"breast cancer\" is a popular topic where researchers share and exchange ideas and facts related to breast cancer.",
"In relation to the topic \"breast cancer\", the active authors of breast cancer are Johnston Stephen RD, Colditz Graham A, and Sternlicht Mark D, and they share ideas with others on breast cancer from our results.",
"In terms of journals that provide a common place for idea sharing and communication, the journal \"Breast Cancer Research\" is the top journal of topic 1, and its impact factor is 5.49.",
"Authors such as Kurzrock Razelle, and Axelrod Haley in group 4 are the leading researchers sharing ideas on the topic \"targeted therapy.\"",
"The topic 4 is associated with the targeted therapy represented by words like mutations, treatments, therapy and disease.",
"The two most influential journals in topic 4 are \"Oncotarget\" and \"Journal of Thoracic Oncology\" whose impact factors are 6.36 and 5.28 respectively, which indicates that these two journals are the major journals encouraging authors to share ideas and collaborate with each other on cancer targeted therapy subject area.",
"Authors like Zitgel Laurence, Galluzzi Lorenzo, and Kroemer Guido in the author group 7 are the ones that actively share ideas about the topic \"Cancer Immunology.\"",
"Top concepts that are related to this topic are cell, immune, clinical and antitumor.",
"The top journal of the topic \"Cancer Immunology\" is Oncoinmmunology whose impact factor is 6.266.",
"Romagnani Paola and Salem Husein K in topic 8 \"Stem Cell\" are the authors that communicate and share ideas actively with each other in the given field, and the journal \"Stem Cells\" (impact factor: 6.523) is the leading journal.",
"We visualize topic keywords obtained from results of AJT-based topic model.",
"We construct the co-occurrence network and analyze which topic words play an important role in this domain.",
"Each node in the network represents a topic word, and an edge represents a co-occurrence frequency between keywords.",
"The size of nodes represents degree centrality and the color means network clusters obtained by using modularity algorithm.",
"This network consists of 100 nodes and 1,436 edges.",
"As shown in Figure 4 , each topic belongs to a specific community, but shares some important topic keywords.",
"Especially, the topic words positioned at the center is represented core-keywords in Oncology.",
"Figure 4 indicates that these words are the essential concepts of the Oncology domain.",
"Along with the results of AJT-based topic models, we can infer the major journals and authors develop their own research area based on these core-concepts.",
"Fig.",
"4.",
"Network of topic keywords The above results imply that the proposed approach identifies which topics are frequently shared, who facilitates to exchange ideas, and which journals provide a placeholder for it.",
"Identification of the triple relationship among authors, journals, and topics sheds new insight on understanding the well-discussed topics driven by the leading journals and authors that play a mediator role in the development of Oncology.",
"Conclusion One of the major research problems in bibliometrics is how to map out the intellectual structure of a research field.",
"The proposed approach tackles such research problem by utilizing citation sentences and AJT model.",
"By using citation sentences as the input for AJT model to find latent meaning, AJT model suggests a new way to detect leading authors and journals in sub-disciplines represented by discovered topics in a certain field.",
"Achieving this is not feasible by traditional frequency-based citation analysis.",
"One of the interesting observations is that the top-ranked journals in the discovered topics derived from AJT model are not ranked top in terms of JCR.",
"For example, the \"Oncotarget\" journal is the top-ranked journal in three topics in our analysis, but the ranking of the journal is 20 according to JCR.",
"Since we only report on preliminary results of our approach, we undertake in-depth analysis to investigate why this difference exists.",
"We also conduct various statistical tests on the results.",
"Based on the reported results in this paper, though, we claim that AJT can be used for discovering latent meaning associated citation sentences and the major players leading the field.",
"As a follow-up study, we will conduct a comparative study that compares the proposed approach with the general topic modeling technique such as LDA.",
"We also plan to investigate whether there is a different impact of using citation sentences and general meta-data such as abstract and title for topic analysis on facilitating idea sharing and scholarly communication.",
"In addition, we would like to consider the window size of citation sentences enriching citation context and to discover the authors' relationships among the neighboring citation sentences."
]
} | {
"paper_header_number": [
"1",
"2.1",
"2.2",
"2.4",
"4"
],
"paper_header_content": [
"Introduction",
"Main idea",
"Data collection",
"AJT Model",
"Conclusion"
]
} | GEM-SciDuet-train-7#paper-971#slide-2 | Original ACT Model Tang Jin and Zhang 2008 | Purpose of Academic search | Purpose of Academic search | [] |
GEM-SciDuet-train-7#paper-971#slide-3 | 971 | Exploring the leading authors and journals in major topics by citation sentences and topic modeling | Citation plays an important role in understanding the knowledge sharing among scholars. Citation sentences embed useful contents that signify the influence of cited authors on shared ideas, and express own opinion of citing authors to others' articles. The purpose of the study is to provide a new lens to analyze the topical relationship embedded in the citation sentences in an integrated manner. To this end, we extract citation sentences from full-text articles in the field of Oncology. In addition, we adopt Author-Journal-Topic (AJT) model to take both authors and journals into consideration of topic analysis. For the study, we collect the 6,360 full-text articles from PubMed Central and select the top 15 journals on Oncology. By applying AJT model, we identify what the major topics are shared among researchers in Oncology and which authors and journal lead the idea exchange in sub-disciplines of Oncology. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106
],
"paper_content_text": [
"Introduction As the size of data on the web continues to increase in an exponential manner, finding valuable meaning between data becomes of paramount importance in many research areas.",
"In the information science field, citations are challenging, pivotal materials to discover the relationship between academic documents because citations present the description of authors' ideas and the hidden relationship between authors and documents.",
"The earliest works focused mainly on classifying the citation behaviors and discovering the citation reasons with limited data such as the location of citation sentences and the number of references [1, 2] .",
"Since the mid-1990s, with the development of computer technology, citation content analysis was elaborated by applying data analysis techniques like text-mining or natural language processing.",
"Zhang et al.",
"[3] present citation analysis based on sematic and syntactic approaches.",
"Semantic-based citation analysis is performed by qualitative analysis to discover the citation motivation and citation classification.",
"On the other hand, syntactic-based citation analysis can be conducted by citation location and citation frequency, which reveals the hidden relation of authors by using meta-data of documents such as journal, venue of publication, affiliation of authors, etc.",
"Following their study, Ding et al.",
"[4] propose a theoretical methodology through content citation analysis.",
"However, these analyses are somewhat limited to the explicit context that primarily represents their own ideas and arguments.",
"The main goal of the paper is to discover the implicit topical relationships buried in citation sentences by utilizing the citation information from the author's perspective of sharing other authors' point of view.",
"Implicitness of the topical relationship is realized by using citation sentences as the input for the topic modeling technique.",
"In this study, a citation sentence indicates the sentence including citation expression consisting of year and author of the cited work.",
"In general, the citation sentence contains brief content of cited work and opinion that the author of citing work on the cited work.",
"We claim that citation sentences reveal interesting characteristics of scholarly communication such as influence, idea exchange, justification for citer's arguments, etc.",
"We assume that using citation sentences for topic analysis reveals aforementioned characteristics.",
"To explore such intellectual space created by citation sentences, we take both authors and journals into consideration of topic analysis.",
"To this end, we applied Author-Conference-Topic (ACT) model proposed by Tang et al.",
"[5] for our topic analysis in relation with both authors and journals, which is called Author-Journal-Topic (AJT) topic model.",
"ACT model is a probabilistic topic model for simultaneously extracting topics of papers, authors, and conferences.",
"There are a few studies to analyze content of citation sentences.",
"Most of previous studies focus on how the topic of document influences citation and vice versa [6, 7, 8] using Topic Modeling.",
"Kataria, Mitra, and Bhatia [8] adapt citation to Author-Topic model [9] with the assumption that the context surrounding the citation anchor could be used to get topical information about the cited authors.",
"These studies including Tang et al.",
"[10] 's ACT model are the examples of combining topic modelling methods and citation content analysis.",
"However, most previous studies used metadata of documents.",
"In this work, we focus on identifying the landscape of the oncology field from a perspective of citation.",
"By using citation sentences, our results can indicate which authors are actively cited and which journals lead a certain topic.",
"The rest of the paper is organized as follows: Section 2 describes the proposed approach.",
"Section 3 analyzes the topic modeling results.",
"Section 4 concludes the paper with the future work.",
"Methodology Main idea The basic assumption of the proposed approach is that citation sentences embed useful contents signifying the influence of cited authors on shared ideas of citing authors.",
"Citation sentences are also considered as an invisible intellectual place for idea exchanging since citations are effective means of supporting and expressing their own arguments by using other works.",
"In the similar vein, Di Marco and Mercer [11] claim that citation sentences play a major role in creating the relationship among relevant authors within the similar research fields.",
"With these assumptions, we are to explore the implicitness of topic relationships resided in citation sentences from the integrated perspective by incorporating the citing authors and journal titles into interpreting the topical relationships.",
"As shown in Figure 1 , we utilized various features including citing authors, citing sentences and journal titles for topic analysis.",
"Authors in Figure 1 mean the citing authors who write a paper and who cite other's work.",
"Citation sentences are the sentences written by the authors when they cite other's work in the paper, and journal titles are the journal names publishing the citing authors' paper.",
"By employing AJT model with these three parameters , we can discover which topics are the most salient ones referred to frequently by researchers and who are the leading authors sharing other authors' ideas in the research field and which journal leads such endeavor.",
"Fig.",
"1.",
"Three parameters for AJT model Data collection For this study, we compile the dataset on the field of Oncology from PubMed Central that provides the full-text in the biomedical field.",
"We select top 15 journals of Oncology by Thomson Reuter's JCR and journal's impact factor, and from these 15 journals, we are able to collect 6,360 full-text articles.",
"Figure 2 describes the workflow of our study.",
"As mentioned earlier, with the fulltext articles collected from PubMed Central, we extract the citation sentences.",
"Most citation sentences are kept in the following format: (author, year), (reference number) [reference number].",
"An example of such format is \"(<xref rid=\"bib00\" ref-type=\"bibr\">Author name, 2000</xref>)\".",
"We use the regular expression technique to parse and extract the citation sentences, when the tag <xref rid=>, </xref> appears on the sentences after parsing XML records with the Java-based SAX parser.",
"Method Fig.",
"2.",
"Workflow We also parse other metadata for AJT model such as the name of authors and journal titles.",
"The author tags, <surname> </surname> and <given-names></given-names> inside the <contrib-gourp></contrib-group>, denote the list of authors who wrote the paper.",
"For journal, we extract the titles when the journal tags, <journal-title> and </journal-title>, are included in the tag of <journal-meta> and </journal-meta>.",
"We also preprocess extracted sentences by removing both functional and general words and applying the Porter's stemming algorithm to improve the input for AJT Model.",
"AJT Model For our study, we apply ACT [10] model with several metadata such as citation sentences, journal titles and citing authors to develop AJT model.",
"Our AJT model utilizes journal titles and citation sentences instead of conference and abstract on documents.",
"The change of model is needed to analyze most influential topics in Oncology and to find leading authors who frequently mention the active topics and to detect the journals involved in such topics.",
"Figure 1 , Table 1) Like ACT model, AJT model assumes that each citing author is related to distribution over topics and each word in citation sentences is derived from a topic.",
"In the AJT model, the journal titles are related to each word.",
"To determine a word (ω_Si) in citation sentences (S), citing authors (x_Si) are consider for a word.",
"Each citing author is associated with a distributed topic.",
"A topic is generated from the citing author-topic distribution.",
"The words and journal titles are generated from a specific topic.",
"AJT model presents (1) 3 Results and Analyses For AJT model, we set the number of topics to 15 and finally select 8 topics as major topics.",
"Since we discovered that there are similar topics on our results, we calculated the similarity between 15 topics to select the most representative topics.",
"The topical similarities are measured by each word on topics and we calculated the similarities of two topics where each topic are represented in an array of a term vector.",
"Through this process, we chose 8 topics which have high topical similarities (over 0.5).",
"Each topic presents top 5 words from topic-word distribution, and 5 most related authors and journal titles are displayed along with each topic.",
"By performing several times on the pilot studies, we decided to choose top 5 words which are quite appropriate to describe each topics.",
"The results of AJT-based topic modeling is shown in Table 1 .",
"We label topic 1 \"breast cancer\" whose top words include breast, expression women, and growth.",
"Since the dataset is compiled with citation sentences, it implies that the topic \"breast cancer\" is a popular topic where researchers share and exchange ideas and facts related to breast cancer.",
"In relation to the topic \"breast cancer\", the active authors of breast cancer are Johnston Stephen RD, Colditz Graham A, and Sternlicht Mark D, and they share ideas with others on breast cancer from our results.",
"In terms of journals that provide a common place for idea sharing and communication, the journal \"Breast Cancer Research\" is the top journal of topic 1, and its impact factor is 5.49.",
"Authors such as Kurzrock Razelle, and Axelrod Haley in group 4 are the leading researchers sharing ideas on the topic \"targeted therapy.\"",
"The topic 4 is associated with the targeted therapy represented by words like mutations, treatments, therapy and disease.",
"The two most influential journals in topic 4 are \"Oncotarget\" and \"Journal of Thoracic Oncology\" whose impact factors are 6.36 and 5.28 respectively, which indicates that these two journals are the major journals encouraging authors to share ideas and collaborate with each other on cancer targeted therapy subject area.",
"Authors like Zitgel Laurence, Galluzzi Lorenzo, and Kroemer Guido in the author group 7 are the ones that actively share ideas about the topic \"Cancer Immunology.\"",
"Top concepts that are related to this topic are cell, immune, clinical and antitumor.",
"The top journal of the topic \"Cancer Immunology\" is Oncoinmmunology whose impact factor is 6.266.",
"Romagnani Paola and Salem Husein K in topic 8 \"Stem Cell\" are the authors that communicate and share ideas actively with each other in the given field, and the journal \"Stem Cells\" (impact factor: 6.523) is the leading journal.",
"We visualize topic keywords obtained from results of AJT-based topic model.",
"We construct the co-occurrence network and analyze which topic words play an important role in this domain.",
"Each node in the network represents a topic word, and an edge represents a co-occurrence frequency between keywords.",
"The size of nodes represents degree centrality and the color means network clusters obtained by using modularity algorithm.",
"This network consists of 100 nodes and 1,436 edges.",
"As shown in Figure 4 , each topic belongs to a specific community, but shares some important topic keywords.",
"Especially, the topic words positioned at the center is represented core-keywords in Oncology.",
"Figure 4 indicates that these words are the essential concepts of the Oncology domain.",
"Along with the results of AJT-based topic models, we can infer the major journals and authors develop their own research area based on these core-concepts.",
"Fig.",
"4.",
"Network of topic keywords The above results imply that the proposed approach identifies which topics are frequently shared, who facilitates to exchange ideas, and which journals provide a placeholder for it.",
"Identification of the triple relationship among authors, journals, and topics sheds new insight on understanding the well-discussed topics driven by the leading journals and authors that play a mediator role in the development of Oncology.",
"Conclusion One of the major research problems in bibliometrics is how to map out the intellectual structure of a research field.",
"The proposed approach tackles such research problem by utilizing citation sentences and AJT model.",
"By using citation sentences as the input for AJT model to find latent meaning, AJT model suggests a new way to detect leading authors and journals in sub-disciplines represented by discovered topics in a certain field.",
"Achieving this is not feasible by traditional frequency-based citation analysis.",
"One of the interesting observations is that the top-ranked journals in the discovered topics derived from AJT model are not ranked top in terms of JCR.",
"For example, the \"Oncotarget\" journal is the top-ranked journal in three topics in our analysis, but the ranking of the journal is 20 according to JCR.",
"Since we only report on preliminary results of our approach, we undertake in-depth analysis to investigate why this difference exists.",
"We also conduct various statistical tests on the results.",
"Based on the reported results in this paper, though, we claim that AJT can be used for discovering latent meaning associated citation sentences and the major players leading the field.",
"As a follow-up study, we will conduct a comparative study that compares the proposed approach with the general topic modeling technique such as LDA.",
"We also plan to investigate whether there is a different impact of using citation sentences and general meta-data such as abstract and title for topic analysis on facilitating idea sharing and scholarly communication.",
"In addition, we would like to consider the window size of citation sentences enriching citation context and to discover the authors' relationships among the neighboring citation sentences."
]
} | {
"paper_header_number": [
"1",
"2.1",
"2.2",
"2.4",
"4"
],
"paper_header_content": [
"Introduction",
"Main idea",
"Data collection",
"AJT Model",
"Conclusion"
]
} | GEM-SciDuet-train-7#paper-971#slide-3 | Modified AJT Model | 1) Citation Data Extraction
2n d journal Topic 2
Which topic is most salient? Who is the active authors sharing other authors ideas? Which journal leads such endeavor? | 1) Citation Data Extraction
2n d journal Topic 2
Which topic is most salient? Who is the active authors sharing other authors ideas? Which journal leads such endeavor? | [] |
GEM-SciDuet-train-7#paper-971#slide-4 | 971 | Exploring the leading authors and journals in major topics by citation sentences and topic modeling | Citation plays an important role in understanding the knowledge sharing among scholars. Citation sentences embed useful contents that signify the influence of cited authors on shared ideas, and express own opinion of citing authors to others' articles. The purpose of the study is to provide a new lens to analyze the topical relationship embedded in the citation sentences in an integrated manner. To this end, we extract citation sentences from full-text articles in the field of Oncology. In addition, we adopt Author-Journal-Topic (AJT) model to take both authors and journals into consideration of topic analysis. For the study, we collect the 6,360 full-text articles from PubMed Central and select the top 15 journals on Oncology. By applying AJT model, we identify what the major topics are shared among researchers in Oncology and which authors and journal lead the idea exchange in sub-disciplines of Oncology. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106
],
"paper_content_text": [
"Introduction As the size of data on the web continues to increase in an exponential manner, finding valuable meaning between data becomes of paramount importance in many research areas.",
"In the information science field, citations are challenging, pivotal materials to discover the relationship between academic documents because citations present the description of authors' ideas and the hidden relationship between authors and documents.",
"The earliest works focused mainly on classifying the citation behaviors and discovering the citation reasons with limited data such as the location of citation sentences and the number of references [1, 2] .",
"Since the mid-1990s, with the development of computer technology, citation content analysis was elaborated by applying data analysis techniques like text-mining or natural language processing.",
"Zhang et al.",
"[3] present citation analysis based on sematic and syntactic approaches.",
"Semantic-based citation analysis is performed by qualitative analysis to discover the citation motivation and citation classification.",
"On the other hand, syntactic-based citation analysis can be conducted by citation location and citation frequency, which reveals the hidden relation of authors by using meta-data of documents such as journal, venue of publication, affiliation of authors, etc.",
"Following their study, Ding et al.",
"[4] propose a theoretical methodology through content citation analysis.",
"However, these analyses are somewhat limited to the explicit context that primarily represents their own ideas and arguments.",
"The main goal of the paper is to discover the implicit topical relationships buried in citation sentences by utilizing the citation information from the author's perspective of sharing other authors' point of view.",
"Implicitness of the topical relationship is realized by using citation sentences as the input for the topic modeling technique.",
"In this study, a citation sentence indicates the sentence including citation expression consisting of year and author of the cited work.",
"In general, the citation sentence contains brief content of cited work and opinion that the author of citing work on the cited work.",
"We claim that citation sentences reveal interesting characteristics of scholarly communication such as influence, idea exchange, justification for citer's arguments, etc.",
"We assume that using citation sentences for topic analysis reveals aforementioned characteristics.",
"To explore such intellectual space created by citation sentences, we take both authors and journals into consideration of topic analysis.",
"To this end, we applied Author-Conference-Topic (ACT) model proposed by Tang et al.",
"[5] for our topic analysis in relation with both authors and journals, which is called Author-Journal-Topic (AJT) topic model.",
"ACT model is a probabilistic topic model for simultaneously extracting topics of papers, authors, and conferences.",
"There are a few studies to analyze content of citation sentences.",
"Most of previous studies focus on how the topic of document influences citation and vice versa [6, 7, 8] using Topic Modeling.",
"Kataria, Mitra, and Bhatia [8] adapt citation to Author-Topic model [9] with the assumption that the context surrounding the citation anchor could be used to get topical information about the cited authors.",
"These studies including Tang et al.",
"[10] 's ACT model are the examples of combining topic modelling methods and citation content analysis.",
"However, most previous studies used metadata of documents.",
"In this work, we focus on identifying the landscape of the oncology field from a perspective of citation.",
"By using citation sentences, our results can indicate which authors are actively cited and which journals lead a certain topic.",
"The rest of the paper is organized as follows: Section 2 describes the proposed approach.",
"Section 3 analyzes the topic modeling results.",
"Section 4 concludes the paper with the future work.",
"Methodology Main idea The basic assumption of the proposed approach is that citation sentences embed useful contents signifying the influence of cited authors on shared ideas of citing authors.",
"Citation sentences are also considered as an invisible intellectual place for idea exchanging since citations are effective means of supporting and expressing their own arguments by using other works.",
"In the similar vein, Di Marco and Mercer [11] claim that citation sentences play a major role in creating the relationship among relevant authors within the similar research fields.",
"With these assumptions, we are to explore the implicitness of topic relationships resided in citation sentences from the integrated perspective by incorporating the citing authors and journal titles into interpreting the topical relationships.",
"As shown in Figure 1 , we utilized various features including citing authors, citing sentences and journal titles for topic analysis.",
"Authors in Figure 1 mean the citing authors who write a paper and who cite other's work.",
"Citation sentences are the sentences written by the authors when they cite other's work in the paper, and journal titles are the journal names publishing the citing authors' paper.",
"By employing AJT model with these three parameters , we can discover which topics are the most salient ones referred to frequently by researchers and who are the leading authors sharing other authors' ideas in the research field and which journal leads such endeavor.",
"Fig.",
"1.",
"Three parameters for AJT model Data collection For this study, we compile the dataset on the field of Oncology from PubMed Central that provides the full-text in the biomedical field.",
"We select top 15 journals of Oncology by Thomson Reuter's JCR and journal's impact factor, and from these 15 journals, we are able to collect 6,360 full-text articles.",
"Figure 2 describes the workflow of our study.",
"As mentioned earlier, with the fulltext articles collected from PubMed Central, we extract the citation sentences.",
"Most citation sentences are kept in the following format: (author, year), (reference number) [reference number].",
"An example of such format is \"(<xref rid=\"bib00\" ref-type=\"bibr\">Author name, 2000</xref>)\".",
"We use the regular expression technique to parse and extract the citation sentences, when the tag <xref rid=>, </xref> appears on the sentences after parsing XML records with the Java-based SAX parser.",
"Method Fig.",
"2.",
"Workflow We also parse other metadata for AJT model such as the name of authors and journal titles.",
"The author tags, <surname> </surname> and <given-names></given-names> inside the <contrib-gourp></contrib-group>, denote the list of authors who wrote the paper.",
"For journal, we extract the titles when the journal tags, <journal-title> and </journal-title>, are included in the tag of <journal-meta> and </journal-meta>.",
"We also preprocess extracted sentences by removing both functional and general words and applying the Porter's stemming algorithm to improve the input for AJT Model.",
"AJT Model For our study, we apply ACT [10] model with several metadata such as citation sentences, journal titles and citing authors to develop AJT model.",
"Our AJT model utilizes journal titles and citation sentences instead of conference and abstract on documents.",
"The change of model is needed to analyze most influential topics in Oncology and to find leading authors who frequently mention the active topics and to detect the journals involved in such topics.",
"Figure 1 , Table 1) Like ACT model, AJT model assumes that each citing author is related to distribution over topics and each word in citation sentences is derived from a topic.",
"In the AJT model, the journal titles are related to each word.",
"To determine a word (ω_Si) in citation sentences (S), citing authors (x_Si) are consider for a word.",
"Each citing author is associated with a distributed topic.",
"A topic is generated from the citing author-topic distribution.",
"The words and journal titles are generated from a specific topic.",
"AJT model presents (1) 3 Results and Analyses For AJT model, we set the number of topics to 15 and finally select 8 topics as major topics.",
"Since we discovered that there are similar topics on our results, we calculated the similarity between 15 topics to select the most representative topics.",
"The topical similarities are measured by each word on topics and we calculated the similarities of two topics where each topic are represented in an array of a term vector.",
"Through this process, we chose 8 topics which have high topical similarities (over 0.5).",
"Each topic presents top 5 words from topic-word distribution, and 5 most related authors and journal titles are displayed along with each topic.",
"By performing several times on the pilot studies, we decided to choose top 5 words which are quite appropriate to describe each topics.",
"The results of AJT-based topic modeling is shown in Table 1 .",
"We label topic 1 \"breast cancer\" whose top words include breast, expression women, and growth.",
"Since the dataset is compiled with citation sentences, it implies that the topic \"breast cancer\" is a popular topic where researchers share and exchange ideas and facts related to breast cancer.",
"In relation to the topic \"breast cancer\", the active authors of breast cancer are Johnston Stephen RD, Colditz Graham A, and Sternlicht Mark D, and they share ideas with others on breast cancer from our results.",
"In terms of journals that provide a common place for idea sharing and communication, the journal \"Breast Cancer Research\" is the top journal of topic 1, and its impact factor is 5.49.",
"Authors such as Kurzrock Razelle, and Axelrod Haley in group 4 are the leading researchers sharing ideas on the topic \"targeted therapy.\"",
"The topic 4 is associated with the targeted therapy represented by words like mutations, treatments, therapy and disease.",
"The two most influential journals in topic 4 are \"Oncotarget\" and \"Journal of Thoracic Oncology\" whose impact factors are 6.36 and 5.28 respectively, which indicates that these two journals are the major journals encouraging authors to share ideas and collaborate with each other on cancer targeted therapy subject area.",
"Authors like Zitgel Laurence, Galluzzi Lorenzo, and Kroemer Guido in the author group 7 are the ones that actively share ideas about the topic \"Cancer Immunology.\"",
"Top concepts that are related to this topic are cell, immune, clinical and antitumor.",
"The top journal of the topic \"Cancer Immunology\" is Oncoinmmunology whose impact factor is 6.266.",
"Romagnani Paola and Salem Husein K in topic 8 \"Stem Cell\" are the authors that communicate and share ideas actively with each other in the given field, and the journal \"Stem Cells\" (impact factor: 6.523) is the leading journal.",
"We visualize topic keywords obtained from results of AJT-based topic model.",
"We construct the co-occurrence network and analyze which topic words play an important role in this domain.",
"Each node in the network represents a topic word, and an edge represents a co-occurrence frequency between keywords.",
"The size of nodes represents degree centrality and the color means network clusters obtained by using modularity algorithm.",
"This network consists of 100 nodes and 1,436 edges.",
"As shown in Figure 4 , each topic belongs to a specific community, but shares some important topic keywords.",
"Especially, the topic words positioned at the center is represented core-keywords in Oncology.",
"Figure 4 indicates that these words are the essential concepts of the Oncology domain.",
"Along with the results of AJT-based topic models, we can infer the major journals and authors develop their own research area based on these core-concepts.",
"Fig.",
"4.",
"Network of topic keywords The above results imply that the proposed approach identifies which topics are frequently shared, who facilitates to exchange ideas, and which journals provide a placeholder for it.",
"Identification of the triple relationship among authors, journals, and topics sheds new insight on understanding the well-discussed topics driven by the leading journals and authors that play a mediator role in the development of Oncology.",
"Conclusion One of the major research problems in bibliometrics is how to map out the intellectual structure of a research field.",
"The proposed approach tackles such research problem by utilizing citation sentences and AJT model.",
"By using citation sentences as the input for AJT model to find latent meaning, AJT model suggests a new way to detect leading authors and journals in sub-disciplines represented by discovered topics in a certain field.",
"Achieving this is not feasible by traditional frequency-based citation analysis.",
"One of the interesting observations is that the top-ranked journals in the discovered topics derived from AJT model are not ranked top in terms of JCR.",
"For example, the \"Oncotarget\" journal is the top-ranked journal in three topics in our analysis, but the ranking of the journal is 20 according to JCR.",
"Since we only report on preliminary results of our approach, we undertake in-depth analysis to investigate why this difference exists.",
"We also conduct various statistical tests on the results.",
"Based on the reported results in this paper, though, we claim that AJT can be used for discovering latent meaning associated citation sentences and the major players leading the field.",
"As a follow-up study, we will conduct a comparative study that compares the proposed approach with the general topic modeling technique such as LDA.",
"We also plan to investigate whether there is a different impact of using citation sentences and general meta-data such as abstract and title for topic analysis on facilitating idea sharing and scholarly communication.",
"In addition, we would like to consider the window size of citation sentences enriching citation context and to discover the authors' relationships among the neighboring citation sentences."
]
} | {
"paper_header_number": [
"1",
"2.1",
"2.2",
"2.4",
"4"
],
"paper_header_content": [
"Introduction",
"Main idea",
"Data collection",
"AJT Model",
"Conclusion"
]
} | GEM-SciDuet-train-7#paper-971#slide-4 | Method | The 77-SNP PRS was associated with a larg er effect
than previously reported for a 10-SNP-PRS (<xref 3) Citing Authors rid=CIT0020 ref-type=bibr> 20 </xref>). | The 77-SNP PRS was associated with a larg er effect
than previously reported for a 10-SNP-PRS (<xref 3) Citing Authors rid=CIT0020 ref-type=bibr> 20 </xref>). | [] |
GEM-SciDuet-train-7#paper-971#slide-5 | 971 | Exploring the leading authors and journals in major topics by citation sentences and topic modeling | Citation plays an important role in understanding the knowledge sharing among scholars. Citation sentences embed useful contents that signify the influence of cited authors on shared ideas, and express own opinion of citing authors to others' articles. The purpose of the study is to provide a new lens to analyze the topical relationship embedded in the citation sentences in an integrated manner. To this end, we extract citation sentences from full-text articles in the field of Oncology. In addition, we adopt Author-Journal-Topic (AJT) model to take both authors and journals into consideration of topic analysis. For the study, we collect the 6,360 full-text articles from PubMed Central and select the top 15 journals on Oncology. By applying AJT model, we identify what the major topics are shared among researchers in Oncology and which authors and journal lead the idea exchange in sub-disciplines of Oncology. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106
],
"paper_content_text": [
"Introduction As the size of data on the web continues to increase in an exponential manner, finding valuable meaning between data becomes of paramount importance in many research areas.",
"In the information science field, citations are challenging, pivotal materials to discover the relationship between academic documents because citations present the description of authors' ideas and the hidden relationship between authors and documents.",
"The earliest works focused mainly on classifying the citation behaviors and discovering the citation reasons with limited data such as the location of citation sentences and the number of references [1, 2] .",
"Since the mid-1990s, with the development of computer technology, citation content analysis was elaborated by applying data analysis techniques like text-mining or natural language processing.",
"Zhang et al.",
"[3] present citation analysis based on sematic and syntactic approaches.",
"Semantic-based citation analysis is performed by qualitative analysis to discover the citation motivation and citation classification.",
"On the other hand, syntactic-based citation analysis can be conducted by citation location and citation frequency, which reveals the hidden relation of authors by using meta-data of documents such as journal, venue of publication, affiliation of authors, etc.",
"Following their study, Ding et al.",
"[4] propose a theoretical methodology through content citation analysis.",
"However, these analyses are somewhat limited to the explicit context that primarily represents their own ideas and arguments.",
"The main goal of the paper is to discover the implicit topical relationships buried in citation sentences by utilizing the citation information from the author's perspective of sharing other authors' point of view.",
"Implicitness of the topical relationship is realized by using citation sentences as the input for the topic modeling technique.",
"In this study, a citation sentence indicates the sentence including citation expression consisting of year and author of the cited work.",
"In general, the citation sentence contains brief content of cited work and opinion that the author of citing work on the cited work.",
"We claim that citation sentences reveal interesting characteristics of scholarly communication such as influence, idea exchange, justification for citer's arguments, etc.",
"We assume that using citation sentences for topic analysis reveals aforementioned characteristics.",
"To explore such intellectual space created by citation sentences, we take both authors and journals into consideration of topic analysis.",
"To this end, we applied Author-Conference-Topic (ACT) model proposed by Tang et al.",
"[5] for our topic analysis in relation with both authors and journals, which is called Author-Journal-Topic (AJT) topic model.",
"ACT model is a probabilistic topic model for simultaneously extracting topics of papers, authors, and conferences.",
"There are a few studies to analyze content of citation sentences.",
"Most of previous studies focus on how the topic of document influences citation and vice versa [6, 7, 8] using Topic Modeling.",
"Kataria, Mitra, and Bhatia [8] adapt citation to Author-Topic model [9] with the assumption that the context surrounding the citation anchor could be used to get topical information about the cited authors.",
"These studies including Tang et al.",
"[10] 's ACT model are the examples of combining topic modelling methods and citation content analysis.",
"However, most previous studies used metadata of documents.",
"In this work, we focus on identifying the landscape of the oncology field from a perspective of citation.",
"By using citation sentences, our results can indicate which authors are actively cited and which journals lead a certain topic.",
"The rest of the paper is organized as follows: Section 2 describes the proposed approach.",
"Section 3 analyzes the topic modeling results.",
"Section 4 concludes the paper with the future work.",
"Methodology Main idea The basic assumption of the proposed approach is that citation sentences embed useful contents signifying the influence of cited authors on shared ideas of citing authors.",
"Citation sentences are also considered as an invisible intellectual place for idea exchanging since citations are effective means of supporting and expressing their own arguments by using other works.",
"In the similar vein, Di Marco and Mercer [11] claim that citation sentences play a major role in creating the relationship among relevant authors within the similar research fields.",
"With these assumptions, we are to explore the implicitness of topic relationships resided in citation sentences from the integrated perspective by incorporating the citing authors and journal titles into interpreting the topical relationships.",
"As shown in Figure 1 , we utilized various features including citing authors, citing sentences and journal titles for topic analysis.",
"Authors in Figure 1 mean the citing authors who write a paper and who cite other's work.",
"Citation sentences are the sentences written by the authors when they cite other's work in the paper, and journal titles are the journal names publishing the citing authors' paper.",
"By employing AJT model with these three parameters , we can discover which topics are the most salient ones referred to frequently by researchers and who are the leading authors sharing other authors' ideas in the research field and which journal leads such endeavor.",
"Fig.",
"1.",
"Three parameters for AJT model Data collection For this study, we compile the dataset on the field of Oncology from PubMed Central that provides the full-text in the biomedical field.",
"We select top 15 journals of Oncology by Thomson Reuter's JCR and journal's impact factor, and from these 15 journals, we are able to collect 6,360 full-text articles.",
"Figure 2 describes the workflow of our study.",
"As mentioned earlier, with the fulltext articles collected from PubMed Central, we extract the citation sentences.",
"Most citation sentences are kept in the following format: (author, year), (reference number) [reference number].",
"An example of such format is \"(<xref rid=\"bib00\" ref-type=\"bibr\">Author name, 2000</xref>)\".",
"We use the regular expression technique to parse and extract the citation sentences, when the tag <xref rid=>, </xref> appears on the sentences after parsing XML records with the Java-based SAX parser.",
"Method Fig.",
"2.",
"Workflow We also parse other metadata for AJT model such as the name of authors and journal titles.",
"The author tags, <surname> </surname> and <given-names></given-names> inside the <contrib-gourp></contrib-group>, denote the list of authors who wrote the paper.",
"For journal, we extract the titles when the journal tags, <journal-title> and </journal-title>, are included in the tag of <journal-meta> and </journal-meta>.",
"We also preprocess extracted sentences by removing both functional and general words and applying the Porter's stemming algorithm to improve the input for AJT Model.",
"AJT Model For our study, we apply ACT [10] model with several metadata such as citation sentences, journal titles and citing authors to develop AJT model.",
"Our AJT model utilizes journal titles and citation sentences instead of conference and abstract on documents.",
"The change of model is needed to analyze most influential topics in Oncology and to find leading authors who frequently mention the active topics and to detect the journals involved in such topics.",
"Figure 1 , Table 1) Like ACT model, AJT model assumes that each citing author is related to distribution over topics and each word in citation sentences is derived from a topic.",
"In the AJT model, the journal titles are related to each word.",
"To determine a word (ω_Si) in citation sentences (S), citing authors (x_Si) are consider for a word.",
"Each citing author is associated with a distributed topic.",
"A topic is generated from the citing author-topic distribution.",
"The words and journal titles are generated from a specific topic.",
"AJT model presents (1) 3 Results and Analyses For AJT model, we set the number of topics to 15 and finally select 8 topics as major topics.",
"Since we discovered that there are similar topics on our results, we calculated the similarity between 15 topics to select the most representative topics.",
"The topical similarities are measured by each word on topics and we calculated the similarities of two topics where each topic are represented in an array of a term vector.",
"Through this process, we chose 8 topics which have high topical similarities (over 0.5).",
"Each topic presents top 5 words from topic-word distribution, and 5 most related authors and journal titles are displayed along with each topic.",
"By performing several times on the pilot studies, we decided to choose top 5 words which are quite appropriate to describe each topics.",
"The results of AJT-based topic modeling is shown in Table 1 .",
"We label topic 1 \"breast cancer\" whose top words include breast, expression women, and growth.",
"Since the dataset is compiled with citation sentences, it implies that the topic \"breast cancer\" is a popular topic where researchers share and exchange ideas and facts related to breast cancer.",
"In relation to the topic \"breast cancer\", the active authors of breast cancer are Johnston Stephen RD, Colditz Graham A, and Sternlicht Mark D, and they share ideas with others on breast cancer from our results.",
"In terms of journals that provide a common place for idea sharing and communication, the journal \"Breast Cancer Research\" is the top journal of topic 1, and its impact factor is 5.49.",
"Authors such as Kurzrock Razelle, and Axelrod Haley in group 4 are the leading researchers sharing ideas on the topic \"targeted therapy.\"",
"The topic 4 is associated with the targeted therapy represented by words like mutations, treatments, therapy and disease.",
"The two most influential journals in topic 4 are \"Oncotarget\" and \"Journal of Thoracic Oncology\" whose impact factors are 6.36 and 5.28 respectively, which indicates that these two journals are the major journals encouraging authors to share ideas and collaborate with each other on cancer targeted therapy subject area.",
"Authors like Zitgel Laurence, Galluzzi Lorenzo, and Kroemer Guido in the author group 7 are the ones that actively share ideas about the topic \"Cancer Immunology.\"",
"Top concepts that are related to this topic are cell, immune, clinical and antitumor.",
"The top journal of the topic \"Cancer Immunology\" is Oncoinmmunology whose impact factor is 6.266.",
"Romagnani Paola and Salem Husein K in topic 8 \"Stem Cell\" are the authors that communicate and share ideas actively with each other in the given field, and the journal \"Stem Cells\" (impact factor: 6.523) is the leading journal.",
"We visualize topic keywords obtained from results of AJT-based topic model.",
"We construct the co-occurrence network and analyze which topic words play an important role in this domain.",
"Each node in the network represents a topic word, and an edge represents a co-occurrence frequency between keywords.",
"The size of nodes represents degree centrality and the color means network clusters obtained by using modularity algorithm.",
"This network consists of 100 nodes and 1,436 edges.",
"As shown in Figure 4 , each topic belongs to a specific community, but shares some important topic keywords.",
"Especially, the topic words positioned at the center is represented core-keywords in Oncology.",
"Figure 4 indicates that these words are the essential concepts of the Oncology domain.",
"Along with the results of AJT-based topic models, we can infer the major journals and authors develop their own research area based on these core-concepts.",
"Fig.",
"4.",
"Network of topic keywords The above results imply that the proposed approach identifies which topics are frequently shared, who facilitates to exchange ideas, and which journals provide a placeholder for it.",
"Identification of the triple relationship among authors, journals, and topics sheds new insight on understanding the well-discussed topics driven by the leading journals and authors that play a mediator role in the development of Oncology.",
"Conclusion One of the major research problems in bibliometrics is how to map out the intellectual structure of a research field.",
"The proposed approach tackles such research problem by utilizing citation sentences and AJT model.",
"By using citation sentences as the input for AJT model to find latent meaning, AJT model suggests a new way to detect leading authors and journals in sub-disciplines represented by discovered topics in a certain field.",
"Achieving this is not feasible by traditional frequency-based citation analysis.",
"One of the interesting observations is that the top-ranked journals in the discovered topics derived from AJT model are not ranked top in terms of JCR.",
"For example, the \"Oncotarget\" journal is the top-ranked journal in three topics in our analysis, but the ranking of the journal is 20 according to JCR.",
"Since we only report on preliminary results of our approach, we undertake in-depth analysis to investigate why this difference exists.",
"We also conduct various statistical tests on the results.",
"Based on the reported results in this paper, though, we claim that AJT can be used for discovering latent meaning associated citation sentences and the major players leading the field.",
"As a follow-up study, we will conduct a comparative study that compares the proposed approach with the general topic modeling technique such as LDA.",
"We also plan to investigate whether there is a different impact of using citation sentences and general meta-data such as abstract and title for topic analysis on facilitating idea sharing and scholarly communication.",
"In addition, we would like to consider the window size of citation sentences enriching citation context and to discover the authors' relationships among the neighboring citation sentences."
]
} | {
"paper_header_number": [
"1",
"2.1",
"2.2",
"2.4",
"4"
],
"paper_header_content": [
"Introduction",
"Main idea",
"Data collection",
"AJT Model",
"Conclusion"
]
} | GEM-SciDuet-train-7#paper-971#slide-5 | Data collection | PubMed Central: 6,360 full-text articles
15 journals of Oncology: by Thomson Reuters JCR & journals impact factor
Cancer Cell, Journal of the National Cancer Institute, Leukemia, Oncogene,
Annals of Oncology, Neuro-Oncology, Stem Cells, Oncotarget, OncoInnunology,
Molecular Oncology, Breast Cancer Research Journal of Thoracic Oncology,
Pigment Cell & Melanoma Resaerch, Clinical Epigenetics, Molecular Cancer | PubMed Central: 6,360 full-text articles
15 journals of Oncology: by Thomson Reuters JCR & journals impact factor
Cancer Cell, Journal of the National Cancer Institute, Leukemia, Oncogene,
Annals of Oncology, Neuro-Oncology, Stem Cells, Oncotarget, OncoInnunology,
Molecular Oncology, Breast Cancer Research Journal of Thoracic Oncology,
Pigment Cell & Melanoma Resaerch, Clinical Epigenetics, Molecular Cancer | [] |
GEM-SciDuet-train-7#paper-971#slide-6 | 971 | Exploring the leading authors and journals in major topics by citation sentences and topic modeling | Citation plays an important role in understanding the knowledge sharing among scholars. Citation sentences embed useful contents that signify the influence of cited authors on shared ideas, and express own opinion of citing authors to others' articles. The purpose of the study is to provide a new lens to analyze the topical relationship embedded in the citation sentences in an integrated manner. To this end, we extract citation sentences from full-text articles in the field of Oncology. In addition, we adopt Author-Journal-Topic (AJT) model to take both authors and journals into consideration of topic analysis. For the study, we collect the 6,360 full-text articles from PubMed Central and select the top 15 journals on Oncology. By applying AJT model, we identify what the major topics are shared among researchers in Oncology and which authors and journal lead the idea exchange in sub-disciplines of Oncology. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106
],
"paper_content_text": [
"Introduction As the size of data on the web continues to increase in an exponential manner, finding valuable meaning between data becomes of paramount importance in many research areas.",
"In the information science field, citations are challenging, pivotal materials to discover the relationship between academic documents because citations present the description of authors' ideas and the hidden relationship between authors and documents.",
"The earliest works focused mainly on classifying the citation behaviors and discovering the citation reasons with limited data such as the location of citation sentences and the number of references [1, 2] .",
"Since the mid-1990s, with the development of computer technology, citation content analysis was elaborated by applying data analysis techniques like text-mining or natural language processing.",
"Zhang et al.",
"[3] present citation analysis based on sematic and syntactic approaches.",
"Semantic-based citation analysis is performed by qualitative analysis to discover the citation motivation and citation classification.",
"On the other hand, syntactic-based citation analysis can be conducted by citation location and citation frequency, which reveals the hidden relation of authors by using meta-data of documents such as journal, venue of publication, affiliation of authors, etc.",
"Following their study, Ding et al.",
"[4] propose a theoretical methodology through content citation analysis.",
"However, these analyses are somewhat limited to the explicit context that primarily represents their own ideas and arguments.",
"The main goal of the paper is to discover the implicit topical relationships buried in citation sentences by utilizing the citation information from the author's perspective of sharing other authors' point of view.",
"Implicitness of the topical relationship is realized by using citation sentences as the input for the topic modeling technique.",
"In this study, a citation sentence indicates the sentence including citation expression consisting of year and author of the cited work.",
"In general, the citation sentence contains brief content of cited work and opinion that the author of citing work on the cited work.",
"We claim that citation sentences reveal interesting characteristics of scholarly communication such as influence, idea exchange, justification for citer's arguments, etc.",
"We assume that using citation sentences for topic analysis reveals aforementioned characteristics.",
"To explore such intellectual space created by citation sentences, we take both authors and journals into consideration of topic analysis.",
"To this end, we applied Author-Conference-Topic (ACT) model proposed by Tang et al.",
"[5] for our topic analysis in relation with both authors and journals, which is called Author-Journal-Topic (AJT) topic model.",
"ACT model is a probabilistic topic model for simultaneously extracting topics of papers, authors, and conferences.",
"There are a few studies to analyze content of citation sentences.",
"Most of previous studies focus on how the topic of document influences citation and vice versa [6, 7, 8] using Topic Modeling.",
"Kataria, Mitra, and Bhatia [8] adapt citation to Author-Topic model [9] with the assumption that the context surrounding the citation anchor could be used to get topical information about the cited authors.",
"These studies including Tang et al.",
"[10] 's ACT model are the examples of combining topic modelling methods and citation content analysis.",
"However, most previous studies used metadata of documents.",
"In this work, we focus on identifying the landscape of the oncology field from a perspective of citation.",
"By using citation sentences, our results can indicate which authors are actively cited and which journals lead a certain topic.",
"The rest of the paper is organized as follows: Section 2 describes the proposed approach.",
"Section 3 analyzes the topic modeling results.",
"Section 4 concludes the paper with the future work.",
"Methodology Main idea The basic assumption of the proposed approach is that citation sentences embed useful contents signifying the influence of cited authors on shared ideas of citing authors.",
"Citation sentences are also considered as an invisible intellectual place for idea exchanging since citations are effective means of supporting and expressing their own arguments by using other works.",
"In the similar vein, Di Marco and Mercer [11] claim that citation sentences play a major role in creating the relationship among relevant authors within the similar research fields.",
"With these assumptions, we are to explore the implicitness of topic relationships resided in citation sentences from the integrated perspective by incorporating the citing authors and journal titles into interpreting the topical relationships.",
"As shown in Figure 1 , we utilized various features including citing authors, citing sentences and journal titles for topic analysis.",
"Authors in Figure 1 mean the citing authors who write a paper and who cite other's work.",
"Citation sentences are the sentences written by the authors when they cite other's work in the paper, and journal titles are the journal names publishing the citing authors' paper.",
"By employing AJT model with these three parameters , we can discover which topics are the most salient ones referred to frequently by researchers and who are the leading authors sharing other authors' ideas in the research field and which journal leads such endeavor.",
"Fig.",
"1.",
"Three parameters for AJT model Data collection For this study, we compile the dataset on the field of Oncology from PubMed Central that provides the full-text in the biomedical field.",
"We select top 15 journals of Oncology by Thomson Reuter's JCR and journal's impact factor, and from these 15 journals, we are able to collect 6,360 full-text articles.",
"Figure 2 describes the workflow of our study.",
"As mentioned earlier, with the fulltext articles collected from PubMed Central, we extract the citation sentences.",
"Most citation sentences are kept in the following format: (author, year), (reference number) [reference number].",
"An example of such format is \"(<xref rid=\"bib00\" ref-type=\"bibr\">Author name, 2000</xref>)\".",
"We use the regular expression technique to parse and extract the citation sentences, when the tag <xref rid=>, </xref> appears on the sentences after parsing XML records with the Java-based SAX parser.",
"Method Fig.",
"2.",
"Workflow We also parse other metadata for AJT model such as the name of authors and journal titles.",
"The author tags, <surname> </surname> and <given-names></given-names> inside the <contrib-gourp></contrib-group>, denote the list of authors who wrote the paper.",
"For journal, we extract the titles when the journal tags, <journal-title> and </journal-title>, are included in the tag of <journal-meta> and </journal-meta>.",
"We also preprocess extracted sentences by removing both functional and general words and applying the Porter's stemming algorithm to improve the input for AJT Model.",
"AJT Model For our study, we apply ACT [10] model with several metadata such as citation sentences, journal titles and citing authors to develop AJT model.",
"Our AJT model utilizes journal titles and citation sentences instead of conference and abstract on documents.",
"The change of model is needed to analyze most influential topics in Oncology and to find leading authors who frequently mention the active topics and to detect the journals involved in such topics.",
"Figure 1 , Table 1) Like ACT model, AJT model assumes that each citing author is related to distribution over topics and each word in citation sentences is derived from a topic.",
"In the AJT model, the journal titles are related to each word.",
"To determine a word (ω_Si) in citation sentences (S), citing authors (x_Si) are consider for a word.",
"Each citing author is associated with a distributed topic.",
"A topic is generated from the citing author-topic distribution.",
"The words and journal titles are generated from a specific topic.",
"AJT model presents (1) 3 Results and Analyses For AJT model, we set the number of topics to 15 and finally select 8 topics as major topics.",
"Since we discovered that there are similar topics on our results, we calculated the similarity between 15 topics to select the most representative topics.",
"The topical similarities are measured by each word on topics and we calculated the similarities of two topics where each topic are represented in an array of a term vector.",
"Through this process, we chose 8 topics which have high topical similarities (over 0.5).",
"Each topic presents top 5 words from topic-word distribution, and 5 most related authors and journal titles are displayed along with each topic.",
"By performing several times on the pilot studies, we decided to choose top 5 words which are quite appropriate to describe each topics.",
"The results of AJT-based topic modeling is shown in Table 1 .",
"We label topic 1 \"breast cancer\" whose top words include breast, expression women, and growth.",
"Since the dataset is compiled with citation sentences, it implies that the topic \"breast cancer\" is a popular topic where researchers share and exchange ideas and facts related to breast cancer.",
"In relation to the topic \"breast cancer\", the active authors of breast cancer are Johnston Stephen RD, Colditz Graham A, and Sternlicht Mark D, and they share ideas with others on breast cancer from our results.",
"In terms of journals that provide a common place for idea sharing and communication, the journal \"Breast Cancer Research\" is the top journal of topic 1, and its impact factor is 5.49.",
"Authors such as Kurzrock Razelle, and Axelrod Haley in group 4 are the leading researchers sharing ideas on the topic \"targeted therapy.\"",
"The topic 4 is associated with the targeted therapy represented by words like mutations, treatments, therapy and disease.",
"The two most influential journals in topic 4 are \"Oncotarget\" and \"Journal of Thoracic Oncology\" whose impact factors are 6.36 and 5.28 respectively, which indicates that these two journals are the major journals encouraging authors to share ideas and collaborate with each other on cancer targeted therapy subject area.",
"Authors like Zitgel Laurence, Galluzzi Lorenzo, and Kroemer Guido in the author group 7 are the ones that actively share ideas about the topic \"Cancer Immunology.\"",
"Top concepts that are related to this topic are cell, immune, clinical and antitumor.",
"The top journal of the topic \"Cancer Immunology\" is Oncoinmmunology whose impact factor is 6.266.",
"Romagnani Paola and Salem Husein K in topic 8 \"Stem Cell\" are the authors that communicate and share ideas actively with each other in the given field, and the journal \"Stem Cells\" (impact factor: 6.523) is the leading journal.",
"We visualize topic keywords obtained from results of AJT-based topic model.",
"We construct the co-occurrence network and analyze which topic words play an important role in this domain.",
"Each node in the network represents a topic word, and an edge represents a co-occurrence frequency between keywords.",
"The size of nodes represents degree centrality and the color means network clusters obtained by using modularity algorithm.",
"This network consists of 100 nodes and 1,436 edges.",
"As shown in Figure 4 , each topic belongs to a specific community, but shares some important topic keywords.",
"Especially, the topic words positioned at the center is represented core-keywords in Oncology.",
"Figure 4 indicates that these words are the essential concepts of the Oncology domain.",
"Along with the results of AJT-based topic models, we can infer the major journals and authors develop their own research area based on these core-concepts.",
"Fig.",
"4.",
"Network of topic keywords The above results imply that the proposed approach identifies which topics are frequently shared, who facilitates to exchange ideas, and which journals provide a placeholder for it.",
"Identification of the triple relationship among authors, journals, and topics sheds new insight on understanding the well-discussed topics driven by the leading journals and authors that play a mediator role in the development of Oncology.",
"Conclusion One of the major research problems in bibliometrics is how to map out the intellectual structure of a research field.",
"The proposed approach tackles such research problem by utilizing citation sentences and AJT model.",
"By using citation sentences as the input for AJT model to find latent meaning, AJT model suggests a new way to detect leading authors and journals in sub-disciplines represented by discovered topics in a certain field.",
"Achieving this is not feasible by traditional frequency-based citation analysis.",
"One of the interesting observations is that the top-ranked journals in the discovered topics derived from AJT model are not ranked top in terms of JCR.",
"For example, the \"Oncotarget\" journal is the top-ranked journal in three topics in our analysis, but the ranking of the journal is 20 according to JCR.",
"Since we only report on preliminary results of our approach, we undertake in-depth analysis to investigate why this difference exists.",
"We also conduct various statistical tests on the results.",
"Based on the reported results in this paper, though, we claim that AJT can be used for discovering latent meaning associated citation sentences and the major players leading the field.",
"As a follow-up study, we will conduct a comparative study that compares the proposed approach with the general topic modeling technique such as LDA.",
"We also plan to investigate whether there is a different impact of using citation sentences and general meta-data such as abstract and title for topic analysis on facilitating idea sharing and scholarly communication.",
"In addition, we would like to consider the window size of citation sentences enriching citation context and to discover the authors' relationships among the neighboring citation sentences."
]
} | {
"paper_header_number": [
"1",
"2.1",
"2.2",
"2.4",
"4"
],
"paper_header_content": [
"Introduction",
"Main idea",
"Data collection",
"AJT Model",
"Conclusion"
]
} | GEM-SciDuet-train-7#paper-971#slide-6 | Research Flow | 1) Citation Data Extraction | 1) Citation Data Extraction | [] |
GEM-SciDuet-train-7#paper-971#slide-7 | 971 | Exploring the leading authors and journals in major topics by citation sentences and topic modeling | Citation plays an important role in understanding the knowledge sharing among scholars. Citation sentences embed useful contents that signify the influence of cited authors on shared ideas, and express own opinion of citing authors to others' articles. The purpose of the study is to provide a new lens to analyze the topical relationship embedded in the citation sentences in an integrated manner. To this end, we extract citation sentences from full-text articles in the field of Oncology. In addition, we adopt Author-Journal-Topic (AJT) model to take both authors and journals into consideration of topic analysis. For the study, we collect the 6,360 full-text articles from PubMed Central and select the top 15 journals on Oncology. By applying AJT model, we identify what the major topics are shared among researchers in Oncology and which authors and journal lead the idea exchange in sub-disciplines of Oncology. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106
],
"paper_content_text": [
"Introduction As the size of data on the web continues to increase in an exponential manner, finding valuable meaning between data becomes of paramount importance in many research areas.",
"In the information science field, citations are challenging, pivotal materials to discover the relationship between academic documents because citations present the description of authors' ideas and the hidden relationship between authors and documents.",
"The earliest works focused mainly on classifying the citation behaviors and discovering the citation reasons with limited data such as the location of citation sentences and the number of references [1, 2] .",
"Since the mid-1990s, with the development of computer technology, citation content analysis was elaborated by applying data analysis techniques like text-mining or natural language processing.",
"Zhang et al.",
"[3] present citation analysis based on sematic and syntactic approaches.",
"Semantic-based citation analysis is performed by qualitative analysis to discover the citation motivation and citation classification.",
"On the other hand, syntactic-based citation analysis can be conducted by citation location and citation frequency, which reveals the hidden relation of authors by using meta-data of documents such as journal, venue of publication, affiliation of authors, etc.",
"Following their study, Ding et al.",
"[4] propose a theoretical methodology through content citation analysis.",
"However, these analyses are somewhat limited to the explicit context that primarily represents their own ideas and arguments.",
"The main goal of the paper is to discover the implicit topical relationships buried in citation sentences by utilizing the citation information from the author's perspective of sharing other authors' point of view.",
"Implicitness of the topical relationship is realized by using citation sentences as the input for the topic modeling technique.",
"In this study, a citation sentence indicates the sentence including citation expression consisting of year and author of the cited work.",
"In general, the citation sentence contains brief content of cited work and opinion that the author of citing work on the cited work.",
"We claim that citation sentences reveal interesting characteristics of scholarly communication such as influence, idea exchange, justification for citer's arguments, etc.",
"We assume that using citation sentences for topic analysis reveals aforementioned characteristics.",
"To explore such intellectual space created by citation sentences, we take both authors and journals into consideration of topic analysis.",
"To this end, we applied Author-Conference-Topic (ACT) model proposed by Tang et al.",
"[5] for our topic analysis in relation with both authors and journals, which is called Author-Journal-Topic (AJT) topic model.",
"ACT model is a probabilistic topic model for simultaneously extracting topics of papers, authors, and conferences.",
"There are a few studies to analyze content of citation sentences.",
"Most of previous studies focus on how the topic of document influences citation and vice versa [6, 7, 8] using Topic Modeling.",
"Kataria, Mitra, and Bhatia [8] adapt citation to Author-Topic model [9] with the assumption that the context surrounding the citation anchor could be used to get topical information about the cited authors.",
"These studies including Tang et al.",
"[10] 's ACT model are the examples of combining topic modelling methods and citation content analysis.",
"However, most previous studies used metadata of documents.",
"In this work, we focus on identifying the landscape of the oncology field from a perspective of citation.",
"By using citation sentences, our results can indicate which authors are actively cited and which journals lead a certain topic.",
"The rest of the paper is organized as follows: Section 2 describes the proposed approach.",
"Section 3 analyzes the topic modeling results.",
"Section 4 concludes the paper with the future work.",
"Methodology Main idea The basic assumption of the proposed approach is that citation sentences embed useful contents signifying the influence of cited authors on shared ideas of citing authors.",
"Citation sentences are also considered as an invisible intellectual place for idea exchanging since citations are effective means of supporting and expressing their own arguments by using other works.",
"In the similar vein, Di Marco and Mercer [11] claim that citation sentences play a major role in creating the relationship among relevant authors within the similar research fields.",
"With these assumptions, we are to explore the implicitness of topic relationships resided in citation sentences from the integrated perspective by incorporating the citing authors and journal titles into interpreting the topical relationships.",
"As shown in Figure 1 , we utilized various features including citing authors, citing sentences and journal titles for topic analysis.",
"Authors in Figure 1 mean the citing authors who write a paper and who cite other's work.",
"Citation sentences are the sentences written by the authors when they cite other's work in the paper, and journal titles are the journal names publishing the citing authors' paper.",
"By employing AJT model with these three parameters , we can discover which topics are the most salient ones referred to frequently by researchers and who are the leading authors sharing other authors' ideas in the research field and which journal leads such endeavor.",
"Fig.",
"1.",
"Three parameters for AJT model Data collection For this study, we compile the dataset on the field of Oncology from PubMed Central that provides the full-text in the biomedical field.",
"We select top 15 journals of Oncology by Thomson Reuter's JCR and journal's impact factor, and from these 15 journals, we are able to collect 6,360 full-text articles.",
"Figure 2 describes the workflow of our study.",
"As mentioned earlier, with the fulltext articles collected from PubMed Central, we extract the citation sentences.",
"Most citation sentences are kept in the following format: (author, year), (reference number) [reference number].",
"An example of such format is \"(<xref rid=\"bib00\" ref-type=\"bibr\">Author name, 2000</xref>)\".",
"We use the regular expression technique to parse and extract the citation sentences, when the tag <xref rid=>, </xref> appears on the sentences after parsing XML records with the Java-based SAX parser.",
"Method Fig.",
"2.",
"Workflow We also parse other metadata for AJT model such as the name of authors and journal titles.",
"The author tags, <surname> </surname> and <given-names></given-names> inside the <contrib-gourp></contrib-group>, denote the list of authors who wrote the paper.",
"For journal, we extract the titles when the journal tags, <journal-title> and </journal-title>, are included in the tag of <journal-meta> and </journal-meta>.",
"We also preprocess extracted sentences by removing both functional and general words and applying the Porter's stemming algorithm to improve the input for AJT Model.",
"AJT Model For our study, we apply ACT [10] model with several metadata such as citation sentences, journal titles and citing authors to develop AJT model.",
"Our AJT model utilizes journal titles and citation sentences instead of conference and abstract on documents.",
"The change of model is needed to analyze most influential topics in Oncology and to find leading authors who frequently mention the active topics and to detect the journals involved in such topics.",
"Figure 1 , Table 1) Like ACT model, AJT model assumes that each citing author is related to distribution over topics and each word in citation sentences is derived from a topic.",
"In the AJT model, the journal titles are related to each word.",
"To determine a word (ω_Si) in citation sentences (S), citing authors (x_Si) are consider for a word.",
"Each citing author is associated with a distributed topic.",
"A topic is generated from the citing author-topic distribution.",
"The words and journal titles are generated from a specific topic.",
"AJT model presents (1) 3 Results and Analyses For AJT model, we set the number of topics to 15 and finally select 8 topics as major topics.",
"Since we discovered that there are similar topics on our results, we calculated the similarity between 15 topics to select the most representative topics.",
"The topical similarities are measured by each word on topics and we calculated the similarities of two topics where each topic are represented in an array of a term vector.",
"Through this process, we chose 8 topics which have high topical similarities (over 0.5).",
"Each topic presents top 5 words from topic-word distribution, and 5 most related authors and journal titles are displayed along with each topic.",
"By performing several times on the pilot studies, we decided to choose top 5 words which are quite appropriate to describe each topics.",
"The results of AJT-based topic modeling is shown in Table 1 .",
"We label topic 1 \"breast cancer\" whose top words include breast, expression women, and growth.",
"Since the dataset is compiled with citation sentences, it implies that the topic \"breast cancer\" is a popular topic where researchers share and exchange ideas and facts related to breast cancer.",
"In relation to the topic \"breast cancer\", the active authors of breast cancer are Johnston Stephen RD, Colditz Graham A, and Sternlicht Mark D, and they share ideas with others on breast cancer from our results.",
"In terms of journals that provide a common place for idea sharing and communication, the journal \"Breast Cancer Research\" is the top journal of topic 1, and its impact factor is 5.49.",
"Authors such as Kurzrock Razelle, and Axelrod Haley in group 4 are the leading researchers sharing ideas on the topic \"targeted therapy.\"",
"The topic 4 is associated with the targeted therapy represented by words like mutations, treatments, therapy and disease.",
"The two most influential journals in topic 4 are \"Oncotarget\" and \"Journal of Thoracic Oncology\" whose impact factors are 6.36 and 5.28 respectively, which indicates that these two journals are the major journals encouraging authors to share ideas and collaborate with each other on cancer targeted therapy subject area.",
"Authors like Zitgel Laurence, Galluzzi Lorenzo, and Kroemer Guido in the author group 7 are the ones that actively share ideas about the topic \"Cancer Immunology.\"",
"Top concepts that are related to this topic are cell, immune, clinical and antitumor.",
"The top journal of the topic \"Cancer Immunology\" is Oncoinmmunology whose impact factor is 6.266.",
"Romagnani Paola and Salem Husein K in topic 8 \"Stem Cell\" are the authors that communicate and share ideas actively with each other in the given field, and the journal \"Stem Cells\" (impact factor: 6.523) is the leading journal.",
"We visualize topic keywords obtained from results of AJT-based topic model.",
"We construct the co-occurrence network and analyze which topic words play an important role in this domain.",
"Each node in the network represents a topic word, and an edge represents a co-occurrence frequency between keywords.",
"The size of nodes represents degree centrality and the color means network clusters obtained by using modularity algorithm.",
"This network consists of 100 nodes and 1,436 edges.",
"As shown in Figure 4 , each topic belongs to a specific community, but shares some important topic keywords.",
"Especially, the topic words positioned at the center is represented core-keywords in Oncology.",
"Figure 4 indicates that these words are the essential concepts of the Oncology domain.",
"Along with the results of AJT-based topic models, we can infer the major journals and authors develop their own research area based on these core-concepts.",
"Fig.",
"4.",
"Network of topic keywords The above results imply that the proposed approach identifies which topics are frequently shared, who facilitates to exchange ideas, and which journals provide a placeholder for it.",
"Identification of the triple relationship among authors, journals, and topics sheds new insight on understanding the well-discussed topics driven by the leading journals and authors that play a mediator role in the development of Oncology.",
"Conclusion One of the major research problems in bibliometrics is how to map out the intellectual structure of a research field.",
"The proposed approach tackles such research problem by utilizing citation sentences and AJT model.",
"By using citation sentences as the input for AJT model to find latent meaning, AJT model suggests a new way to detect leading authors and journals in sub-disciplines represented by discovered topics in a certain field.",
"Achieving this is not feasible by traditional frequency-based citation analysis.",
"One of the interesting observations is that the top-ranked journals in the discovered topics derived from AJT model are not ranked top in terms of JCR.",
"For example, the \"Oncotarget\" journal is the top-ranked journal in three topics in our analysis, but the ranking of the journal is 20 according to JCR.",
"Since we only report on preliminary results of our approach, we undertake in-depth analysis to investigate why this difference exists.",
"We also conduct various statistical tests on the results.",
"Based on the reported results in this paper, though, we claim that AJT can be used for discovering latent meaning associated citation sentences and the major players leading the field.",
"As a follow-up study, we will conduct a comparative study that compares the proposed approach with the general topic modeling technique such as LDA.",
"We also plan to investigate whether there is a different impact of using citation sentences and general meta-data such as abstract and title for topic analysis on facilitating idea sharing and scholarly communication.",
"In addition, we would like to consider the window size of citation sentences enriching citation context and to discover the authors' relationships among the neighboring citation sentences."
]
} | {
"paper_header_number": [
"1",
"2.1",
"2.2",
"2.4",
"4"
],
"paper_header_content": [
"Introduction",
"Main idea",
"Data collection",
"AJT Model",
"Conclusion"
]
} | GEM-SciDuet-train-7#paper-971#slide-7 | Results 8 Topics | Labeled by 3 Experts
Author Group 1 Author Group 2 Author Group 3 Author Group 4
Journal Group 1 Journal Group 2 Journal Group 3 Journal Group 4
Research Annals of Oncology
Pigment Cell & Melanoma Research
Journal of Thoracic Oncology | Labeled by 3 Experts
Author Group 1 Author Group 2 Author Group 3 Author Group 4
Journal Group 1 Journal Group 2 Journal Group 3 Journal Group 4
Research Annals of Oncology
Pigment Cell & Melanoma Research
Journal of Thoracic Oncology | [] |
GEM-SciDuet-train-7#paper-971#slide-8 | 971 | Exploring the leading authors and journals in major topics by citation sentences and topic modeling | Citation plays an important role in understanding the knowledge sharing among scholars. Citation sentences embed useful contents that signify the influence of cited authors on shared ideas, and express own opinion of citing authors to others' articles. The purpose of the study is to provide a new lens to analyze the topical relationship embedded in the citation sentences in an integrated manner. To this end, we extract citation sentences from full-text articles in the field of Oncology. In addition, we adopt Author-Journal-Topic (AJT) model to take both authors and journals into consideration of topic analysis. For the study, we collect the 6,360 full-text articles from PubMed Central and select the top 15 journals on Oncology. By applying AJT model, we identify what the major topics are shared among researchers in Oncology and which authors and journal lead the idea exchange in sub-disciplines of Oncology. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106
],
"paper_content_text": [
"Introduction As the size of data on the web continues to increase in an exponential manner, finding valuable meaning between data becomes of paramount importance in many research areas.",
"In the information science field, citations are challenging, pivotal materials to discover the relationship between academic documents because citations present the description of authors' ideas and the hidden relationship between authors and documents.",
"The earliest works focused mainly on classifying the citation behaviors and discovering the citation reasons with limited data such as the location of citation sentences and the number of references [1, 2] .",
"Since the mid-1990s, with the development of computer technology, citation content analysis was elaborated by applying data analysis techniques like text-mining or natural language processing.",
"Zhang et al.",
"[3] present citation analysis based on sematic and syntactic approaches.",
"Semantic-based citation analysis is performed by qualitative analysis to discover the citation motivation and citation classification.",
"On the other hand, syntactic-based citation analysis can be conducted by citation location and citation frequency, which reveals the hidden relation of authors by using meta-data of documents such as journal, venue of publication, affiliation of authors, etc.",
"Following their study, Ding et al.",
"[4] propose a theoretical methodology through content citation analysis.",
"However, these analyses are somewhat limited to the explicit context that primarily represents their own ideas and arguments.",
"The main goal of the paper is to discover the implicit topical relationships buried in citation sentences by utilizing the citation information from the author's perspective of sharing other authors' point of view.",
"Implicitness of the topical relationship is realized by using citation sentences as the input for the topic modeling technique.",
"In this study, a citation sentence indicates the sentence including citation expression consisting of year and author of the cited work.",
"In general, the citation sentence contains brief content of cited work and opinion that the author of citing work on the cited work.",
"We claim that citation sentences reveal interesting characteristics of scholarly communication such as influence, idea exchange, justification for citer's arguments, etc.",
"We assume that using citation sentences for topic analysis reveals aforementioned characteristics.",
"To explore such intellectual space created by citation sentences, we take both authors and journals into consideration of topic analysis.",
"To this end, we applied Author-Conference-Topic (ACT) model proposed by Tang et al.",
"[5] for our topic analysis in relation with both authors and journals, which is called Author-Journal-Topic (AJT) topic model.",
"ACT model is a probabilistic topic model for simultaneously extracting topics of papers, authors, and conferences.",
"There are a few studies to analyze content of citation sentences.",
"Most of previous studies focus on how the topic of document influences citation and vice versa [6, 7, 8] using Topic Modeling.",
"Kataria, Mitra, and Bhatia [8] adapt citation to Author-Topic model [9] with the assumption that the context surrounding the citation anchor could be used to get topical information about the cited authors.",
"These studies including Tang et al.",
"[10] 's ACT model are the examples of combining topic modelling methods and citation content analysis.",
"However, most previous studies used metadata of documents.",
"In this work, we focus on identifying the landscape of the oncology field from a perspective of citation.",
"By using citation sentences, our results can indicate which authors are actively cited and which journals lead a certain topic.",
"The rest of the paper is organized as follows: Section 2 describes the proposed approach.",
"Section 3 analyzes the topic modeling results.",
"Section 4 concludes the paper with the future work.",
"Methodology Main idea The basic assumption of the proposed approach is that citation sentences embed useful contents signifying the influence of cited authors on shared ideas of citing authors.",
"Citation sentences are also considered as an invisible intellectual place for idea exchanging since citations are effective means of supporting and expressing their own arguments by using other works.",
"In the similar vein, Di Marco and Mercer [11] claim that citation sentences play a major role in creating the relationship among relevant authors within the similar research fields.",
"With these assumptions, we are to explore the implicitness of topic relationships resided in citation sentences from the integrated perspective by incorporating the citing authors and journal titles into interpreting the topical relationships.",
"As shown in Figure 1 , we utilized various features including citing authors, citing sentences and journal titles for topic analysis.",
"Authors in Figure 1 mean the citing authors who write a paper and who cite other's work.",
"Citation sentences are the sentences written by the authors when they cite other's work in the paper, and journal titles are the journal names publishing the citing authors' paper.",
"By employing AJT model with these three parameters , we can discover which topics are the most salient ones referred to frequently by researchers and who are the leading authors sharing other authors' ideas in the research field and which journal leads such endeavor.",
"Fig.",
"1.",
"Three parameters for AJT model Data collection For this study, we compile the dataset on the field of Oncology from PubMed Central that provides the full-text in the biomedical field.",
"We select top 15 journals of Oncology by Thomson Reuter's JCR and journal's impact factor, and from these 15 journals, we are able to collect 6,360 full-text articles.",
"Figure 2 describes the workflow of our study.",
"As mentioned earlier, with the fulltext articles collected from PubMed Central, we extract the citation sentences.",
"Most citation sentences are kept in the following format: (author, year), (reference number) [reference number].",
"An example of such format is \"(<xref rid=\"bib00\" ref-type=\"bibr\">Author name, 2000</xref>)\".",
"We use the regular expression technique to parse and extract the citation sentences, when the tag <xref rid=>, </xref> appears on the sentences after parsing XML records with the Java-based SAX parser.",
"Method Fig.",
"2.",
"Workflow We also parse other metadata for AJT model such as the name of authors and journal titles.",
"The author tags, <surname> </surname> and <given-names></given-names> inside the <contrib-gourp></contrib-group>, denote the list of authors who wrote the paper.",
"For journal, we extract the titles when the journal tags, <journal-title> and </journal-title>, are included in the tag of <journal-meta> and </journal-meta>.",
"We also preprocess extracted sentences by removing both functional and general words and applying the Porter's stemming algorithm to improve the input for AJT Model.",
"AJT Model For our study, we apply ACT [10] model with several metadata such as citation sentences, journal titles and citing authors to develop AJT model.",
"Our AJT model utilizes journal titles and citation sentences instead of conference and abstract on documents.",
"The change of model is needed to analyze most influential topics in Oncology and to find leading authors who frequently mention the active topics and to detect the journals involved in such topics.",
"Figure 1 , Table 1) Like ACT model, AJT model assumes that each citing author is related to distribution over topics and each word in citation sentences is derived from a topic.",
"In the AJT model, the journal titles are related to each word.",
"To determine a word (ω_Si) in citation sentences (S), citing authors (x_Si) are consider for a word.",
"Each citing author is associated with a distributed topic.",
"A topic is generated from the citing author-topic distribution.",
"The words and journal titles are generated from a specific topic.",
"AJT model presents (1) 3 Results and Analyses For AJT model, we set the number of topics to 15 and finally select 8 topics as major topics.",
"Since we discovered that there are similar topics on our results, we calculated the similarity between 15 topics to select the most representative topics.",
"The topical similarities are measured by each word on topics and we calculated the similarities of two topics where each topic are represented in an array of a term vector.",
"Through this process, we chose 8 topics which have high topical similarities (over 0.5).",
"Each topic presents top 5 words from topic-word distribution, and 5 most related authors and journal titles are displayed along with each topic.",
"By performing several times on the pilot studies, we decided to choose top 5 words which are quite appropriate to describe each topics.",
"The results of AJT-based topic modeling is shown in Table 1 .",
"We label topic 1 \"breast cancer\" whose top words include breast, expression women, and growth.",
"Since the dataset is compiled with citation sentences, it implies that the topic \"breast cancer\" is a popular topic where researchers share and exchange ideas and facts related to breast cancer.",
"In relation to the topic \"breast cancer\", the active authors of breast cancer are Johnston Stephen RD, Colditz Graham A, and Sternlicht Mark D, and they share ideas with others on breast cancer from our results.",
"In terms of journals that provide a common place for idea sharing and communication, the journal \"Breast Cancer Research\" is the top journal of topic 1, and its impact factor is 5.49.",
"Authors such as Kurzrock Razelle, and Axelrod Haley in group 4 are the leading researchers sharing ideas on the topic \"targeted therapy.\"",
"The topic 4 is associated with the targeted therapy represented by words like mutations, treatments, therapy and disease.",
"The two most influential journals in topic 4 are \"Oncotarget\" and \"Journal of Thoracic Oncology\" whose impact factors are 6.36 and 5.28 respectively, which indicates that these two journals are the major journals encouraging authors to share ideas and collaborate with each other on cancer targeted therapy subject area.",
"Authors like Zitgel Laurence, Galluzzi Lorenzo, and Kroemer Guido in the author group 7 are the ones that actively share ideas about the topic \"Cancer Immunology.\"",
"Top concepts that are related to this topic are cell, immune, clinical and antitumor.",
"The top journal of the topic \"Cancer Immunology\" is Oncoinmmunology whose impact factor is 6.266.",
"Romagnani Paola and Salem Husein K in topic 8 \"Stem Cell\" are the authors that communicate and share ideas actively with each other in the given field, and the journal \"Stem Cells\" (impact factor: 6.523) is the leading journal.",
"We visualize topic keywords obtained from results of AJT-based topic model.",
"We construct the co-occurrence network and analyze which topic words play an important role in this domain.",
"Each node in the network represents a topic word, and an edge represents a co-occurrence frequency between keywords.",
"The size of nodes represents degree centrality and the color means network clusters obtained by using modularity algorithm.",
"This network consists of 100 nodes and 1,436 edges.",
"As shown in Figure 4 , each topic belongs to a specific community, but shares some important topic keywords.",
"Especially, the topic words positioned at the center is represented core-keywords in Oncology.",
"Figure 4 indicates that these words are the essential concepts of the Oncology domain.",
"Along with the results of AJT-based topic models, we can infer the major journals and authors develop their own research area based on these core-concepts.",
"Fig.",
"4.",
"Network of topic keywords The above results imply that the proposed approach identifies which topics are frequently shared, who facilitates to exchange ideas, and which journals provide a placeholder for it.",
"Identification of the triple relationship among authors, journals, and topics sheds new insight on understanding the well-discussed topics driven by the leading journals and authors that play a mediator role in the development of Oncology.",
"Conclusion One of the major research problems in bibliometrics is how to map out the intellectual structure of a research field.",
"The proposed approach tackles such research problem by utilizing citation sentences and AJT model.",
"By using citation sentences as the input for AJT model to find latent meaning, AJT model suggests a new way to detect leading authors and journals in sub-disciplines represented by discovered topics in a certain field.",
"Achieving this is not feasible by traditional frequency-based citation analysis.",
"One of the interesting observations is that the top-ranked journals in the discovered topics derived from AJT model are not ranked top in terms of JCR.",
"For example, the \"Oncotarget\" journal is the top-ranked journal in three topics in our analysis, but the ranking of the journal is 20 according to JCR.",
"Since we only report on preliminary results of our approach, we undertake in-depth analysis to investigate why this difference exists.",
"We also conduct various statistical tests on the results.",
"Based on the reported results in this paper, though, we claim that AJT can be used for discovering latent meaning associated citation sentences and the major players leading the field.",
"As a follow-up study, we will conduct a comparative study that compares the proposed approach with the general topic modeling technique such as LDA.",
"We also plan to investigate whether there is a different impact of using citation sentences and general meta-data such as abstract and title for topic analysis on facilitating idea sharing and scholarly communication.",
"In addition, we would like to consider the window size of citation sentences enriching citation context and to discover the authors' relationships among the neighboring citation sentences."
]
} | {
"paper_header_number": [
"1",
"2.1",
"2.2",
"2.4",
"4"
],
"paper_header_content": [
"Introduction",
"Main idea",
"Data collection",
"AJT Model",
"Conclusion"
]
} | GEM-SciDuet-train-7#paper-971#slide-8 | Results contd | Author Group 5 Author Group 6 Author Group Author Group 8
Journal Group 5 Journal Group 6 Journal Group 7 Journal Group 8
Annals of Oncology Cancer Cell
Annals of Oncology Breast Cancer Research | Author Group 5 Author Group 6 Author Group Author Group 8
Journal Group 5 Journal Group 6 Journal Group 7 Journal Group 8
Annals of Oncology Cancer Cell
Annals of Oncology Breast Cancer Research | [] |
GEM-SciDuet-train-7#paper-971#slide-9 | 971 | Exploring the leading authors and journals in major topics by citation sentences and topic modeling | Citation plays an important role in understanding the knowledge sharing among scholars. Citation sentences embed useful contents that signify the influence of cited authors on shared ideas, and express own opinion of citing authors to others' articles. The purpose of the study is to provide a new lens to analyze the topical relationship embedded in the citation sentences in an integrated manner. To this end, we extract citation sentences from full-text articles in the field of Oncology. In addition, we adopt Author-Journal-Topic (AJT) model to take both authors and journals into consideration of topic analysis. For the study, we collect the 6,360 full-text articles from PubMed Central and select the top 15 journals on Oncology. By applying AJT model, we identify what the major topics are shared among researchers in Oncology and which authors and journal lead the idea exchange in sub-disciplines of Oncology. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106
],
"paper_content_text": [
"Introduction As the size of data on the web continues to increase in an exponential manner, finding valuable meaning between data becomes of paramount importance in many research areas.",
"In the information science field, citations are challenging, pivotal materials to discover the relationship between academic documents because citations present the description of authors' ideas and the hidden relationship between authors and documents.",
"The earliest works focused mainly on classifying the citation behaviors and discovering the citation reasons with limited data such as the location of citation sentences and the number of references [1, 2] .",
"Since the mid-1990s, with the development of computer technology, citation content analysis was elaborated by applying data analysis techniques like text-mining or natural language processing.",
"Zhang et al.",
"[3] present citation analysis based on sematic and syntactic approaches.",
"Semantic-based citation analysis is performed by qualitative analysis to discover the citation motivation and citation classification.",
"On the other hand, syntactic-based citation analysis can be conducted by citation location and citation frequency, which reveals the hidden relation of authors by using meta-data of documents such as journal, venue of publication, affiliation of authors, etc.",
"Following their study, Ding et al.",
"[4] propose a theoretical methodology through content citation analysis.",
"However, these analyses are somewhat limited to the explicit context that primarily represents their own ideas and arguments.",
"The main goal of the paper is to discover the implicit topical relationships buried in citation sentences by utilizing the citation information from the author's perspective of sharing other authors' point of view.",
"Implicitness of the topical relationship is realized by using citation sentences as the input for the topic modeling technique.",
"In this study, a citation sentence indicates the sentence including citation expression consisting of year and author of the cited work.",
"In general, the citation sentence contains brief content of cited work and opinion that the author of citing work on the cited work.",
"We claim that citation sentences reveal interesting characteristics of scholarly communication such as influence, idea exchange, justification for citer's arguments, etc.",
"We assume that using citation sentences for topic analysis reveals aforementioned characteristics.",
"To explore such intellectual space created by citation sentences, we take both authors and journals into consideration of topic analysis.",
"To this end, we applied Author-Conference-Topic (ACT) model proposed by Tang et al.",
"[5] for our topic analysis in relation with both authors and journals, which is called Author-Journal-Topic (AJT) topic model.",
"ACT model is a probabilistic topic model for simultaneously extracting topics of papers, authors, and conferences.",
"There are a few studies to analyze content of citation sentences.",
"Most of previous studies focus on how the topic of document influences citation and vice versa [6, 7, 8] using Topic Modeling.",
"Kataria, Mitra, and Bhatia [8] adapt citation to Author-Topic model [9] with the assumption that the context surrounding the citation anchor could be used to get topical information about the cited authors.",
"These studies including Tang et al.",
"[10] 's ACT model are the examples of combining topic modelling methods and citation content analysis.",
"However, most previous studies used metadata of documents.",
"In this work, we focus on identifying the landscape of the oncology field from a perspective of citation.",
"By using citation sentences, our results can indicate which authors are actively cited and which journals lead a certain topic.",
"The rest of the paper is organized as follows: Section 2 describes the proposed approach.",
"Section 3 analyzes the topic modeling results.",
"Section 4 concludes the paper with the future work.",
"Methodology Main idea The basic assumption of the proposed approach is that citation sentences embed useful contents signifying the influence of cited authors on shared ideas of citing authors.",
"Citation sentences are also considered as an invisible intellectual place for idea exchanging since citations are effective means of supporting and expressing their own arguments by using other works.",
"In the similar vein, Di Marco and Mercer [11] claim that citation sentences play a major role in creating the relationship among relevant authors within the similar research fields.",
"With these assumptions, we are to explore the implicitness of topic relationships resided in citation sentences from the integrated perspective by incorporating the citing authors and journal titles into interpreting the topical relationships.",
"As shown in Figure 1 , we utilized various features including citing authors, citing sentences and journal titles for topic analysis.",
"Authors in Figure 1 mean the citing authors who write a paper and who cite other's work.",
"Citation sentences are the sentences written by the authors when they cite other's work in the paper, and journal titles are the journal names publishing the citing authors' paper.",
"By employing AJT model with these three parameters , we can discover which topics are the most salient ones referred to frequently by researchers and who are the leading authors sharing other authors' ideas in the research field and which journal leads such endeavor.",
"Fig.",
"1.",
"Three parameters for AJT model Data collection For this study, we compile the dataset on the field of Oncology from PubMed Central that provides the full-text in the biomedical field.",
"We select top 15 journals of Oncology by Thomson Reuter's JCR and journal's impact factor, and from these 15 journals, we are able to collect 6,360 full-text articles.",
"Figure 2 describes the workflow of our study.",
"As mentioned earlier, with the fulltext articles collected from PubMed Central, we extract the citation sentences.",
"Most citation sentences are kept in the following format: (author, year), (reference number) [reference number].",
"An example of such format is \"(<xref rid=\"bib00\" ref-type=\"bibr\">Author name, 2000</xref>)\".",
"We use the regular expression technique to parse and extract the citation sentences, when the tag <xref rid=>, </xref> appears on the sentences after parsing XML records with the Java-based SAX parser.",
"Method Fig.",
"2.",
"Workflow We also parse other metadata for AJT model such as the name of authors and journal titles.",
"The author tags, <surname> </surname> and <given-names></given-names> inside the <contrib-gourp></contrib-group>, denote the list of authors who wrote the paper.",
"For journal, we extract the titles when the journal tags, <journal-title> and </journal-title>, are included in the tag of <journal-meta> and </journal-meta>.",
"We also preprocess extracted sentences by removing both functional and general words and applying the Porter's stemming algorithm to improve the input for AJT Model.",
"AJT Model For our study, we apply ACT [10] model with several metadata such as citation sentences, journal titles and citing authors to develop AJT model.",
"Our AJT model utilizes journal titles and citation sentences instead of conference and abstract on documents.",
"The change of model is needed to analyze most influential topics in Oncology and to find leading authors who frequently mention the active topics and to detect the journals involved in such topics.",
"Figure 1 , Table 1) Like ACT model, AJT model assumes that each citing author is related to distribution over topics and each word in citation sentences is derived from a topic.",
"In the AJT model, the journal titles are related to each word.",
"To determine a word (ω_Si) in citation sentences (S), citing authors (x_Si) are consider for a word.",
"Each citing author is associated with a distributed topic.",
"A topic is generated from the citing author-topic distribution.",
"The words and journal titles are generated from a specific topic.",
"AJT model presents (1) 3 Results and Analyses For AJT model, we set the number of topics to 15 and finally select 8 topics as major topics.",
"Since we discovered that there are similar topics on our results, we calculated the similarity between 15 topics to select the most representative topics.",
"The topical similarities are measured by each word on topics and we calculated the similarities of two topics where each topic are represented in an array of a term vector.",
"Through this process, we chose 8 topics which have high topical similarities (over 0.5).",
"Each topic presents top 5 words from topic-word distribution, and 5 most related authors and journal titles are displayed along with each topic.",
"By performing several times on the pilot studies, we decided to choose top 5 words which are quite appropriate to describe each topics.",
"The results of AJT-based topic modeling is shown in Table 1 .",
"We label topic 1 \"breast cancer\" whose top words include breast, expression women, and growth.",
"Since the dataset is compiled with citation sentences, it implies that the topic \"breast cancer\" is a popular topic where researchers share and exchange ideas and facts related to breast cancer.",
"In relation to the topic \"breast cancer\", the active authors of breast cancer are Johnston Stephen RD, Colditz Graham A, and Sternlicht Mark D, and they share ideas with others on breast cancer from our results.",
"In terms of journals that provide a common place for idea sharing and communication, the journal \"Breast Cancer Research\" is the top journal of topic 1, and its impact factor is 5.49.",
"Authors such as Kurzrock Razelle, and Axelrod Haley in group 4 are the leading researchers sharing ideas on the topic \"targeted therapy.\"",
"The topic 4 is associated with the targeted therapy represented by words like mutations, treatments, therapy and disease.",
"The two most influential journals in topic 4 are \"Oncotarget\" and \"Journal of Thoracic Oncology\" whose impact factors are 6.36 and 5.28 respectively, which indicates that these two journals are the major journals encouraging authors to share ideas and collaborate with each other on cancer targeted therapy subject area.",
"Authors like Zitgel Laurence, Galluzzi Lorenzo, and Kroemer Guido in the author group 7 are the ones that actively share ideas about the topic \"Cancer Immunology.\"",
"Top concepts that are related to this topic are cell, immune, clinical and antitumor.",
"The top journal of the topic \"Cancer Immunology\" is Oncoinmmunology whose impact factor is 6.266.",
"Romagnani Paola and Salem Husein K in topic 8 \"Stem Cell\" are the authors that communicate and share ideas actively with each other in the given field, and the journal \"Stem Cells\" (impact factor: 6.523) is the leading journal.",
"We visualize topic keywords obtained from results of AJT-based topic model.",
"We construct the co-occurrence network and analyze which topic words play an important role in this domain.",
"Each node in the network represents a topic word, and an edge represents a co-occurrence frequency between keywords.",
"The size of nodes represents degree centrality and the color means network clusters obtained by using modularity algorithm.",
"This network consists of 100 nodes and 1,436 edges.",
"As shown in Figure 4 , each topic belongs to a specific community, but shares some important topic keywords.",
"Especially, the topic words positioned at the center is represented core-keywords in Oncology.",
"Figure 4 indicates that these words are the essential concepts of the Oncology domain.",
"Along with the results of AJT-based topic models, we can infer the major journals and authors develop their own research area based on these core-concepts.",
"Fig.",
"4.",
"Network of topic keywords The above results imply that the proposed approach identifies which topics are frequently shared, who facilitates to exchange ideas, and which journals provide a placeholder for it.",
"Identification of the triple relationship among authors, journals, and topics sheds new insight on understanding the well-discussed topics driven by the leading journals and authors that play a mediator role in the development of Oncology.",
"Conclusion One of the major research problems in bibliometrics is how to map out the intellectual structure of a research field.",
"The proposed approach tackles such research problem by utilizing citation sentences and AJT model.",
"By using citation sentences as the input for AJT model to find latent meaning, AJT model suggests a new way to detect leading authors and journals in sub-disciplines represented by discovered topics in a certain field.",
"Achieving this is not feasible by traditional frequency-based citation analysis.",
"One of the interesting observations is that the top-ranked journals in the discovered topics derived from AJT model are not ranked top in terms of JCR.",
"For example, the \"Oncotarget\" journal is the top-ranked journal in three topics in our analysis, but the ranking of the journal is 20 according to JCR.",
"Since we only report on preliminary results of our approach, we undertake in-depth analysis to investigate why this difference exists.",
"We also conduct various statistical tests on the results.",
"Based on the reported results in this paper, though, we claim that AJT can be used for discovering latent meaning associated citation sentences and the major players leading the field.",
"As a follow-up study, we will conduct a comparative study that compares the proposed approach with the general topic modeling technique such as LDA.",
"We also plan to investigate whether there is a different impact of using citation sentences and general meta-data such as abstract and title for topic analysis on facilitating idea sharing and scholarly communication.",
"In addition, we would like to consider the window size of citation sentences enriching citation context and to discover the authors' relationships among the neighboring citation sentences."
]
} | {
"paper_header_number": [
"1",
"2.1",
"2.2",
"2.4",
"4"
],
"paper_header_content": [
"Introduction",
"Main idea",
"Data collection",
"AJT Model",
"Conclusion"
]
} | GEM-SciDuet-train-7#paper-971#slide-9 | Conclusion | AJT model: to detect leading authors and journals in sub-disciplines
represented by discovered topics in a certain field
Citation sentences: Discovering latent meaning associated citation sentences
and the major players leading the field | AJT model: to detect leading authors and journals in sub-disciplines
represented by discovered topics in a certain field
Citation sentences: Discovering latent meaning associated citation sentences
and the major players leading the field | [] |
GEM-SciDuet-train-7#paper-971#slide-10 | 971 | Exploring the leading authors and journals in major topics by citation sentences and topic modeling | Citation plays an important role in understanding the knowledge sharing among scholars. Citation sentences embed useful contents that signify the influence of cited authors on shared ideas, and express own opinion of citing authors to others' articles. The purpose of the study is to provide a new lens to analyze the topical relationship embedded in the citation sentences in an integrated manner. To this end, we extract citation sentences from full-text articles in the field of Oncology. In addition, we adopt Author-Journal-Topic (AJT) model to take both authors and journals into consideration of topic analysis. For the study, we collect the 6,360 full-text articles from PubMed Central and select the top 15 journals on Oncology. By applying AJT model, we identify what the major topics are shared among researchers in Oncology and which authors and journal lead the idea exchange in sub-disciplines of Oncology. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106
],
"paper_content_text": [
"Introduction As the size of data on the web continues to increase in an exponential manner, finding valuable meaning between data becomes of paramount importance in many research areas.",
"In the information science field, citations are challenging, pivotal materials to discover the relationship between academic documents because citations present the description of authors' ideas and the hidden relationship between authors and documents.",
"The earliest works focused mainly on classifying the citation behaviors and discovering the citation reasons with limited data such as the location of citation sentences and the number of references [1, 2] .",
"Since the mid-1990s, with the development of computer technology, citation content analysis was elaborated by applying data analysis techniques like text-mining or natural language processing.",
"Zhang et al.",
"[3] present citation analysis based on sematic and syntactic approaches.",
"Semantic-based citation analysis is performed by qualitative analysis to discover the citation motivation and citation classification.",
"On the other hand, syntactic-based citation analysis can be conducted by citation location and citation frequency, which reveals the hidden relation of authors by using meta-data of documents such as journal, venue of publication, affiliation of authors, etc.",
"Following their study, Ding et al.",
"[4] propose a theoretical methodology through content citation analysis.",
"However, these analyses are somewhat limited to the explicit context that primarily represents their own ideas and arguments.",
"The main goal of the paper is to discover the implicit topical relationships buried in citation sentences by utilizing the citation information from the author's perspective of sharing other authors' point of view.",
"Implicitness of the topical relationship is realized by using citation sentences as the input for the topic modeling technique.",
"In this study, a citation sentence indicates the sentence including citation expression consisting of year and author of the cited work.",
"In general, the citation sentence contains brief content of cited work and opinion that the author of citing work on the cited work.",
"We claim that citation sentences reveal interesting characteristics of scholarly communication such as influence, idea exchange, justification for citer's arguments, etc.",
"We assume that using citation sentences for topic analysis reveals aforementioned characteristics.",
"To explore such intellectual space created by citation sentences, we take both authors and journals into consideration of topic analysis.",
"To this end, we applied Author-Conference-Topic (ACT) model proposed by Tang et al.",
"[5] for our topic analysis in relation with both authors and journals, which is called Author-Journal-Topic (AJT) topic model.",
"ACT model is a probabilistic topic model for simultaneously extracting topics of papers, authors, and conferences.",
"There are a few studies to analyze content of citation sentences.",
"Most of previous studies focus on how the topic of document influences citation and vice versa [6, 7, 8] using Topic Modeling.",
"Kataria, Mitra, and Bhatia [8] adapt citation to Author-Topic model [9] with the assumption that the context surrounding the citation anchor could be used to get topical information about the cited authors.",
"These studies including Tang et al.",
"[10] 's ACT model are the examples of combining topic modelling methods and citation content analysis.",
"However, most previous studies used metadata of documents.",
"In this work, we focus on identifying the landscape of the oncology field from a perspective of citation.",
"By using citation sentences, our results can indicate which authors are actively cited and which journals lead a certain topic.",
"The rest of the paper is organized as follows: Section 2 describes the proposed approach.",
"Section 3 analyzes the topic modeling results.",
"Section 4 concludes the paper with the future work.",
"Methodology Main idea The basic assumption of the proposed approach is that citation sentences embed useful contents signifying the influence of cited authors on shared ideas of citing authors.",
"Citation sentences are also considered as an invisible intellectual place for idea exchanging since citations are effective means of supporting and expressing their own arguments by using other works.",
"In the similar vein, Di Marco and Mercer [11] claim that citation sentences play a major role in creating the relationship among relevant authors within the similar research fields.",
"With these assumptions, we are to explore the implicitness of topic relationships resided in citation sentences from the integrated perspective by incorporating the citing authors and journal titles into interpreting the topical relationships.",
"As shown in Figure 1 , we utilized various features including citing authors, citing sentences and journal titles for topic analysis.",
"Authors in Figure 1 mean the citing authors who write a paper and who cite other's work.",
"Citation sentences are the sentences written by the authors when they cite other's work in the paper, and journal titles are the journal names publishing the citing authors' paper.",
"By employing AJT model with these three parameters , we can discover which topics are the most salient ones referred to frequently by researchers and who are the leading authors sharing other authors' ideas in the research field and which journal leads such endeavor.",
"Fig.",
"1.",
"Three parameters for AJT model Data collection For this study, we compile the dataset on the field of Oncology from PubMed Central that provides the full-text in the biomedical field.",
"We select top 15 journals of Oncology by Thomson Reuter's JCR and journal's impact factor, and from these 15 journals, we are able to collect 6,360 full-text articles.",
"Figure 2 describes the workflow of our study.",
"As mentioned earlier, with the fulltext articles collected from PubMed Central, we extract the citation sentences.",
"Most citation sentences are kept in the following format: (author, year), (reference number) [reference number].",
"An example of such format is \"(<xref rid=\"bib00\" ref-type=\"bibr\">Author name, 2000</xref>)\".",
"We use the regular expression technique to parse and extract the citation sentences, when the tag <xref rid=>, </xref> appears on the sentences after parsing XML records with the Java-based SAX parser.",
"Method Fig.",
"2.",
"Workflow We also parse other metadata for AJT model such as the name of authors and journal titles.",
"The author tags, <surname> </surname> and <given-names></given-names> inside the <contrib-gourp></contrib-group>, denote the list of authors who wrote the paper.",
"For journal, we extract the titles when the journal tags, <journal-title> and </journal-title>, are included in the tag of <journal-meta> and </journal-meta>.",
"We also preprocess extracted sentences by removing both functional and general words and applying the Porter's stemming algorithm to improve the input for AJT Model.",
"AJT Model For our study, we apply ACT [10] model with several metadata such as citation sentences, journal titles and citing authors to develop AJT model.",
"Our AJT model utilizes journal titles and citation sentences instead of conference and abstract on documents.",
"The change of model is needed to analyze most influential topics in Oncology and to find leading authors who frequently mention the active topics and to detect the journals involved in such topics.",
"Figure 1 , Table 1) Like ACT model, AJT model assumes that each citing author is related to distribution over topics and each word in citation sentences is derived from a topic.",
"In the AJT model, the journal titles are related to each word.",
"To determine a word (ω_Si) in citation sentences (S), citing authors (x_Si) are consider for a word.",
"Each citing author is associated with a distributed topic.",
"A topic is generated from the citing author-topic distribution.",
"The words and journal titles are generated from a specific topic.",
"AJT model presents (1) 3 Results and Analyses For AJT model, we set the number of topics to 15 and finally select 8 topics as major topics.",
"Since we discovered that there are similar topics on our results, we calculated the similarity between 15 topics to select the most representative topics.",
"The topical similarities are measured by each word on topics and we calculated the similarities of two topics where each topic are represented in an array of a term vector.",
"Through this process, we chose 8 topics which have high topical similarities (over 0.5).",
"Each topic presents top 5 words from topic-word distribution, and 5 most related authors and journal titles are displayed along with each topic.",
"By performing several times on the pilot studies, we decided to choose top 5 words which are quite appropriate to describe each topics.",
"The results of AJT-based topic modeling is shown in Table 1 .",
"We label topic 1 \"breast cancer\" whose top words include breast, expression women, and growth.",
"Since the dataset is compiled with citation sentences, it implies that the topic \"breast cancer\" is a popular topic where researchers share and exchange ideas and facts related to breast cancer.",
"In relation to the topic \"breast cancer\", the active authors of breast cancer are Johnston Stephen RD, Colditz Graham A, and Sternlicht Mark D, and they share ideas with others on breast cancer from our results.",
"In terms of journals that provide a common place for idea sharing and communication, the journal \"Breast Cancer Research\" is the top journal of topic 1, and its impact factor is 5.49.",
"Authors such as Kurzrock Razelle, and Axelrod Haley in group 4 are the leading researchers sharing ideas on the topic \"targeted therapy.\"",
"The topic 4 is associated with the targeted therapy represented by words like mutations, treatments, therapy and disease.",
"The two most influential journals in topic 4 are \"Oncotarget\" and \"Journal of Thoracic Oncology\" whose impact factors are 6.36 and 5.28 respectively, which indicates that these two journals are the major journals encouraging authors to share ideas and collaborate with each other on cancer targeted therapy subject area.",
"Authors like Zitgel Laurence, Galluzzi Lorenzo, and Kroemer Guido in the author group 7 are the ones that actively share ideas about the topic \"Cancer Immunology.\"",
"Top concepts that are related to this topic are cell, immune, clinical and antitumor.",
"The top journal of the topic \"Cancer Immunology\" is Oncoinmmunology whose impact factor is 6.266.",
"Romagnani Paola and Salem Husein K in topic 8 \"Stem Cell\" are the authors that communicate and share ideas actively with each other in the given field, and the journal \"Stem Cells\" (impact factor: 6.523) is the leading journal.",
"We visualize topic keywords obtained from results of AJT-based topic model.",
"We construct the co-occurrence network and analyze which topic words play an important role in this domain.",
"Each node in the network represents a topic word, and an edge represents a co-occurrence frequency between keywords.",
"The size of nodes represents degree centrality and the color means network clusters obtained by using modularity algorithm.",
"This network consists of 100 nodes and 1,436 edges.",
"As shown in Figure 4 , each topic belongs to a specific community, but shares some important topic keywords.",
"Especially, the topic words positioned at the center is represented core-keywords in Oncology.",
"Figure 4 indicates that these words are the essential concepts of the Oncology domain.",
"Along with the results of AJT-based topic models, we can infer the major journals and authors develop their own research area based on these core-concepts.",
"Fig.",
"4.",
"Network of topic keywords The above results imply that the proposed approach identifies which topics are frequently shared, who facilitates to exchange ideas, and which journals provide a placeholder for it.",
"Identification of the triple relationship among authors, journals, and topics sheds new insight on understanding the well-discussed topics driven by the leading journals and authors that play a mediator role in the development of Oncology.",
"Conclusion One of the major research problems in bibliometrics is how to map out the intellectual structure of a research field.",
"The proposed approach tackles such research problem by utilizing citation sentences and AJT model.",
"By using citation sentences as the input for AJT model to find latent meaning, AJT model suggests a new way to detect leading authors and journals in sub-disciplines represented by discovered topics in a certain field.",
"Achieving this is not feasible by traditional frequency-based citation analysis.",
"One of the interesting observations is that the top-ranked journals in the discovered topics derived from AJT model are not ranked top in terms of JCR.",
"For example, the \"Oncotarget\" journal is the top-ranked journal in three topics in our analysis, but the ranking of the journal is 20 according to JCR.",
"Since we only report on preliminary results of our approach, we undertake in-depth analysis to investigate why this difference exists.",
"We also conduct various statistical tests on the results.",
"Based on the reported results in this paper, though, we claim that AJT can be used for discovering latent meaning associated citation sentences and the major players leading the field.",
"As a follow-up study, we will conduct a comparative study that compares the proposed approach with the general topic modeling technique such as LDA.",
"We also plan to investigate whether there is a different impact of using citation sentences and general meta-data such as abstract and title for topic analysis on facilitating idea sharing and scholarly communication.",
"In addition, we would like to consider the window size of citation sentences enriching citation context and to discover the authors' relationships among the neighboring citation sentences."
]
} | {
"paper_header_number": [
"1",
"2.1",
"2.2",
"2.4",
"4"
],
"paper_header_content": [
"Introduction",
"Main idea",
"Data collection",
"AJT Model",
"Conclusion"
]
} | GEM-SciDuet-train-7#paper-971#slide-10 | Future works | Comparing the proposed approach with the general topic modeling
Investigating whether there is a different impact of using citation
sentences and general meta-data (abstract and title)
Considering the window size of citation sentences enriching citation | Comparing the proposed approach with the general topic modeling
Investigating whether there is a different impact of using citation
sentences and general meta-data (abstract and title)
Considering the window size of citation sentences enriching citation | [] |
GEM-SciDuet-train-8#paper-972#slide-0 | 972 | Obtaining SMT dictionaries for related languages | This study explores methods for developing Machine Translation dictionaries on the basis of word frequency lists coming from comparable corpora. We investigate (1) various methods to measure the similarity of cognates between related languages, (2) detection and removal of noisy cognate translations using SVM ranking. We show preliminary results on several Romance and Slavonic languages. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119
],
"paper_content_text": [
"Introduction Cognates are words having similarities in their spelling and meaning in two languages, either because the two languages are typologically related, e.g., maladie vs malattia ('disease'), or because they were both borrowed from the same source (informatique vs informatica).",
"The advantage of their use in Statistical Machine Translation (SMT) is that the procedure can be based on comparable corpora, i.e., similar corpora which are not translations of each other (Sharoff et al., 2013) .",
"Given that there are more sources of comparable corpora in comparison to parallel ones, the lexicon obtained from them is likely to be richer and more variable.",
"Detection of cognates is a well-known task, which has been explored for a range of languages using different methods.",
"The two main approaches applied to detection of the cognates are the generative and discriminative paradigms.",
"The first one is based on detection of the edit distance between potential candidate pairs.",
"The distance can be a simple Levenshtein distance, or a distance measure with the scores learned from an existing parallel set (Tiedemann, 1999; Mann and Yarowsky, 2001) .",
"The discriminative paradigm uses standard approaches to machine learning, which are based on (1) extracting features, e.g., character n-grams, and (2) learning to predict the transformations of the source word needed to (Jiampojamarn et al., 2010; Frunza and Inkpen, 2009) .",
"Given that SMT is usually based on a full-form lexicon, one of the possible issues in generation of cognates concerns the similarity of words in their root form vs the similarity in endings.",
"For example, the Ukrainian wordform áëèaeíüîãî 'near gen ' is cognate to Russian áëèaeíåãî, the root is identical, while the ending is considerably different (üîãî vs åãî).",
"Regular differences in the endings, which are shared across a large number of words, can be learned separately from the regular differences in the roots.",
"One also needs to take into account the false friends among cognates.",
"For example, diseñar means 'to design' in Spanish vs desenhar in Portuguese means 'to draw'.",
"There are also often cases of partial cognates, when the words share the meaning in some contexts, but not in others, e.g., aeåíà in Russian means 'wife', while its Bulgarian cognate aeåíà has two meanings: 'wife' and 'woman'.",
"Yet another complexity concerns a frequency mismatch.",
"Two cognates might differ in their frequency.",
"For example, dibujo in Spanish ('a drawing', rank 1779 in the Wikipedia frequency list) corresponds to a relatively rare cognate word debuxo in Portuguese (rank 104,514 in Wikipedia), while another Portuguese word desenho is more commonly used in this sense (rank 884 in the Portuguese Wikipedia).",
"For MT tasks we need translations that are equally appropriate in the source and target language, therefore cognates useful for a high-quality dictionary for SMT need to have roughly the same frequency in comparable corpora and they need to be used in similar contexts.",
"This study investigates the settings for extracting cognates for related languages in Romance and Slavonic language families for the task of reducing the number of unknown words for SMT.",
"This in-cludes the effects of having constraints for the cognates to be similar in their roots and in the endings, to occur in distributionally similar contexts and to have similar frequency.",
"Methodology The methodology for producing the list of cognates is based on the following steps: 1) Produce several lists of cognates using a family of distance measures, discussed in Section 2.1 from comparable corpora, 2) Prune the candidate lists by ranking items, this is done using a Machine Learning (ML) algorithm trained over parallel corpora for detecting the outliers, discussed in Section 2.2; The initial frequency lists for alignment are based Wikipedia dumps for the following languages: Romance (French, Italian, Spanish, Portuguese) and Slavonic (Bulgarian, Russian, Ukrainian), where the target languages are Spanish and Russian 1 .",
"Cognate detection We extract possible lists of cognates from comparable corpora by using a family of similarity measures: L direct matching between the languages using Levenshtein distance (Levenshtein, 1966) ; L(w s , w t ) = 1 − ed(w s , w t ) L-R Levenshtein distance with weights computed separately for the roots and for the endings; LR(r s , r t , e s , e t ) = α×ed(rs,rt)+β×ed(es,et) α+β L-C Levenshtein distance over word with similar number of starting characters (i.e.",
"prefix); LC(c s , c t ) = 1 − ed(c s , c t ), same prefix 0, otherwise where ed(., .)",
"is the normalised Levenshtein distance in characters between the source word w s and the target word w t .",
"The r s and r t are the stems produced by the Snowball stemmer 2 .",
"Since the Snowball stemmer does not support Ukrainian and Bulgarian, we used the Russian model for making the stem/ending split.",
"e s , e t are the characters at the end of a word form given a stem and c s , c t are the first n characters of a word.",
"In this work, we set the weights α = 0.6 and β = 0.4 giving more importance to the roots.",
"We set a higher weight to roots on the L-R, which is language dependent, and compare to the L-C metric, which is language independent.",
"We transform the Levenshtein distances into similarity metrics by subtracting the normalised distance score from one.",
"The produced lists contain for each source word the possible n-best target words accordingly to the maximum scores with one of the previous measures.",
"The n-best list allows possible cognate translations to a given source word that share a part of the surface form.",
"Different from (Mann and Yarowsky, 2001) , we produce n-best cognate lists scored by edit distance instead of 1-best.",
"An important problem when comparing comparable corpora is the way of representing the search space, where an exhaustive method compares all the combinations of source and target words (Mann and Yarowsky, 2001) .",
"We constraint the search space by comparing each source word against the target words that belong to a frequency window around the frequency of the source word.",
"This constraint only applies for the L and L-R metrics.",
"We use Wikipedia dumps for the source and target side processed in the form frequency lists.",
"We order the target side list into bins of similar frequency and for the source side we filter words that appear only once.",
"We use the window approach given that the frequency between the corpora under study can not be directly comparable.",
"During testing we use a wide window of ±200 bins to minimise the loss of good candidate translations.",
"The second search space constraint heuristic is the L-C metric.",
"This metric only compares source words with the target words upto a given n prefix.",
"For c s , c t in L-C , we use the first four characters to compare groups of words as suggested in (Kondrak et al., 2003) .",
"Cognate Ranking Given that the n-best lists contain noise, we aim to prune them by an ML ranking model.",
"However, there is a lack of resources to train a classification model for cognates (i.e.",
"cognate vs. false friend), as mentioned in (Fišer and Ljubešić, 2013) .",
"Available data that can be used to judge the cognate lists are the alignment pairs extracted from parallel data.",
"We decide to use a ranking model to avoid data imbalance present in classification and to use the probability scores of the alignment pairs as ranks, as opposed to the classification model used by (Irvine and Callison-Burch, 2013) .",
"Moreover, we also use a popular domain adaptation technique (Daumé et al., 2010) given that we have access to different domains of parallel training data that might be compatible with our comparable corpora.",
"The training data are the alignments between pairs of words where we rank them accordingly to their correspondent alignment probability from the output of GIZA++ (Och and Ney, 2003) .",
"We then use a heuristic to prune training data in order to simulate cognate words.",
"Pairs of words scored below the Levenshtein similarity threshold of 0.5 are not considered as cognates given that they are likely to have a different surface form.",
"We represent the training and test data with features extracted from different edit distance scores and distributional measures.",
"The edit distances features are as follows: 1) Similarity measure L and 2) Number of times of each edit operation.",
"Thus, the model assigns a different importance to each operation.",
"The distributional feature is based on the cosine between the distributional vectors of a window of n words around the word currently under comparison.",
"We train distributional similarity models with word2vec (Mikolov et al., 2013a) for the source and target side separately.",
"We extract the continuous vector for each word in the window, concatenate it and then compute the cosine between the concatenated vectors of the source and the target.",
"We suspect that the vectors will have similar behaviour between the source and the target given that they are trained under parallel Wikipedia articles.",
"We develop two ML models: 1) Edit distance scores and 2) Edit distance scores and distributional similarity score.",
"We use SVMlight (Joachims, 1998) Results and Discussion In this section we describe the data used to produce the n-best lists and train the cognate ranking models.",
"We evaluate the ranking models with heldout data from each training domain.",
"We also provide manual evaluation over the ranked n-best lists for error analysis.",
"Data The n-best lists to detect cognates were extracted from the respective Wikipedias by using the method described in Section 2.1.",
"The training data for the ranking model consists of different types of parallel corpora.",
"The parallel corpora are as follows: 1) Wiki-titles we use the inter language links to create a parallel corpus from the tittles of the Wikipedia articles, with about 500K aligned links (i.e.",
"'sentences') per language pair (about 200k for bg-ru), giving us about 200K training instances per language pair 3 , 2) Opensubs is an open source corpus of subtitles built by the fan community, with 1M sentences, 6M tokens, 100K words, giving about 90K training instances (Tiedemann, 2012) and 3) Zoo is a proprietary corpus of subtitles produced by professional translators, with 100K sentences, 700K tokens, 40K words and giving about 20K training instances per language pair.",
"Our objective is to create MT dictionaries from the produced n-best lists and we use parallel data as a source of training to prune them.",
"We are interested in the corpora of subtitles because the chosen domain of our SMT experiments is subtitling, while the proposed ranking method can be used in other application domains as well.",
"We consider Zoo and Opensubs as two different domains given that they were built by different types of translators and they differ in size and quality.",
"The heldout data consists of 2K instances for each corpus.",
"We use Wikipedia documents and Opensusbs subtitles to train word2vec for the distributional similarity feature.",
"We use the continuous bag-ofwords algorithm for word2vec and set the parameters for training to 200 dimensions and a window of 8 words.",
"The Wikipedia documents with an average number of 70K documents for each language, and Opensubs subtitles with 1M sentences for each language.",
"In practice we only use the Wikipedia data given that for Opensubs the model is able to find relatively few vectors, for example a vector is found only for 20% of the words in the pt-es pair.",
"Evaluation of the Ranking Model We define two ranking models as: model E for edit distance features and model EC for both edit Table 1 shows the results of the ranking procedure.",
"For the Romance family language pairs the model EC with context features consistently reduces the error compared to the solely use of edit distance metrics.",
"The only exception is the it-es EC model with poor results for the domain of Wiki-titles.",
"The models for the Slavonic family behave similarly to the Romance family, where the use of context features reduces the ranking error.",
"The exception is the bg-ru model on the Opensubs domain.",
"A possible reason for the poor results on the ites and bg-ru models is that the model often assigns a high similarity score to unrelated words.",
"For example, in it-es, mortes 'deads' is treated as close to categoria 'category'.",
"A possible solution is to map the vectors form the source side into the space of the target side via a learned transformation matrix (Mikolov et al., 2013b) .",
"Preliminary Results on Comparable Corpora After we extracted the n-best lists for the Romance family comparable corpora, we applied one of the ranking models on these lists and we manually evaluated over a sample of 50 words 4 .",
"We set n to 10 for the n-best lists.",
"We use a frequency window of 200 for the n-best list search heuristic and the domain of the comparable corpora to Wiki-titles 4 The sample consists of words with a frequency between 2K and 5. for the domain adaptation technique.",
"The purpose of manual evaluation is to decide whether the ML setup is sensible on the objective task.",
"Each list is evaluated by accuracy at 1 and accuracy at 10.",
"We also show success and failure examples of the ranking and the n-best lists.",
"Table 2 shows the preliminary results of the ML model E on a sample of Wikipedia dumps.",
"The L and L-R lists consistently show poor results.",
"A possible reason is the amount of errors given the first step to extract the n-best lists.",
"For example, in pt-es, for the word vivem 'live' the 10-best list only contain one word with a similar meaning viva 'living' but it can be also translated as 'cheers'.",
"In the pt-es list for the word representação 'description' the correct translation representación is not among the 10-best in the L list.",
"However, it is present in the 10-best for the L-C list and the ML model EC ranks it in the first place.",
"The edit distance model E still makes mistakes like with the list L-C, the word vivem 'live' translates into viven 'living' and the correct translation is vivir.",
"However, given a certain context/sense the previous translation can be correct.",
"The ranking scores given by the SVM varies from each list version.",
"For the L-C lists the scores are more uniform in increasing order and with a small variance.",
"The L and L-R lists show the opposite behaviour.",
"We add the produced Wikipedia n-best lists with the L metric into a SMT training dataset for the ptes pair.",
"We use the Moses SMT toolkit (Koehn et al., 2007) to test the augmented datasets.",
"We compare the augmented model with a baseline both trained by using the Zoo corpus of subtitles.",
"We use a 1-best list consisting of 100K pairs.",
"Te dataset used for pt-es baseline is: 80K training sentences, 1K sentences for tuning and 2K sen- Lang Pairs acc@1 acc@10 acc@1 acc@10 acc@1 acc@10 pt-es 20 60 22 59 32 70 it-es 16 53 18 45 44 66 fr-es 10 48 12 51 29 59 A possible reason for low improvement in terms of the BLEU scores is because MT evaluation metrics, such as BLEU, compare the MT output with a human reference.",
"The human reference translations in our corpus have been done from English (e.g., En→Es), while the test translations come from a related language (En→Pt→Es), often resulting in different paraphrases of the same English source.",
"While our OOV rate improved, the evaluation scores did not reflected this, because our MT output was still far from the reference even in cases it was otherwise acceptable.",
"List L List L-R List L-C Conclusions and future Work We have presented work in progress for developing MT dictionaries extracted from comparable resources for related languages.",
"The extraction heuristic show positive results on the n-best lists that group words with the same starting char-5 https://github.com/clab/fast_align 6 https://kheafield.com/code/kenlm/ 7 The p-value for the uk-ru pair is 0.06 we do not consider this result as statistically significant.",
"acters, because the used comparable corpora consist of related languages that share a similar orthography.",
"However, the lists based on the frequency window heuristic show poor results to include the correct translations during the extraction step.",
"Our ML models based on similarity metrics over parallel corpora show rankings similar to heldout data.",
"However, we created our training data using simple heuristics that simulate cognate words (i.e.",
"pairs of words with a small surface difference).",
"The ML models are able to rank similar words on the top of the list and they give a reliable score to discriminate wrong translations.",
"Preliminary results on the addition of the n-best lists into an SMT system show modest improvements compare to the baseline.",
"However, the OOV rate shows improvements around 10% reduction on word types, because of the wide variety of lexical choices introduced by the MT dictionaries.",
"Future work involves the addition of unsupervised morphology features for the n-best list extraction, i.e.",
"first step, given that the use of starting characters shows to be an effective heuristic to prune the search space and language independent.",
"Finally, we will measure the contribution for all the produced cognate lists, where we can try different strategies to add the dictionaries into an SMT system (Irvine and Callison-Burch, 2014) ."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"3",
"3.1",
"3.2",
"3.3",
"4"
],
"paper_header_content": [
"Introduction",
"Methodology",
"Cognate detection",
"Cognate Ranking",
"Results and Discussion",
"Data",
"Evaluation of the Ranking Model",
"Preliminary Results on Comparable Corpora",
"Conclusions and future Work"
]
} | GEM-SciDuet-train-8#paper-972#slide-0 | Motivation | Extracting cognates for related languages in Romance and
Reducing the number of unknown words on SMT training data
Learning regular differences in words roots/endings shared across related languages | Extracting cognates for related languages in Romance and
Reducing the number of unknown words on SMT training data
Learning regular differences in words roots/endings shared across related languages | [] |
GEM-SciDuet-train-8#paper-972#slide-1 | 972 | Obtaining SMT dictionaries for related languages | This study explores methods for developing Machine Translation dictionaries on the basis of word frequency lists coming from comparable corpora. We investigate (1) various methods to measure the similarity of cognates between related languages, (2) detection and removal of noisy cognate translations using SVM ranking. We show preliminary results on several Romance and Slavonic languages. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119
],
"paper_content_text": [
"Introduction Cognates are words having similarities in their spelling and meaning in two languages, either because the two languages are typologically related, e.g., maladie vs malattia ('disease'), or because they were both borrowed from the same source (informatique vs informatica).",
"The advantage of their use in Statistical Machine Translation (SMT) is that the procedure can be based on comparable corpora, i.e., similar corpora which are not translations of each other (Sharoff et al., 2013) .",
"Given that there are more sources of comparable corpora in comparison to parallel ones, the lexicon obtained from them is likely to be richer and more variable.",
"Detection of cognates is a well-known task, which has been explored for a range of languages using different methods.",
"The two main approaches applied to detection of the cognates are the generative and discriminative paradigms.",
"The first one is based on detection of the edit distance between potential candidate pairs.",
"The distance can be a simple Levenshtein distance, or a distance measure with the scores learned from an existing parallel set (Tiedemann, 1999; Mann and Yarowsky, 2001) .",
"The discriminative paradigm uses standard approaches to machine learning, which are based on (1) extracting features, e.g., character n-grams, and (2) learning to predict the transformations of the source word needed to (Jiampojamarn et al., 2010; Frunza and Inkpen, 2009) .",
"Given that SMT is usually based on a full-form lexicon, one of the possible issues in generation of cognates concerns the similarity of words in their root form vs the similarity in endings.",
"For example, the Ukrainian wordform áëèaeíüîãî 'near gen ' is cognate to Russian áëèaeíåãî, the root is identical, while the ending is considerably different (üîãî vs åãî).",
"Regular differences in the endings, which are shared across a large number of words, can be learned separately from the regular differences in the roots.",
"One also needs to take into account the false friends among cognates.",
"For example, diseñar means 'to design' in Spanish vs desenhar in Portuguese means 'to draw'.",
"There are also often cases of partial cognates, when the words share the meaning in some contexts, but not in others, e.g., aeåíà in Russian means 'wife', while its Bulgarian cognate aeåíà has two meanings: 'wife' and 'woman'.",
"Yet another complexity concerns a frequency mismatch.",
"Two cognates might differ in their frequency.",
"For example, dibujo in Spanish ('a drawing', rank 1779 in the Wikipedia frequency list) corresponds to a relatively rare cognate word debuxo in Portuguese (rank 104,514 in Wikipedia), while another Portuguese word desenho is more commonly used in this sense (rank 884 in the Portuguese Wikipedia).",
"For MT tasks we need translations that are equally appropriate in the source and target language, therefore cognates useful for a high-quality dictionary for SMT need to have roughly the same frequency in comparable corpora and they need to be used in similar contexts.",
"This study investigates the settings for extracting cognates for related languages in Romance and Slavonic language families for the task of reducing the number of unknown words for SMT.",
"This in-cludes the effects of having constraints for the cognates to be similar in their roots and in the endings, to occur in distributionally similar contexts and to have similar frequency.",
"Methodology The methodology for producing the list of cognates is based on the following steps: 1) Produce several lists of cognates using a family of distance measures, discussed in Section 2.1 from comparable corpora, 2) Prune the candidate lists by ranking items, this is done using a Machine Learning (ML) algorithm trained over parallel corpora for detecting the outliers, discussed in Section 2.2; The initial frequency lists for alignment are based Wikipedia dumps for the following languages: Romance (French, Italian, Spanish, Portuguese) and Slavonic (Bulgarian, Russian, Ukrainian), where the target languages are Spanish and Russian 1 .",
"Cognate detection We extract possible lists of cognates from comparable corpora by using a family of similarity measures: L direct matching between the languages using Levenshtein distance (Levenshtein, 1966) ; L(w s , w t ) = 1 − ed(w s , w t ) L-R Levenshtein distance with weights computed separately for the roots and for the endings; LR(r s , r t , e s , e t ) = α×ed(rs,rt)+β×ed(es,et) α+β L-C Levenshtein distance over word with similar number of starting characters (i.e.",
"prefix); LC(c s , c t ) = 1 − ed(c s , c t ), same prefix 0, otherwise where ed(., .)",
"is the normalised Levenshtein distance in characters between the source word w s and the target word w t .",
"The r s and r t are the stems produced by the Snowball stemmer 2 .",
"Since the Snowball stemmer does not support Ukrainian and Bulgarian, we used the Russian model for making the stem/ending split.",
"e s , e t are the characters at the end of a word form given a stem and c s , c t are the first n characters of a word.",
"In this work, we set the weights α = 0.6 and β = 0.4 giving more importance to the roots.",
"We set a higher weight to roots on the L-R, which is language dependent, and compare to the L-C metric, which is language independent.",
"We transform the Levenshtein distances into similarity metrics by subtracting the normalised distance score from one.",
"The produced lists contain for each source word the possible n-best target words accordingly to the maximum scores with one of the previous measures.",
"The n-best list allows possible cognate translations to a given source word that share a part of the surface form.",
"Different from (Mann and Yarowsky, 2001) , we produce n-best cognate lists scored by edit distance instead of 1-best.",
"An important problem when comparing comparable corpora is the way of representing the search space, where an exhaustive method compares all the combinations of source and target words (Mann and Yarowsky, 2001) .",
"We constraint the search space by comparing each source word against the target words that belong to a frequency window around the frequency of the source word.",
"This constraint only applies for the L and L-R metrics.",
"We use Wikipedia dumps for the source and target side processed in the form frequency lists.",
"We order the target side list into bins of similar frequency and for the source side we filter words that appear only once.",
"We use the window approach given that the frequency between the corpora under study can not be directly comparable.",
"During testing we use a wide window of ±200 bins to minimise the loss of good candidate translations.",
"The second search space constraint heuristic is the L-C metric.",
"This metric only compares source words with the target words upto a given n prefix.",
"For c s , c t in L-C , we use the first four characters to compare groups of words as suggested in (Kondrak et al., 2003) .",
"Cognate Ranking Given that the n-best lists contain noise, we aim to prune them by an ML ranking model.",
"However, there is a lack of resources to train a classification model for cognates (i.e.",
"cognate vs. false friend), as mentioned in (Fišer and Ljubešić, 2013) .",
"Available data that can be used to judge the cognate lists are the alignment pairs extracted from parallel data.",
"We decide to use a ranking model to avoid data imbalance present in classification and to use the probability scores of the alignment pairs as ranks, as opposed to the classification model used by (Irvine and Callison-Burch, 2013) .",
"Moreover, we also use a popular domain adaptation technique (Daumé et al., 2010) given that we have access to different domains of parallel training data that might be compatible with our comparable corpora.",
"The training data are the alignments between pairs of words where we rank them accordingly to their correspondent alignment probability from the output of GIZA++ (Och and Ney, 2003) .",
"We then use a heuristic to prune training data in order to simulate cognate words.",
"Pairs of words scored below the Levenshtein similarity threshold of 0.5 are not considered as cognates given that they are likely to have a different surface form.",
"We represent the training and test data with features extracted from different edit distance scores and distributional measures.",
"The edit distances features are as follows: 1) Similarity measure L and 2) Number of times of each edit operation.",
"Thus, the model assigns a different importance to each operation.",
"The distributional feature is based on the cosine between the distributional vectors of a window of n words around the word currently under comparison.",
"We train distributional similarity models with word2vec (Mikolov et al., 2013a) for the source and target side separately.",
"We extract the continuous vector for each word in the window, concatenate it and then compute the cosine between the concatenated vectors of the source and the target.",
"We suspect that the vectors will have similar behaviour between the source and the target given that they are trained under parallel Wikipedia articles.",
"We develop two ML models: 1) Edit distance scores and 2) Edit distance scores and distributional similarity score.",
"We use SVMlight (Joachims, 1998) Results and Discussion In this section we describe the data used to produce the n-best lists and train the cognate ranking models.",
"We evaluate the ranking models with heldout data from each training domain.",
"We also provide manual evaluation over the ranked n-best lists for error analysis.",
"Data The n-best lists to detect cognates were extracted from the respective Wikipedias by using the method described in Section 2.1.",
"The training data for the ranking model consists of different types of parallel corpora.",
"The parallel corpora are as follows: 1) Wiki-titles we use the inter language links to create a parallel corpus from the tittles of the Wikipedia articles, with about 500K aligned links (i.e.",
"'sentences') per language pair (about 200k for bg-ru), giving us about 200K training instances per language pair 3 , 2) Opensubs is an open source corpus of subtitles built by the fan community, with 1M sentences, 6M tokens, 100K words, giving about 90K training instances (Tiedemann, 2012) and 3) Zoo is a proprietary corpus of subtitles produced by professional translators, with 100K sentences, 700K tokens, 40K words and giving about 20K training instances per language pair.",
"Our objective is to create MT dictionaries from the produced n-best lists and we use parallel data as a source of training to prune them.",
"We are interested in the corpora of subtitles because the chosen domain of our SMT experiments is subtitling, while the proposed ranking method can be used in other application domains as well.",
"We consider Zoo and Opensubs as two different domains given that they were built by different types of translators and they differ in size and quality.",
"The heldout data consists of 2K instances for each corpus.",
"We use Wikipedia documents and Opensusbs subtitles to train word2vec for the distributional similarity feature.",
"We use the continuous bag-ofwords algorithm for word2vec and set the parameters for training to 200 dimensions and a window of 8 words.",
"The Wikipedia documents with an average number of 70K documents for each language, and Opensubs subtitles with 1M sentences for each language.",
"In practice we only use the Wikipedia data given that for Opensubs the model is able to find relatively few vectors, for example a vector is found only for 20% of the words in the pt-es pair.",
"Evaluation of the Ranking Model We define two ranking models as: model E for edit distance features and model EC for both edit Table 1 shows the results of the ranking procedure.",
"For the Romance family language pairs the model EC with context features consistently reduces the error compared to the solely use of edit distance metrics.",
"The only exception is the it-es EC model with poor results for the domain of Wiki-titles.",
"The models for the Slavonic family behave similarly to the Romance family, where the use of context features reduces the ranking error.",
"The exception is the bg-ru model on the Opensubs domain.",
"A possible reason for the poor results on the ites and bg-ru models is that the model often assigns a high similarity score to unrelated words.",
"For example, in it-es, mortes 'deads' is treated as close to categoria 'category'.",
"A possible solution is to map the vectors form the source side into the space of the target side via a learned transformation matrix (Mikolov et al., 2013b) .",
"Preliminary Results on Comparable Corpora After we extracted the n-best lists for the Romance family comparable corpora, we applied one of the ranking models on these lists and we manually evaluated over a sample of 50 words 4 .",
"We set n to 10 for the n-best lists.",
"We use a frequency window of 200 for the n-best list search heuristic and the domain of the comparable corpora to Wiki-titles 4 The sample consists of words with a frequency between 2K and 5. for the domain adaptation technique.",
"The purpose of manual evaluation is to decide whether the ML setup is sensible on the objective task.",
"Each list is evaluated by accuracy at 1 and accuracy at 10.",
"We also show success and failure examples of the ranking and the n-best lists.",
"Table 2 shows the preliminary results of the ML model E on a sample of Wikipedia dumps.",
"The L and L-R lists consistently show poor results.",
"A possible reason is the amount of errors given the first step to extract the n-best lists.",
"For example, in pt-es, for the word vivem 'live' the 10-best list only contain one word with a similar meaning viva 'living' but it can be also translated as 'cheers'.",
"In the pt-es list for the word representação 'description' the correct translation representación is not among the 10-best in the L list.",
"However, it is present in the 10-best for the L-C list and the ML model EC ranks it in the first place.",
"The edit distance model E still makes mistakes like with the list L-C, the word vivem 'live' translates into viven 'living' and the correct translation is vivir.",
"However, given a certain context/sense the previous translation can be correct.",
"The ranking scores given by the SVM varies from each list version.",
"For the L-C lists the scores are more uniform in increasing order and with a small variance.",
"The L and L-R lists show the opposite behaviour.",
"We add the produced Wikipedia n-best lists with the L metric into a SMT training dataset for the ptes pair.",
"We use the Moses SMT toolkit (Koehn et al., 2007) to test the augmented datasets.",
"We compare the augmented model with a baseline both trained by using the Zoo corpus of subtitles.",
"We use a 1-best list consisting of 100K pairs.",
"Te dataset used for pt-es baseline is: 80K training sentences, 1K sentences for tuning and 2K sen- Lang Pairs acc@1 acc@10 acc@1 acc@10 acc@1 acc@10 pt-es 20 60 22 59 32 70 it-es 16 53 18 45 44 66 fr-es 10 48 12 51 29 59 A possible reason for low improvement in terms of the BLEU scores is because MT evaluation metrics, such as BLEU, compare the MT output with a human reference.",
"The human reference translations in our corpus have been done from English (e.g., En→Es), while the test translations come from a related language (En→Pt→Es), often resulting in different paraphrases of the same English source.",
"While our OOV rate improved, the evaluation scores did not reflected this, because our MT output was still far from the reference even in cases it was otherwise acceptable.",
"List L List L-R List L-C Conclusions and future Work We have presented work in progress for developing MT dictionaries extracted from comparable resources for related languages.",
"The extraction heuristic show positive results on the n-best lists that group words with the same starting char-5 https://github.com/clab/fast_align 6 https://kheafield.com/code/kenlm/ 7 The p-value for the uk-ru pair is 0.06 we do not consider this result as statistically significant.",
"acters, because the used comparable corpora consist of related languages that share a similar orthography.",
"However, the lists based on the frequency window heuristic show poor results to include the correct translations during the extraction step.",
"Our ML models based on similarity metrics over parallel corpora show rankings similar to heldout data.",
"However, we created our training data using simple heuristics that simulate cognate words (i.e.",
"pairs of words with a small surface difference).",
"The ML models are able to rank similar words on the top of the list and they give a reliable score to discriminate wrong translations.",
"Preliminary results on the addition of the n-best lists into an SMT system show modest improvements compare to the baseline.",
"However, the OOV rate shows improvements around 10% reduction on word types, because of the wide variety of lexical choices introduced by the MT dictionaries.",
"Future work involves the addition of unsupervised morphology features for the n-best list extraction, i.e.",
"first step, given that the use of starting characters shows to be an effective heuristic to prune the search space and language independent.",
"Finally, we will measure the contribution for all the produced cognate lists, where we can try different strategies to add the dictionaries into an SMT system (Irvine and Callison-Burch, 2014) ."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"3",
"3.1",
"3.2",
"3.3",
"4"
],
"paper_header_content": [
"Introduction",
"Methodology",
"Cognate detection",
"Cognate Ranking",
"Results and Discussion",
"Data",
"Evaluation of the Ranking Model",
"Preliminary Results on Comparable Corpora",
"Conclusions and future Work"
]
} | GEM-SciDuet-train-8#paper-972#slide-1 | Method | Produce n-best lists of cognates using a family of distance measures from comparable corpora
Prune the n-best lists by ranking Machine Learning (ML) algorithm trained over parallel corpora
Motivation n-best list allows surface variation on possible cognate translations | Produce n-best lists of cognates using a family of distance measures from comparable corpora
Prune the n-best lists by ranking Machine Learning (ML) algorithm trained over parallel corpora
Motivation n-best list allows surface variation on possible cognate translations | [] |
GEM-SciDuet-train-8#paper-972#slide-2 | 972 | Obtaining SMT dictionaries for related languages | This study explores methods for developing Machine Translation dictionaries on the basis of word frequency lists coming from comparable corpora. We investigate (1) various methods to measure the similarity of cognates between related languages, (2) detection and removal of noisy cognate translations using SVM ranking. We show preliminary results on several Romance and Slavonic languages. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119
],
"paper_content_text": [
"Introduction Cognates are words having similarities in their spelling and meaning in two languages, either because the two languages are typologically related, e.g., maladie vs malattia ('disease'), or because they were both borrowed from the same source (informatique vs informatica).",
"The advantage of their use in Statistical Machine Translation (SMT) is that the procedure can be based on comparable corpora, i.e., similar corpora which are not translations of each other (Sharoff et al., 2013) .",
"Given that there are more sources of comparable corpora in comparison to parallel ones, the lexicon obtained from them is likely to be richer and more variable.",
"Detection of cognates is a well-known task, which has been explored for a range of languages using different methods.",
"The two main approaches applied to detection of the cognates are the generative and discriminative paradigms.",
"The first one is based on detection of the edit distance between potential candidate pairs.",
"The distance can be a simple Levenshtein distance, or a distance measure with the scores learned from an existing parallel set (Tiedemann, 1999; Mann and Yarowsky, 2001) .",
"The discriminative paradigm uses standard approaches to machine learning, which are based on (1) extracting features, e.g., character n-grams, and (2) learning to predict the transformations of the source word needed to (Jiampojamarn et al., 2010; Frunza and Inkpen, 2009) .",
"Given that SMT is usually based on a full-form lexicon, one of the possible issues in generation of cognates concerns the similarity of words in their root form vs the similarity in endings.",
"For example, the Ukrainian wordform áëèaeíüîãî 'near gen ' is cognate to Russian áëèaeíåãî, the root is identical, while the ending is considerably different (üîãî vs åãî).",
"Regular differences in the endings, which are shared across a large number of words, can be learned separately from the regular differences in the roots.",
"One also needs to take into account the false friends among cognates.",
"For example, diseñar means 'to design' in Spanish vs desenhar in Portuguese means 'to draw'.",
"There are also often cases of partial cognates, when the words share the meaning in some contexts, but not in others, e.g., aeåíà in Russian means 'wife', while its Bulgarian cognate aeåíà has two meanings: 'wife' and 'woman'.",
"Yet another complexity concerns a frequency mismatch.",
"Two cognates might differ in their frequency.",
"For example, dibujo in Spanish ('a drawing', rank 1779 in the Wikipedia frequency list) corresponds to a relatively rare cognate word debuxo in Portuguese (rank 104,514 in Wikipedia), while another Portuguese word desenho is more commonly used in this sense (rank 884 in the Portuguese Wikipedia).",
"For MT tasks we need translations that are equally appropriate in the source and target language, therefore cognates useful for a high-quality dictionary for SMT need to have roughly the same frequency in comparable corpora and they need to be used in similar contexts.",
"This study investigates the settings for extracting cognates for related languages in Romance and Slavonic language families for the task of reducing the number of unknown words for SMT.",
"This in-cludes the effects of having constraints for the cognates to be similar in their roots and in the endings, to occur in distributionally similar contexts and to have similar frequency.",
"Methodology The methodology for producing the list of cognates is based on the following steps: 1) Produce several lists of cognates using a family of distance measures, discussed in Section 2.1 from comparable corpora, 2) Prune the candidate lists by ranking items, this is done using a Machine Learning (ML) algorithm trained over parallel corpora for detecting the outliers, discussed in Section 2.2; The initial frequency lists for alignment are based Wikipedia dumps for the following languages: Romance (French, Italian, Spanish, Portuguese) and Slavonic (Bulgarian, Russian, Ukrainian), where the target languages are Spanish and Russian 1 .",
"Cognate detection We extract possible lists of cognates from comparable corpora by using a family of similarity measures: L direct matching between the languages using Levenshtein distance (Levenshtein, 1966) ; L(w s , w t ) = 1 − ed(w s , w t ) L-R Levenshtein distance with weights computed separately for the roots and for the endings; LR(r s , r t , e s , e t ) = α×ed(rs,rt)+β×ed(es,et) α+β L-C Levenshtein distance over word with similar number of starting characters (i.e.",
"prefix); LC(c s , c t ) = 1 − ed(c s , c t ), same prefix 0, otherwise where ed(., .)",
"is the normalised Levenshtein distance in characters between the source word w s and the target word w t .",
"The r s and r t are the stems produced by the Snowball stemmer 2 .",
"Since the Snowball stemmer does not support Ukrainian and Bulgarian, we used the Russian model for making the stem/ending split.",
"e s , e t are the characters at the end of a word form given a stem and c s , c t are the first n characters of a word.",
"In this work, we set the weights α = 0.6 and β = 0.4 giving more importance to the roots.",
"We set a higher weight to roots on the L-R, which is language dependent, and compare to the L-C metric, which is language independent.",
"We transform the Levenshtein distances into similarity metrics by subtracting the normalised distance score from one.",
"The produced lists contain for each source word the possible n-best target words accordingly to the maximum scores with one of the previous measures.",
"The n-best list allows possible cognate translations to a given source word that share a part of the surface form.",
"Different from (Mann and Yarowsky, 2001) , we produce n-best cognate lists scored by edit distance instead of 1-best.",
"An important problem when comparing comparable corpora is the way of representing the search space, where an exhaustive method compares all the combinations of source and target words (Mann and Yarowsky, 2001) .",
"We constraint the search space by comparing each source word against the target words that belong to a frequency window around the frequency of the source word.",
"This constraint only applies for the L and L-R metrics.",
"We use Wikipedia dumps for the source and target side processed in the form frequency lists.",
"We order the target side list into bins of similar frequency and for the source side we filter words that appear only once.",
"We use the window approach given that the frequency between the corpora under study can not be directly comparable.",
"During testing we use a wide window of ±200 bins to minimise the loss of good candidate translations.",
"The second search space constraint heuristic is the L-C metric.",
"This metric only compares source words with the target words upto a given n prefix.",
"For c s , c t in L-C , we use the first four characters to compare groups of words as suggested in (Kondrak et al., 2003) .",
"Cognate Ranking Given that the n-best lists contain noise, we aim to prune them by an ML ranking model.",
"However, there is a lack of resources to train a classification model for cognates (i.e.",
"cognate vs. false friend), as mentioned in (Fišer and Ljubešić, 2013) .",
"Available data that can be used to judge the cognate lists are the alignment pairs extracted from parallel data.",
"We decide to use a ranking model to avoid data imbalance present in classification and to use the probability scores of the alignment pairs as ranks, as opposed to the classification model used by (Irvine and Callison-Burch, 2013) .",
"Moreover, we also use a popular domain adaptation technique (Daumé et al., 2010) given that we have access to different domains of parallel training data that might be compatible with our comparable corpora.",
"The training data are the alignments between pairs of words where we rank them accordingly to their correspondent alignment probability from the output of GIZA++ (Och and Ney, 2003) .",
"We then use a heuristic to prune training data in order to simulate cognate words.",
"Pairs of words scored below the Levenshtein similarity threshold of 0.5 are not considered as cognates given that they are likely to have a different surface form.",
"We represent the training and test data with features extracted from different edit distance scores and distributional measures.",
"The edit distances features are as follows: 1) Similarity measure L and 2) Number of times of each edit operation.",
"Thus, the model assigns a different importance to each operation.",
"The distributional feature is based on the cosine between the distributional vectors of a window of n words around the word currently under comparison.",
"We train distributional similarity models with word2vec (Mikolov et al., 2013a) for the source and target side separately.",
"We extract the continuous vector for each word in the window, concatenate it and then compute the cosine between the concatenated vectors of the source and the target.",
"We suspect that the vectors will have similar behaviour between the source and the target given that they are trained under parallel Wikipedia articles.",
"We develop two ML models: 1) Edit distance scores and 2) Edit distance scores and distributional similarity score.",
"We use SVMlight (Joachims, 1998) Results and Discussion In this section we describe the data used to produce the n-best lists and train the cognate ranking models.",
"We evaluate the ranking models with heldout data from each training domain.",
"We also provide manual evaluation over the ranked n-best lists for error analysis.",
"Data The n-best lists to detect cognates were extracted from the respective Wikipedias by using the method described in Section 2.1.",
"The training data for the ranking model consists of different types of parallel corpora.",
"The parallel corpora are as follows: 1) Wiki-titles we use the inter language links to create a parallel corpus from the tittles of the Wikipedia articles, with about 500K aligned links (i.e.",
"'sentences') per language pair (about 200k for bg-ru), giving us about 200K training instances per language pair 3 , 2) Opensubs is an open source corpus of subtitles built by the fan community, with 1M sentences, 6M tokens, 100K words, giving about 90K training instances (Tiedemann, 2012) and 3) Zoo is a proprietary corpus of subtitles produced by professional translators, with 100K sentences, 700K tokens, 40K words and giving about 20K training instances per language pair.",
"Our objective is to create MT dictionaries from the produced n-best lists and we use parallel data as a source of training to prune them.",
"We are interested in the corpora of subtitles because the chosen domain of our SMT experiments is subtitling, while the proposed ranking method can be used in other application domains as well.",
"We consider Zoo and Opensubs as two different domains given that they were built by different types of translators and they differ in size and quality.",
"The heldout data consists of 2K instances for each corpus.",
"We use Wikipedia documents and Opensusbs subtitles to train word2vec for the distributional similarity feature.",
"We use the continuous bag-ofwords algorithm for word2vec and set the parameters for training to 200 dimensions and a window of 8 words.",
"The Wikipedia documents with an average number of 70K documents for each language, and Opensubs subtitles with 1M sentences for each language.",
"In practice we only use the Wikipedia data given that for Opensubs the model is able to find relatively few vectors, for example a vector is found only for 20% of the words in the pt-es pair.",
"Evaluation of the Ranking Model We define two ranking models as: model E for edit distance features and model EC for both edit Table 1 shows the results of the ranking procedure.",
"For the Romance family language pairs the model EC with context features consistently reduces the error compared to the solely use of edit distance metrics.",
"The only exception is the it-es EC model with poor results for the domain of Wiki-titles.",
"The models for the Slavonic family behave similarly to the Romance family, where the use of context features reduces the ranking error.",
"The exception is the bg-ru model on the Opensubs domain.",
"A possible reason for the poor results on the ites and bg-ru models is that the model often assigns a high similarity score to unrelated words.",
"For example, in it-es, mortes 'deads' is treated as close to categoria 'category'.",
"A possible solution is to map the vectors form the source side into the space of the target side via a learned transformation matrix (Mikolov et al., 2013b) .",
"Preliminary Results on Comparable Corpora After we extracted the n-best lists for the Romance family comparable corpora, we applied one of the ranking models on these lists and we manually evaluated over a sample of 50 words 4 .",
"We set n to 10 for the n-best lists.",
"We use a frequency window of 200 for the n-best list search heuristic and the domain of the comparable corpora to Wiki-titles 4 The sample consists of words with a frequency between 2K and 5. for the domain adaptation technique.",
"The purpose of manual evaluation is to decide whether the ML setup is sensible on the objective task.",
"Each list is evaluated by accuracy at 1 and accuracy at 10.",
"We also show success and failure examples of the ranking and the n-best lists.",
"Table 2 shows the preliminary results of the ML model E on a sample of Wikipedia dumps.",
"The L and L-R lists consistently show poor results.",
"A possible reason is the amount of errors given the first step to extract the n-best lists.",
"For example, in pt-es, for the word vivem 'live' the 10-best list only contain one word with a similar meaning viva 'living' but it can be also translated as 'cheers'.",
"In the pt-es list for the word representação 'description' the correct translation representación is not among the 10-best in the L list.",
"However, it is present in the 10-best for the L-C list and the ML model EC ranks it in the first place.",
"The edit distance model E still makes mistakes like with the list L-C, the word vivem 'live' translates into viven 'living' and the correct translation is vivir.",
"However, given a certain context/sense the previous translation can be correct.",
"The ranking scores given by the SVM varies from each list version.",
"For the L-C lists the scores are more uniform in increasing order and with a small variance.",
"The L and L-R lists show the opposite behaviour.",
"We add the produced Wikipedia n-best lists with the L metric into a SMT training dataset for the ptes pair.",
"We use the Moses SMT toolkit (Koehn et al., 2007) to test the augmented datasets.",
"We compare the augmented model with a baseline both trained by using the Zoo corpus of subtitles.",
"We use a 1-best list consisting of 100K pairs.",
"Te dataset used for pt-es baseline is: 80K training sentences, 1K sentences for tuning and 2K sen- Lang Pairs acc@1 acc@10 acc@1 acc@10 acc@1 acc@10 pt-es 20 60 22 59 32 70 it-es 16 53 18 45 44 66 fr-es 10 48 12 51 29 59 A possible reason for low improvement in terms of the BLEU scores is because MT evaluation metrics, such as BLEU, compare the MT output with a human reference.",
"The human reference translations in our corpus have been done from English (e.g., En→Es), while the test translations come from a related language (En→Pt→Es), often resulting in different paraphrases of the same English source.",
"While our OOV rate improved, the evaluation scores did not reflected this, because our MT output was still far from the reference even in cases it was otherwise acceptable.",
"List L List L-R List L-C Conclusions and future Work We have presented work in progress for developing MT dictionaries extracted from comparable resources for related languages.",
"The extraction heuristic show positive results on the n-best lists that group words with the same starting char-5 https://github.com/clab/fast_align 6 https://kheafield.com/code/kenlm/ 7 The p-value for the uk-ru pair is 0.06 we do not consider this result as statistically significant.",
"acters, because the used comparable corpora consist of related languages that share a similar orthography.",
"However, the lists based on the frequency window heuristic show poor results to include the correct translations during the extraction step.",
"Our ML models based on similarity metrics over parallel corpora show rankings similar to heldout data.",
"However, we created our training data using simple heuristics that simulate cognate words (i.e.",
"pairs of words with a small surface difference).",
"The ML models are able to rank similar words on the top of the list and they give a reliable score to discriminate wrong translations.",
"Preliminary results on the addition of the n-best lists into an SMT system show modest improvements compare to the baseline.",
"However, the OOV rate shows improvements around 10% reduction on word types, because of the wide variety of lexical choices introduced by the MT dictionaries.",
"Future work involves the addition of unsupervised morphology features for the n-best list extraction, i.e.",
"first step, given that the use of starting characters shows to be an effective heuristic to prune the search space and language independent.",
"Finally, we will measure the contribution for all the produced cognate lists, where we can try different strategies to add the dictionaries into an SMT system (Irvine and Callison-Burch, 2014) ."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"3",
"3.1",
"3.2",
"3.3",
"4"
],
"paper_header_content": [
"Introduction",
"Methodology",
"Cognate detection",
"Cognate Ranking",
"Results and Discussion",
"Data",
"Evaluation of the Ranking Model",
"Preliminary Results on Comparable Corpora",
"Conclusions and future Work"
]
} | GEM-SciDuet-train-8#paper-972#slide-2 | Similarity metrics | Compare words between frequency lists over comparable corpora
L matching between the languages using Levenshtein distance:
L-R Levenshtein distance computed separately for the roots and for the endings: aceito (pt) vs acepto (es) rejeito (pt) vs rechazo (es)
L-C Levenshtein distance over words with similar number of starting characters (i.e. prefix): introducao (pt) vs introduccion (es) introduziu (pt) vs introdujo (es) | Compare words between frequency lists over comparable corpora
L matching between the languages using Levenshtein distance:
L-R Levenshtein distance computed separately for the roots and for the endings: aceito (pt) vs acepto (es) rejeito (pt) vs rechazo (es)
L-C Levenshtein distance over words with similar number of starting characters (i.e. prefix): introducao (pt) vs introduccion (es) introduziu (pt) vs introdujo (es) | [] |
GEM-SciDuet-train-8#paper-972#slide-3 | 972 | Obtaining SMT dictionaries for related languages | This study explores methods for developing Machine Translation dictionaries on the basis of word frequency lists coming from comparable corpora. We investigate (1) various methods to measure the similarity of cognates between related languages, (2) detection and removal of noisy cognate translations using SVM ranking. We show preliminary results on several Romance and Slavonic languages. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119
],
"paper_content_text": [
"Introduction Cognates are words having similarities in their spelling and meaning in two languages, either because the two languages are typologically related, e.g., maladie vs malattia ('disease'), or because they were both borrowed from the same source (informatique vs informatica).",
"The advantage of their use in Statistical Machine Translation (SMT) is that the procedure can be based on comparable corpora, i.e., similar corpora which are not translations of each other (Sharoff et al., 2013) .",
"Given that there are more sources of comparable corpora in comparison to parallel ones, the lexicon obtained from them is likely to be richer and more variable.",
"Detection of cognates is a well-known task, which has been explored for a range of languages using different methods.",
"The two main approaches applied to detection of the cognates are the generative and discriminative paradigms.",
"The first one is based on detection of the edit distance between potential candidate pairs.",
"The distance can be a simple Levenshtein distance, or a distance measure with the scores learned from an existing parallel set (Tiedemann, 1999; Mann and Yarowsky, 2001) .",
"The discriminative paradigm uses standard approaches to machine learning, which are based on (1) extracting features, e.g., character n-grams, and (2) learning to predict the transformations of the source word needed to (Jiampojamarn et al., 2010; Frunza and Inkpen, 2009) .",
"Given that SMT is usually based on a full-form lexicon, one of the possible issues in generation of cognates concerns the similarity of words in their root form vs the similarity in endings.",
"For example, the Ukrainian wordform áëèaeíüîãî 'near gen ' is cognate to Russian áëèaeíåãî, the root is identical, while the ending is considerably different (üîãî vs åãî).",
"Regular differences in the endings, which are shared across a large number of words, can be learned separately from the regular differences in the roots.",
"One also needs to take into account the false friends among cognates.",
"For example, diseñar means 'to design' in Spanish vs desenhar in Portuguese means 'to draw'.",
"There are also often cases of partial cognates, when the words share the meaning in some contexts, but not in others, e.g., aeåíà in Russian means 'wife', while its Bulgarian cognate aeåíà has two meanings: 'wife' and 'woman'.",
"Yet another complexity concerns a frequency mismatch.",
"Two cognates might differ in their frequency.",
"For example, dibujo in Spanish ('a drawing', rank 1779 in the Wikipedia frequency list) corresponds to a relatively rare cognate word debuxo in Portuguese (rank 104,514 in Wikipedia), while another Portuguese word desenho is more commonly used in this sense (rank 884 in the Portuguese Wikipedia).",
"For MT tasks we need translations that are equally appropriate in the source and target language, therefore cognates useful for a high-quality dictionary for SMT need to have roughly the same frequency in comparable corpora and they need to be used in similar contexts.",
"This study investigates the settings for extracting cognates for related languages in Romance and Slavonic language families for the task of reducing the number of unknown words for SMT.",
"This in-cludes the effects of having constraints for the cognates to be similar in their roots and in the endings, to occur in distributionally similar contexts and to have similar frequency.",
"Methodology The methodology for producing the list of cognates is based on the following steps: 1) Produce several lists of cognates using a family of distance measures, discussed in Section 2.1 from comparable corpora, 2) Prune the candidate lists by ranking items, this is done using a Machine Learning (ML) algorithm trained over parallel corpora for detecting the outliers, discussed in Section 2.2; The initial frequency lists for alignment are based Wikipedia dumps for the following languages: Romance (French, Italian, Spanish, Portuguese) and Slavonic (Bulgarian, Russian, Ukrainian), where the target languages are Spanish and Russian 1 .",
"Cognate detection We extract possible lists of cognates from comparable corpora by using a family of similarity measures: L direct matching between the languages using Levenshtein distance (Levenshtein, 1966) ; L(w s , w t ) = 1 − ed(w s , w t ) L-R Levenshtein distance with weights computed separately for the roots and for the endings; LR(r s , r t , e s , e t ) = α×ed(rs,rt)+β×ed(es,et) α+β L-C Levenshtein distance over word with similar number of starting characters (i.e.",
"prefix); LC(c s , c t ) = 1 − ed(c s , c t ), same prefix 0, otherwise where ed(., .)",
"is the normalised Levenshtein distance in characters between the source word w s and the target word w t .",
"The r s and r t are the stems produced by the Snowball stemmer 2 .",
"Since the Snowball stemmer does not support Ukrainian and Bulgarian, we used the Russian model for making the stem/ending split.",
"e s , e t are the characters at the end of a word form given a stem and c s , c t are the first n characters of a word.",
"In this work, we set the weights α = 0.6 and β = 0.4 giving more importance to the roots.",
"We set a higher weight to roots on the L-R, which is language dependent, and compare to the L-C metric, which is language independent.",
"We transform the Levenshtein distances into similarity metrics by subtracting the normalised distance score from one.",
"The produced lists contain for each source word the possible n-best target words accordingly to the maximum scores with one of the previous measures.",
"The n-best list allows possible cognate translations to a given source word that share a part of the surface form.",
"Different from (Mann and Yarowsky, 2001) , we produce n-best cognate lists scored by edit distance instead of 1-best.",
"An important problem when comparing comparable corpora is the way of representing the search space, where an exhaustive method compares all the combinations of source and target words (Mann and Yarowsky, 2001) .",
"We constraint the search space by comparing each source word against the target words that belong to a frequency window around the frequency of the source word.",
"This constraint only applies for the L and L-R metrics.",
"We use Wikipedia dumps for the source and target side processed in the form frequency lists.",
"We order the target side list into bins of similar frequency and for the source side we filter words that appear only once.",
"We use the window approach given that the frequency between the corpora under study can not be directly comparable.",
"During testing we use a wide window of ±200 bins to minimise the loss of good candidate translations.",
"The second search space constraint heuristic is the L-C metric.",
"This metric only compares source words with the target words upto a given n prefix.",
"For c s , c t in L-C , we use the first four characters to compare groups of words as suggested in (Kondrak et al., 2003) .",
"Cognate Ranking Given that the n-best lists contain noise, we aim to prune them by an ML ranking model.",
"However, there is a lack of resources to train a classification model for cognates (i.e.",
"cognate vs. false friend), as mentioned in (Fišer and Ljubešić, 2013) .",
"Available data that can be used to judge the cognate lists are the alignment pairs extracted from parallel data.",
"We decide to use a ranking model to avoid data imbalance present in classification and to use the probability scores of the alignment pairs as ranks, as opposed to the classification model used by (Irvine and Callison-Burch, 2013) .",
"Moreover, we also use a popular domain adaptation technique (Daumé et al., 2010) given that we have access to different domains of parallel training data that might be compatible with our comparable corpora.",
"The training data are the alignments between pairs of words where we rank them accordingly to their correspondent alignment probability from the output of GIZA++ (Och and Ney, 2003) .",
"We then use a heuristic to prune training data in order to simulate cognate words.",
"Pairs of words scored below the Levenshtein similarity threshold of 0.5 are not considered as cognates given that they are likely to have a different surface form.",
"We represent the training and test data with features extracted from different edit distance scores and distributional measures.",
"The edit distances features are as follows: 1) Similarity measure L and 2) Number of times of each edit operation.",
"Thus, the model assigns a different importance to each operation.",
"The distributional feature is based on the cosine between the distributional vectors of a window of n words around the word currently under comparison.",
"We train distributional similarity models with word2vec (Mikolov et al., 2013a) for the source and target side separately.",
"We extract the continuous vector for each word in the window, concatenate it and then compute the cosine between the concatenated vectors of the source and the target.",
"We suspect that the vectors will have similar behaviour between the source and the target given that they are trained under parallel Wikipedia articles.",
"We develop two ML models: 1) Edit distance scores and 2) Edit distance scores and distributional similarity score.",
"We use SVMlight (Joachims, 1998) Results and Discussion In this section we describe the data used to produce the n-best lists and train the cognate ranking models.",
"We evaluate the ranking models with heldout data from each training domain.",
"We also provide manual evaluation over the ranked n-best lists for error analysis.",
"Data The n-best lists to detect cognates were extracted from the respective Wikipedias by using the method described in Section 2.1.",
"The training data for the ranking model consists of different types of parallel corpora.",
"The parallel corpora are as follows: 1) Wiki-titles we use the inter language links to create a parallel corpus from the tittles of the Wikipedia articles, with about 500K aligned links (i.e.",
"'sentences') per language pair (about 200k for bg-ru), giving us about 200K training instances per language pair 3 , 2) Opensubs is an open source corpus of subtitles built by the fan community, with 1M sentences, 6M tokens, 100K words, giving about 90K training instances (Tiedemann, 2012) and 3) Zoo is a proprietary corpus of subtitles produced by professional translators, with 100K sentences, 700K tokens, 40K words and giving about 20K training instances per language pair.",
"Our objective is to create MT dictionaries from the produced n-best lists and we use parallel data as a source of training to prune them.",
"We are interested in the corpora of subtitles because the chosen domain of our SMT experiments is subtitling, while the proposed ranking method can be used in other application domains as well.",
"We consider Zoo and Opensubs as two different domains given that they were built by different types of translators and they differ in size and quality.",
"The heldout data consists of 2K instances for each corpus.",
"We use Wikipedia documents and Opensusbs subtitles to train word2vec for the distributional similarity feature.",
"We use the continuous bag-ofwords algorithm for word2vec and set the parameters for training to 200 dimensions and a window of 8 words.",
"The Wikipedia documents with an average number of 70K documents for each language, and Opensubs subtitles with 1M sentences for each language.",
"In practice we only use the Wikipedia data given that for Opensubs the model is able to find relatively few vectors, for example a vector is found only for 20% of the words in the pt-es pair.",
"Evaluation of the Ranking Model We define two ranking models as: model E for edit distance features and model EC for both edit Table 1 shows the results of the ranking procedure.",
"For the Romance family language pairs the model EC with context features consistently reduces the error compared to the solely use of edit distance metrics.",
"The only exception is the it-es EC model with poor results for the domain of Wiki-titles.",
"The models for the Slavonic family behave similarly to the Romance family, where the use of context features reduces the ranking error.",
"The exception is the bg-ru model on the Opensubs domain.",
"A possible reason for the poor results on the ites and bg-ru models is that the model often assigns a high similarity score to unrelated words.",
"For example, in it-es, mortes 'deads' is treated as close to categoria 'category'.",
"A possible solution is to map the vectors form the source side into the space of the target side via a learned transformation matrix (Mikolov et al., 2013b) .",
"Preliminary Results on Comparable Corpora After we extracted the n-best lists for the Romance family comparable corpora, we applied one of the ranking models on these lists and we manually evaluated over a sample of 50 words 4 .",
"We set n to 10 for the n-best lists.",
"We use a frequency window of 200 for the n-best list search heuristic and the domain of the comparable corpora to Wiki-titles 4 The sample consists of words with a frequency between 2K and 5. for the domain adaptation technique.",
"The purpose of manual evaluation is to decide whether the ML setup is sensible on the objective task.",
"Each list is evaluated by accuracy at 1 and accuracy at 10.",
"We also show success and failure examples of the ranking and the n-best lists.",
"Table 2 shows the preliminary results of the ML model E on a sample of Wikipedia dumps.",
"The L and L-R lists consistently show poor results.",
"A possible reason is the amount of errors given the first step to extract the n-best lists.",
"For example, in pt-es, for the word vivem 'live' the 10-best list only contain one word with a similar meaning viva 'living' but it can be also translated as 'cheers'.",
"In the pt-es list for the word representação 'description' the correct translation representación is not among the 10-best in the L list.",
"However, it is present in the 10-best for the L-C list and the ML model EC ranks it in the first place.",
"The edit distance model E still makes mistakes like with the list L-C, the word vivem 'live' translates into viven 'living' and the correct translation is vivir.",
"However, given a certain context/sense the previous translation can be correct.",
"The ranking scores given by the SVM varies from each list version.",
"For the L-C lists the scores are more uniform in increasing order and with a small variance.",
"The L and L-R lists show the opposite behaviour.",
"We add the produced Wikipedia n-best lists with the L metric into a SMT training dataset for the ptes pair.",
"We use the Moses SMT toolkit (Koehn et al., 2007) to test the augmented datasets.",
"We compare the augmented model with a baseline both trained by using the Zoo corpus of subtitles.",
"We use a 1-best list consisting of 100K pairs.",
"Te dataset used for pt-es baseline is: 80K training sentences, 1K sentences for tuning and 2K sen- Lang Pairs acc@1 acc@10 acc@1 acc@10 acc@1 acc@10 pt-es 20 60 22 59 32 70 it-es 16 53 18 45 44 66 fr-es 10 48 12 51 29 59 A possible reason for low improvement in terms of the BLEU scores is because MT evaluation metrics, such as BLEU, compare the MT output with a human reference.",
"The human reference translations in our corpus have been done from English (e.g., En→Es), while the test translations come from a related language (En→Pt→Es), often resulting in different paraphrases of the same English source.",
"While our OOV rate improved, the evaluation scores did not reflected this, because our MT output was still far from the reference even in cases it was otherwise acceptable.",
"List L List L-R List L-C Conclusions and future Work We have presented work in progress for developing MT dictionaries extracted from comparable resources for related languages.",
"The extraction heuristic show positive results on the n-best lists that group words with the same starting char-5 https://github.com/clab/fast_align 6 https://kheafield.com/code/kenlm/ 7 The p-value for the uk-ru pair is 0.06 we do not consider this result as statistically significant.",
"acters, because the used comparable corpora consist of related languages that share a similar orthography.",
"However, the lists based on the frequency window heuristic show poor results to include the correct translations during the extraction step.",
"Our ML models based on similarity metrics over parallel corpora show rankings similar to heldout data.",
"However, we created our training data using simple heuristics that simulate cognate words (i.e.",
"pairs of words with a small surface difference).",
"The ML models are able to rank similar words on the top of the list and they give a reliable score to discriminate wrong translations.",
"Preliminary results on the addition of the n-best lists into an SMT system show modest improvements compare to the baseline.",
"However, the OOV rate shows improvements around 10% reduction on word types, because of the wide variety of lexical choices introduced by the MT dictionaries.",
"Future work involves the addition of unsupervised morphology features for the n-best list extraction, i.e.",
"first step, given that the use of starting characters shows to be an effective heuristic to prune the search space and language independent.",
"Finally, we will measure the contribution for all the produced cognate lists, where we can try different strategies to add the dictionaries into an SMT system (Irvine and Callison-Burch, 2014) ."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"3",
"3.1",
"3.2",
"3.3",
"4"
],
"paper_header_content": [
"Introduction",
"Methodology",
"Cognate detection",
"Cognate Ranking",
"Results and Discussion",
"Data",
"Evaluation of the Ranking Model",
"Preliminary Results on Comparable Corpora",
"Conclusions and future Work"
]
} | GEM-SciDuet-train-8#paper-972#slide-3 | Search space constraints | Motivation Exhaustive method compares all the combinations of source and target words
Order the target side frequency list into bins of similar frequency
Compare each source word with target bins of similar frequency around a window
L-C metric only compares words that share a given n prefix | Motivation Exhaustive method compares all the combinations of source and target words
Order the target side frequency list into bins of similar frequency
Compare each source word with target bins of similar frequency around a window
L-C metric only compares words that share a given n prefix | [] |
GEM-SciDuet-train-8#paper-972#slide-4 | 972 | Obtaining SMT dictionaries for related languages | This study explores methods for developing Machine Translation dictionaries on the basis of word frequency lists coming from comparable corpora. We investigate (1) various methods to measure the similarity of cognates between related languages, (2) detection and removal of noisy cognate translations using SVM ranking. We show preliminary results on several Romance and Slavonic languages. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119
],
"paper_content_text": [
"Introduction Cognates are words having similarities in their spelling and meaning in two languages, either because the two languages are typologically related, e.g., maladie vs malattia ('disease'), or because they were both borrowed from the same source (informatique vs informatica).",
"The advantage of their use in Statistical Machine Translation (SMT) is that the procedure can be based on comparable corpora, i.e., similar corpora which are not translations of each other (Sharoff et al., 2013) .",
"Given that there are more sources of comparable corpora in comparison to parallel ones, the lexicon obtained from them is likely to be richer and more variable.",
"Detection of cognates is a well-known task, which has been explored for a range of languages using different methods.",
"The two main approaches applied to detection of the cognates are the generative and discriminative paradigms.",
"The first one is based on detection of the edit distance between potential candidate pairs.",
"The distance can be a simple Levenshtein distance, or a distance measure with the scores learned from an existing parallel set (Tiedemann, 1999; Mann and Yarowsky, 2001) .",
"The discriminative paradigm uses standard approaches to machine learning, which are based on (1) extracting features, e.g., character n-grams, and (2) learning to predict the transformations of the source word needed to (Jiampojamarn et al., 2010; Frunza and Inkpen, 2009) .",
"Given that SMT is usually based on a full-form lexicon, one of the possible issues in generation of cognates concerns the similarity of words in their root form vs the similarity in endings.",
"For example, the Ukrainian wordform áëèaeíüîãî 'near gen ' is cognate to Russian áëèaeíåãî, the root is identical, while the ending is considerably different (üîãî vs åãî).",
"Regular differences in the endings, which are shared across a large number of words, can be learned separately from the regular differences in the roots.",
"One also needs to take into account the false friends among cognates.",
"For example, diseñar means 'to design' in Spanish vs desenhar in Portuguese means 'to draw'.",
"There are also often cases of partial cognates, when the words share the meaning in some contexts, but not in others, e.g., aeåíà in Russian means 'wife', while its Bulgarian cognate aeåíà has two meanings: 'wife' and 'woman'.",
"Yet another complexity concerns a frequency mismatch.",
"Two cognates might differ in their frequency.",
"For example, dibujo in Spanish ('a drawing', rank 1779 in the Wikipedia frequency list) corresponds to a relatively rare cognate word debuxo in Portuguese (rank 104,514 in Wikipedia), while another Portuguese word desenho is more commonly used in this sense (rank 884 in the Portuguese Wikipedia).",
"For MT tasks we need translations that are equally appropriate in the source and target language, therefore cognates useful for a high-quality dictionary for SMT need to have roughly the same frequency in comparable corpora and they need to be used in similar contexts.",
"This study investigates the settings for extracting cognates for related languages in Romance and Slavonic language families for the task of reducing the number of unknown words for SMT.",
"This in-cludes the effects of having constraints for the cognates to be similar in their roots and in the endings, to occur in distributionally similar contexts and to have similar frequency.",
"Methodology The methodology for producing the list of cognates is based on the following steps: 1) Produce several lists of cognates using a family of distance measures, discussed in Section 2.1 from comparable corpora, 2) Prune the candidate lists by ranking items, this is done using a Machine Learning (ML) algorithm trained over parallel corpora for detecting the outliers, discussed in Section 2.2; The initial frequency lists for alignment are based Wikipedia dumps for the following languages: Romance (French, Italian, Spanish, Portuguese) and Slavonic (Bulgarian, Russian, Ukrainian), where the target languages are Spanish and Russian 1 .",
"Cognate detection We extract possible lists of cognates from comparable corpora by using a family of similarity measures: L direct matching between the languages using Levenshtein distance (Levenshtein, 1966) ; L(w s , w t ) = 1 − ed(w s , w t ) L-R Levenshtein distance with weights computed separately for the roots and for the endings; LR(r s , r t , e s , e t ) = α×ed(rs,rt)+β×ed(es,et) α+β L-C Levenshtein distance over word with similar number of starting characters (i.e.",
"prefix); LC(c s , c t ) = 1 − ed(c s , c t ), same prefix 0, otherwise where ed(., .)",
"is the normalised Levenshtein distance in characters between the source word w s and the target word w t .",
"The r s and r t are the stems produced by the Snowball stemmer 2 .",
"Since the Snowball stemmer does not support Ukrainian and Bulgarian, we used the Russian model for making the stem/ending split.",
"e s , e t are the characters at the end of a word form given a stem and c s , c t are the first n characters of a word.",
"In this work, we set the weights α = 0.6 and β = 0.4 giving more importance to the roots.",
"We set a higher weight to roots on the L-R, which is language dependent, and compare to the L-C metric, which is language independent.",
"We transform the Levenshtein distances into similarity metrics by subtracting the normalised distance score from one.",
"The produced lists contain for each source word the possible n-best target words accordingly to the maximum scores with one of the previous measures.",
"The n-best list allows possible cognate translations to a given source word that share a part of the surface form.",
"Different from (Mann and Yarowsky, 2001) , we produce n-best cognate lists scored by edit distance instead of 1-best.",
"An important problem when comparing comparable corpora is the way of representing the search space, where an exhaustive method compares all the combinations of source and target words (Mann and Yarowsky, 2001) .",
"We constraint the search space by comparing each source word against the target words that belong to a frequency window around the frequency of the source word.",
"This constraint only applies for the L and L-R metrics.",
"We use Wikipedia dumps for the source and target side processed in the form frequency lists.",
"We order the target side list into bins of similar frequency and for the source side we filter words that appear only once.",
"We use the window approach given that the frequency between the corpora under study can not be directly comparable.",
"During testing we use a wide window of ±200 bins to minimise the loss of good candidate translations.",
"The second search space constraint heuristic is the L-C metric.",
"This metric only compares source words with the target words upto a given n prefix.",
"For c s , c t in L-C , we use the first four characters to compare groups of words as suggested in (Kondrak et al., 2003) .",
"Cognate Ranking Given that the n-best lists contain noise, we aim to prune them by an ML ranking model.",
"However, there is a lack of resources to train a classification model for cognates (i.e.",
"cognate vs. false friend), as mentioned in (Fišer and Ljubešić, 2013) .",
"Available data that can be used to judge the cognate lists are the alignment pairs extracted from parallel data.",
"We decide to use a ranking model to avoid data imbalance present in classification and to use the probability scores of the alignment pairs as ranks, as opposed to the classification model used by (Irvine and Callison-Burch, 2013) .",
"Moreover, we also use a popular domain adaptation technique (Daumé et al., 2010) given that we have access to different domains of parallel training data that might be compatible with our comparable corpora.",
"The training data are the alignments between pairs of words where we rank them accordingly to their correspondent alignment probability from the output of GIZA++ (Och and Ney, 2003) .",
"We then use a heuristic to prune training data in order to simulate cognate words.",
"Pairs of words scored below the Levenshtein similarity threshold of 0.5 are not considered as cognates given that they are likely to have a different surface form.",
"We represent the training and test data with features extracted from different edit distance scores and distributional measures.",
"The edit distances features are as follows: 1) Similarity measure L and 2) Number of times of each edit operation.",
"Thus, the model assigns a different importance to each operation.",
"The distributional feature is based on the cosine between the distributional vectors of a window of n words around the word currently under comparison.",
"We train distributional similarity models with word2vec (Mikolov et al., 2013a) for the source and target side separately.",
"We extract the continuous vector for each word in the window, concatenate it and then compute the cosine between the concatenated vectors of the source and the target.",
"We suspect that the vectors will have similar behaviour between the source and the target given that they are trained under parallel Wikipedia articles.",
"We develop two ML models: 1) Edit distance scores and 2) Edit distance scores and distributional similarity score.",
"We use SVMlight (Joachims, 1998) Results and Discussion In this section we describe the data used to produce the n-best lists and train the cognate ranking models.",
"We evaluate the ranking models with heldout data from each training domain.",
"We also provide manual evaluation over the ranked n-best lists for error analysis.",
"Data The n-best lists to detect cognates were extracted from the respective Wikipedias by using the method described in Section 2.1.",
"The training data for the ranking model consists of different types of parallel corpora.",
"The parallel corpora are as follows: 1) Wiki-titles we use the inter language links to create a parallel corpus from the tittles of the Wikipedia articles, with about 500K aligned links (i.e.",
"'sentences') per language pair (about 200k for bg-ru), giving us about 200K training instances per language pair 3 , 2) Opensubs is an open source corpus of subtitles built by the fan community, with 1M sentences, 6M tokens, 100K words, giving about 90K training instances (Tiedemann, 2012) and 3) Zoo is a proprietary corpus of subtitles produced by professional translators, with 100K sentences, 700K tokens, 40K words and giving about 20K training instances per language pair.",
"Our objective is to create MT dictionaries from the produced n-best lists and we use parallel data as a source of training to prune them.",
"We are interested in the corpora of subtitles because the chosen domain of our SMT experiments is subtitling, while the proposed ranking method can be used in other application domains as well.",
"We consider Zoo and Opensubs as two different domains given that they were built by different types of translators and they differ in size and quality.",
"The heldout data consists of 2K instances for each corpus.",
"We use Wikipedia documents and Opensusbs subtitles to train word2vec for the distributional similarity feature.",
"We use the continuous bag-ofwords algorithm for word2vec and set the parameters for training to 200 dimensions and a window of 8 words.",
"The Wikipedia documents with an average number of 70K documents for each language, and Opensubs subtitles with 1M sentences for each language.",
"In practice we only use the Wikipedia data given that for Opensubs the model is able to find relatively few vectors, for example a vector is found only for 20% of the words in the pt-es pair.",
"Evaluation of the Ranking Model We define two ranking models as: model E for edit distance features and model EC for both edit Table 1 shows the results of the ranking procedure.",
"For the Romance family language pairs the model EC with context features consistently reduces the error compared to the solely use of edit distance metrics.",
"The only exception is the it-es EC model with poor results for the domain of Wiki-titles.",
"The models for the Slavonic family behave similarly to the Romance family, where the use of context features reduces the ranking error.",
"The exception is the bg-ru model on the Opensubs domain.",
"A possible reason for the poor results on the ites and bg-ru models is that the model often assigns a high similarity score to unrelated words.",
"For example, in it-es, mortes 'deads' is treated as close to categoria 'category'.",
"A possible solution is to map the vectors form the source side into the space of the target side via a learned transformation matrix (Mikolov et al., 2013b) .",
"Preliminary Results on Comparable Corpora After we extracted the n-best lists for the Romance family comparable corpora, we applied one of the ranking models on these lists and we manually evaluated over a sample of 50 words 4 .",
"We set n to 10 for the n-best lists.",
"We use a frequency window of 200 for the n-best list search heuristic and the domain of the comparable corpora to Wiki-titles 4 The sample consists of words with a frequency between 2K and 5. for the domain adaptation technique.",
"The purpose of manual evaluation is to decide whether the ML setup is sensible on the objective task.",
"Each list is evaluated by accuracy at 1 and accuracy at 10.",
"We also show success and failure examples of the ranking and the n-best lists.",
"Table 2 shows the preliminary results of the ML model E on a sample of Wikipedia dumps.",
"The L and L-R lists consistently show poor results.",
"A possible reason is the amount of errors given the first step to extract the n-best lists.",
"For example, in pt-es, for the word vivem 'live' the 10-best list only contain one word with a similar meaning viva 'living' but it can be also translated as 'cheers'.",
"In the pt-es list for the word representação 'description' the correct translation representación is not among the 10-best in the L list.",
"However, it is present in the 10-best for the L-C list and the ML model EC ranks it in the first place.",
"The edit distance model E still makes mistakes like with the list L-C, the word vivem 'live' translates into viven 'living' and the correct translation is vivir.",
"However, given a certain context/sense the previous translation can be correct.",
"The ranking scores given by the SVM varies from each list version.",
"For the L-C lists the scores are more uniform in increasing order and with a small variance.",
"The L and L-R lists show the opposite behaviour.",
"We add the produced Wikipedia n-best lists with the L metric into a SMT training dataset for the ptes pair.",
"We use the Moses SMT toolkit (Koehn et al., 2007) to test the augmented datasets.",
"We compare the augmented model with a baseline both trained by using the Zoo corpus of subtitles.",
"We use a 1-best list consisting of 100K pairs.",
"Te dataset used for pt-es baseline is: 80K training sentences, 1K sentences for tuning and 2K sen- Lang Pairs acc@1 acc@10 acc@1 acc@10 acc@1 acc@10 pt-es 20 60 22 59 32 70 it-es 16 53 18 45 44 66 fr-es 10 48 12 51 29 59 A possible reason for low improvement in terms of the BLEU scores is because MT evaluation metrics, such as BLEU, compare the MT output with a human reference.",
"The human reference translations in our corpus have been done from English (e.g., En→Es), while the test translations come from a related language (En→Pt→Es), often resulting in different paraphrases of the same English source.",
"While our OOV rate improved, the evaluation scores did not reflected this, because our MT output was still far from the reference even in cases it was otherwise acceptable.",
"List L List L-R List L-C Conclusions and future Work We have presented work in progress for developing MT dictionaries extracted from comparable resources for related languages.",
"The extraction heuristic show positive results on the n-best lists that group words with the same starting char-5 https://github.com/clab/fast_align 6 https://kheafield.com/code/kenlm/ 7 The p-value for the uk-ru pair is 0.06 we do not consider this result as statistically significant.",
"acters, because the used comparable corpora consist of related languages that share a similar orthography.",
"However, the lists based on the frequency window heuristic show poor results to include the correct translations during the extraction step.",
"Our ML models based on similarity metrics over parallel corpora show rankings similar to heldout data.",
"However, we created our training data using simple heuristics that simulate cognate words (i.e.",
"pairs of words with a small surface difference).",
"The ML models are able to rank similar words on the top of the list and they give a reliable score to discriminate wrong translations.",
"Preliminary results on the addition of the n-best lists into an SMT system show modest improvements compare to the baseline.",
"However, the OOV rate shows improvements around 10% reduction on word types, because of the wide variety of lexical choices introduced by the MT dictionaries.",
"Future work involves the addition of unsupervised morphology features for the n-best list extraction, i.e.",
"first step, given that the use of starting characters shows to be an effective heuristic to prune the search space and language independent.",
"Finally, we will measure the contribution for all the produced cognate lists, where we can try different strategies to add the dictionaries into an SMT system (Irvine and Callison-Burch, 2014) ."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"3",
"3.1",
"3.2",
"3.3",
"4"
],
"paper_header_content": [
"Introduction",
"Methodology",
"Cognate detection",
"Cognate Ranking",
"Results and Discussion",
"Data",
"Evaluation of the Ranking Model",
"Preliminary Results on Comparable Corpora",
"Conclusions and future Work"
]
} | GEM-SciDuet-train-8#paper-972#slide-4 | Ranking | Motivation Prune n-best lists by ranking ML algorithm
Training data come from aligned parallel corpora where the rank is given by the alignment probability from GIZA++
Simulate cognate training data by pruning pairs of words below a Levenshtein threshold | Motivation Prune n-best lists by ranking ML algorithm
Training data come from aligned parallel corpora where the rank is given by the alignment probability from GIZA++
Simulate cognate training data by pruning pairs of words below a Levenshtein threshold | [] |
GEM-SciDuet-train-8#paper-972#slide-5 | 972 | Obtaining SMT dictionaries for related languages | This study explores methods for developing Machine Translation dictionaries on the basis of word frequency lists coming from comparable corpora. We investigate (1) various methods to measure the similarity of cognates between related languages, (2) detection and removal of noisy cognate translations using SVM ranking. We show preliminary results on several Romance and Slavonic languages. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119
],
"paper_content_text": [
"Introduction Cognates are words having similarities in their spelling and meaning in two languages, either because the two languages are typologically related, e.g., maladie vs malattia ('disease'), or because they were both borrowed from the same source (informatique vs informatica).",
"The advantage of their use in Statistical Machine Translation (SMT) is that the procedure can be based on comparable corpora, i.e., similar corpora which are not translations of each other (Sharoff et al., 2013) .",
"Given that there are more sources of comparable corpora in comparison to parallel ones, the lexicon obtained from them is likely to be richer and more variable.",
"Detection of cognates is a well-known task, which has been explored for a range of languages using different methods.",
"The two main approaches applied to detection of the cognates are the generative and discriminative paradigms.",
"The first one is based on detection of the edit distance between potential candidate pairs.",
"The distance can be a simple Levenshtein distance, or a distance measure with the scores learned from an existing parallel set (Tiedemann, 1999; Mann and Yarowsky, 2001) .",
"The discriminative paradigm uses standard approaches to machine learning, which are based on (1) extracting features, e.g., character n-grams, and (2) learning to predict the transformations of the source word needed to (Jiampojamarn et al., 2010; Frunza and Inkpen, 2009) .",
"Given that SMT is usually based on a full-form lexicon, one of the possible issues in generation of cognates concerns the similarity of words in their root form vs the similarity in endings.",
"For example, the Ukrainian wordform áëèaeíüîãî 'near gen ' is cognate to Russian áëèaeíåãî, the root is identical, while the ending is considerably different (üîãî vs åãî).",
"Regular differences in the endings, which are shared across a large number of words, can be learned separately from the regular differences in the roots.",
"One also needs to take into account the false friends among cognates.",
"For example, diseñar means 'to design' in Spanish vs desenhar in Portuguese means 'to draw'.",
"There are also often cases of partial cognates, when the words share the meaning in some contexts, but not in others, e.g., aeåíà in Russian means 'wife', while its Bulgarian cognate aeåíà has two meanings: 'wife' and 'woman'.",
"Yet another complexity concerns a frequency mismatch.",
"Two cognates might differ in their frequency.",
"For example, dibujo in Spanish ('a drawing', rank 1779 in the Wikipedia frequency list) corresponds to a relatively rare cognate word debuxo in Portuguese (rank 104,514 in Wikipedia), while another Portuguese word desenho is more commonly used in this sense (rank 884 in the Portuguese Wikipedia).",
"For MT tasks we need translations that are equally appropriate in the source and target language, therefore cognates useful for a high-quality dictionary for SMT need to have roughly the same frequency in comparable corpora and they need to be used in similar contexts.",
"This study investigates the settings for extracting cognates for related languages in Romance and Slavonic language families for the task of reducing the number of unknown words for SMT.",
"This in-cludes the effects of having constraints for the cognates to be similar in their roots and in the endings, to occur in distributionally similar contexts and to have similar frequency.",
"Methodology The methodology for producing the list of cognates is based on the following steps: 1) Produce several lists of cognates using a family of distance measures, discussed in Section 2.1 from comparable corpora, 2) Prune the candidate lists by ranking items, this is done using a Machine Learning (ML) algorithm trained over parallel corpora for detecting the outliers, discussed in Section 2.2; The initial frequency lists for alignment are based Wikipedia dumps for the following languages: Romance (French, Italian, Spanish, Portuguese) and Slavonic (Bulgarian, Russian, Ukrainian), where the target languages are Spanish and Russian 1 .",
"Cognate detection We extract possible lists of cognates from comparable corpora by using a family of similarity measures: L direct matching between the languages using Levenshtein distance (Levenshtein, 1966) ; L(w s , w t ) = 1 − ed(w s , w t ) L-R Levenshtein distance with weights computed separately for the roots and for the endings; LR(r s , r t , e s , e t ) = α×ed(rs,rt)+β×ed(es,et) α+β L-C Levenshtein distance over word with similar number of starting characters (i.e.",
"prefix); LC(c s , c t ) = 1 − ed(c s , c t ), same prefix 0, otherwise where ed(., .)",
"is the normalised Levenshtein distance in characters between the source word w s and the target word w t .",
"The r s and r t are the stems produced by the Snowball stemmer 2 .",
"Since the Snowball stemmer does not support Ukrainian and Bulgarian, we used the Russian model for making the stem/ending split.",
"e s , e t are the characters at the end of a word form given a stem and c s , c t are the first n characters of a word.",
"In this work, we set the weights α = 0.6 and β = 0.4 giving more importance to the roots.",
"We set a higher weight to roots on the L-R, which is language dependent, and compare to the L-C metric, which is language independent.",
"We transform the Levenshtein distances into similarity metrics by subtracting the normalised distance score from one.",
"The produced lists contain for each source word the possible n-best target words accordingly to the maximum scores with one of the previous measures.",
"The n-best list allows possible cognate translations to a given source word that share a part of the surface form.",
"Different from (Mann and Yarowsky, 2001) , we produce n-best cognate lists scored by edit distance instead of 1-best.",
"An important problem when comparing comparable corpora is the way of representing the search space, where an exhaustive method compares all the combinations of source and target words (Mann and Yarowsky, 2001) .",
"We constraint the search space by comparing each source word against the target words that belong to a frequency window around the frequency of the source word.",
"This constraint only applies for the L and L-R metrics.",
"We use Wikipedia dumps for the source and target side processed in the form frequency lists.",
"We order the target side list into bins of similar frequency and for the source side we filter words that appear only once.",
"We use the window approach given that the frequency between the corpora under study can not be directly comparable.",
"During testing we use a wide window of ±200 bins to minimise the loss of good candidate translations.",
"The second search space constraint heuristic is the L-C metric.",
"This metric only compares source words with the target words upto a given n prefix.",
"For c s , c t in L-C , we use the first four characters to compare groups of words as suggested in (Kondrak et al., 2003) .",
"Cognate Ranking Given that the n-best lists contain noise, we aim to prune them by an ML ranking model.",
"However, there is a lack of resources to train a classification model for cognates (i.e.",
"cognate vs. false friend), as mentioned in (Fišer and Ljubešić, 2013) .",
"Available data that can be used to judge the cognate lists are the alignment pairs extracted from parallel data.",
"We decide to use a ranking model to avoid data imbalance present in classification and to use the probability scores of the alignment pairs as ranks, as opposed to the classification model used by (Irvine and Callison-Burch, 2013) .",
"Moreover, we also use a popular domain adaptation technique (Daumé et al., 2010) given that we have access to different domains of parallel training data that might be compatible with our comparable corpora.",
"The training data are the alignments between pairs of words where we rank them accordingly to their correspondent alignment probability from the output of GIZA++ (Och and Ney, 2003) .",
"We then use a heuristic to prune training data in order to simulate cognate words.",
"Pairs of words scored below the Levenshtein similarity threshold of 0.5 are not considered as cognates given that they are likely to have a different surface form.",
"We represent the training and test data with features extracted from different edit distance scores and distributional measures.",
"The edit distances features are as follows: 1) Similarity measure L and 2) Number of times of each edit operation.",
"Thus, the model assigns a different importance to each operation.",
"The distributional feature is based on the cosine between the distributional vectors of a window of n words around the word currently under comparison.",
"We train distributional similarity models with word2vec (Mikolov et al., 2013a) for the source and target side separately.",
"We extract the continuous vector for each word in the window, concatenate it and then compute the cosine between the concatenated vectors of the source and the target.",
"We suspect that the vectors will have similar behaviour between the source and the target given that they are trained under parallel Wikipedia articles.",
"We develop two ML models: 1) Edit distance scores and 2) Edit distance scores and distributional similarity score.",
"We use SVMlight (Joachims, 1998) Results and Discussion In this section we describe the data used to produce the n-best lists and train the cognate ranking models.",
"We evaluate the ranking models with heldout data from each training domain.",
"We also provide manual evaluation over the ranked n-best lists for error analysis.",
"Data The n-best lists to detect cognates were extracted from the respective Wikipedias by using the method described in Section 2.1.",
"The training data for the ranking model consists of different types of parallel corpora.",
"The parallel corpora are as follows: 1) Wiki-titles we use the inter language links to create a parallel corpus from the tittles of the Wikipedia articles, with about 500K aligned links (i.e.",
"'sentences') per language pair (about 200k for bg-ru), giving us about 200K training instances per language pair 3 , 2) Opensubs is an open source corpus of subtitles built by the fan community, with 1M sentences, 6M tokens, 100K words, giving about 90K training instances (Tiedemann, 2012) and 3) Zoo is a proprietary corpus of subtitles produced by professional translators, with 100K sentences, 700K tokens, 40K words and giving about 20K training instances per language pair.",
"Our objective is to create MT dictionaries from the produced n-best lists and we use parallel data as a source of training to prune them.",
"We are interested in the corpora of subtitles because the chosen domain of our SMT experiments is subtitling, while the proposed ranking method can be used in other application domains as well.",
"We consider Zoo and Opensubs as two different domains given that they were built by different types of translators and they differ in size and quality.",
"The heldout data consists of 2K instances for each corpus.",
"We use Wikipedia documents and Opensusbs subtitles to train word2vec for the distributional similarity feature.",
"We use the continuous bag-ofwords algorithm for word2vec and set the parameters for training to 200 dimensions and a window of 8 words.",
"The Wikipedia documents with an average number of 70K documents for each language, and Opensubs subtitles with 1M sentences for each language.",
"In practice we only use the Wikipedia data given that for Opensubs the model is able to find relatively few vectors, for example a vector is found only for 20% of the words in the pt-es pair.",
"Evaluation of the Ranking Model We define two ranking models as: model E for edit distance features and model EC for both edit Table 1 shows the results of the ranking procedure.",
"For the Romance family language pairs the model EC with context features consistently reduces the error compared to the solely use of edit distance metrics.",
"The only exception is the it-es EC model with poor results for the domain of Wiki-titles.",
"The models for the Slavonic family behave similarly to the Romance family, where the use of context features reduces the ranking error.",
"The exception is the bg-ru model on the Opensubs domain.",
"A possible reason for the poor results on the ites and bg-ru models is that the model often assigns a high similarity score to unrelated words.",
"For example, in it-es, mortes 'deads' is treated as close to categoria 'category'.",
"A possible solution is to map the vectors form the source side into the space of the target side via a learned transformation matrix (Mikolov et al., 2013b) .",
"Preliminary Results on Comparable Corpora After we extracted the n-best lists for the Romance family comparable corpora, we applied one of the ranking models on these lists and we manually evaluated over a sample of 50 words 4 .",
"We set n to 10 for the n-best lists.",
"We use a frequency window of 200 for the n-best list search heuristic and the domain of the comparable corpora to Wiki-titles 4 The sample consists of words with a frequency between 2K and 5. for the domain adaptation technique.",
"The purpose of manual evaluation is to decide whether the ML setup is sensible on the objective task.",
"Each list is evaluated by accuracy at 1 and accuracy at 10.",
"We also show success and failure examples of the ranking and the n-best lists.",
"Table 2 shows the preliminary results of the ML model E on a sample of Wikipedia dumps.",
"The L and L-R lists consistently show poor results.",
"A possible reason is the amount of errors given the first step to extract the n-best lists.",
"For example, in pt-es, for the word vivem 'live' the 10-best list only contain one word with a similar meaning viva 'living' but it can be also translated as 'cheers'.",
"In the pt-es list for the word representação 'description' the correct translation representación is not among the 10-best in the L list.",
"However, it is present in the 10-best for the L-C list and the ML model EC ranks it in the first place.",
"The edit distance model E still makes mistakes like with the list L-C, the word vivem 'live' translates into viven 'living' and the correct translation is vivir.",
"However, given a certain context/sense the previous translation can be correct.",
"The ranking scores given by the SVM varies from each list version.",
"For the L-C lists the scores are more uniform in increasing order and with a small variance.",
"The L and L-R lists show the opposite behaviour.",
"We add the produced Wikipedia n-best lists with the L metric into a SMT training dataset for the ptes pair.",
"We use the Moses SMT toolkit (Koehn et al., 2007) to test the augmented datasets.",
"We compare the augmented model with a baseline both trained by using the Zoo corpus of subtitles.",
"We use a 1-best list consisting of 100K pairs.",
"Te dataset used for pt-es baseline is: 80K training sentences, 1K sentences for tuning and 2K sen- Lang Pairs acc@1 acc@10 acc@1 acc@10 acc@1 acc@10 pt-es 20 60 22 59 32 70 it-es 16 53 18 45 44 66 fr-es 10 48 12 51 29 59 A possible reason for low improvement in terms of the BLEU scores is because MT evaluation metrics, such as BLEU, compare the MT output with a human reference.",
"The human reference translations in our corpus have been done from English (e.g., En→Es), while the test translations come from a related language (En→Pt→Es), often resulting in different paraphrases of the same English source.",
"While our OOV rate improved, the evaluation scores did not reflected this, because our MT output was still far from the reference even in cases it was otherwise acceptable.",
"List L List L-R List L-C Conclusions and future Work We have presented work in progress for developing MT dictionaries extracted from comparable resources for related languages.",
"The extraction heuristic show positive results on the n-best lists that group words with the same starting char-5 https://github.com/clab/fast_align 6 https://kheafield.com/code/kenlm/ 7 The p-value for the uk-ru pair is 0.06 we do not consider this result as statistically significant.",
"acters, because the used comparable corpora consist of related languages that share a similar orthography.",
"However, the lists based on the frequency window heuristic show poor results to include the correct translations during the extraction step.",
"Our ML models based on similarity metrics over parallel corpora show rankings similar to heldout data.",
"However, we created our training data using simple heuristics that simulate cognate words (i.e.",
"pairs of words with a small surface difference).",
"The ML models are able to rank similar words on the top of the list and they give a reliable score to discriminate wrong translations.",
"Preliminary results on the addition of the n-best lists into an SMT system show modest improvements compare to the baseline.",
"However, the OOV rate shows improvements around 10% reduction on word types, because of the wide variety of lexical choices introduced by the MT dictionaries.",
"Future work involves the addition of unsupervised morphology features for the n-best list extraction, i.e.",
"first step, given that the use of starting characters shows to be an effective heuristic to prune the search space and language independent.",
"Finally, we will measure the contribution for all the produced cognate lists, where we can try different strategies to add the dictionaries into an SMT system (Irvine and Callison-Burch, 2014) ."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"3",
"3.1",
"3.2",
"3.3",
"4"
],
"paper_header_content": [
"Introduction",
"Methodology",
"Cognate detection",
"Cognate Ranking",
"Results and Discussion",
"Data",
"Evaluation of the Ranking Model",
"Preliminary Results on Comparable Corpora",
"Conclusions and future Work"
]
} | GEM-SciDuet-train-8#paper-972#slide-5 | Features | Number of times of each edit operation, the model assigns a different weight to each operation
Cosine between the distributional vectors of the source and target words vectors from word2vec mapped to same space via a learned transformation matrix
SVM ranking default configuration (RBF kernel)
Easy-adapt features given different domains (Wikipedia, subtitles) | Number of times of each edit operation, the model assigns a different weight to each operation
Cosine between the distributional vectors of the source and target words vectors from word2vec mapped to same space via a learned transformation matrix
SVM ranking default configuration (RBF kernel)
Easy-adapt features given different domains (Wikipedia, subtitles) | [] |
GEM-SciDuet-train-8#paper-972#slide-6 | 972 | Obtaining SMT dictionaries for related languages | This study explores methods for developing Machine Translation dictionaries on the basis of word frequency lists coming from comparable corpora. We investigate (1) various methods to measure the similarity of cognates between related languages, (2) detection and removal of noisy cognate translations using SVM ranking. We show preliminary results on several Romance and Slavonic languages. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119
],
"paper_content_text": [
"Introduction Cognates are words having similarities in their spelling and meaning in two languages, either because the two languages are typologically related, e.g., maladie vs malattia ('disease'), or because they were both borrowed from the same source (informatique vs informatica).",
"The advantage of their use in Statistical Machine Translation (SMT) is that the procedure can be based on comparable corpora, i.e., similar corpora which are not translations of each other (Sharoff et al., 2013) .",
"Given that there are more sources of comparable corpora in comparison to parallel ones, the lexicon obtained from them is likely to be richer and more variable.",
"Detection of cognates is a well-known task, which has been explored for a range of languages using different methods.",
"The two main approaches applied to detection of the cognates are the generative and discriminative paradigms.",
"The first one is based on detection of the edit distance between potential candidate pairs.",
"The distance can be a simple Levenshtein distance, or a distance measure with the scores learned from an existing parallel set (Tiedemann, 1999; Mann and Yarowsky, 2001) .",
"The discriminative paradigm uses standard approaches to machine learning, which are based on (1) extracting features, e.g., character n-grams, and (2) learning to predict the transformations of the source word needed to (Jiampojamarn et al., 2010; Frunza and Inkpen, 2009) .",
"Given that SMT is usually based on a full-form lexicon, one of the possible issues in generation of cognates concerns the similarity of words in their root form vs the similarity in endings.",
"For example, the Ukrainian wordform áëèaeíüîãî 'near gen ' is cognate to Russian áëèaeíåãî, the root is identical, while the ending is considerably different (üîãî vs åãî).",
"Regular differences in the endings, which are shared across a large number of words, can be learned separately from the regular differences in the roots.",
"One also needs to take into account the false friends among cognates.",
"For example, diseñar means 'to design' in Spanish vs desenhar in Portuguese means 'to draw'.",
"There are also often cases of partial cognates, when the words share the meaning in some contexts, but not in others, e.g., aeåíà in Russian means 'wife', while its Bulgarian cognate aeåíà has two meanings: 'wife' and 'woman'.",
"Yet another complexity concerns a frequency mismatch.",
"Two cognates might differ in their frequency.",
"For example, dibujo in Spanish ('a drawing', rank 1779 in the Wikipedia frequency list) corresponds to a relatively rare cognate word debuxo in Portuguese (rank 104,514 in Wikipedia), while another Portuguese word desenho is more commonly used in this sense (rank 884 in the Portuguese Wikipedia).",
"For MT tasks we need translations that are equally appropriate in the source and target language, therefore cognates useful for a high-quality dictionary for SMT need to have roughly the same frequency in comparable corpora and they need to be used in similar contexts.",
"This study investigates the settings for extracting cognates for related languages in Romance and Slavonic language families for the task of reducing the number of unknown words for SMT.",
"This in-cludes the effects of having constraints for the cognates to be similar in their roots and in the endings, to occur in distributionally similar contexts and to have similar frequency.",
"Methodology The methodology for producing the list of cognates is based on the following steps: 1) Produce several lists of cognates using a family of distance measures, discussed in Section 2.1 from comparable corpora, 2) Prune the candidate lists by ranking items, this is done using a Machine Learning (ML) algorithm trained over parallel corpora for detecting the outliers, discussed in Section 2.2; The initial frequency lists for alignment are based Wikipedia dumps for the following languages: Romance (French, Italian, Spanish, Portuguese) and Slavonic (Bulgarian, Russian, Ukrainian), where the target languages are Spanish and Russian 1 .",
"Cognate detection We extract possible lists of cognates from comparable corpora by using a family of similarity measures: L direct matching between the languages using Levenshtein distance (Levenshtein, 1966) ; L(w s , w t ) = 1 − ed(w s , w t ) L-R Levenshtein distance with weights computed separately for the roots and for the endings; LR(r s , r t , e s , e t ) = α×ed(rs,rt)+β×ed(es,et) α+β L-C Levenshtein distance over word with similar number of starting characters (i.e.",
"prefix); LC(c s , c t ) = 1 − ed(c s , c t ), same prefix 0, otherwise where ed(., .)",
"is the normalised Levenshtein distance in characters between the source word w s and the target word w t .",
"The r s and r t are the stems produced by the Snowball stemmer 2 .",
"Since the Snowball stemmer does not support Ukrainian and Bulgarian, we used the Russian model for making the stem/ending split.",
"e s , e t are the characters at the end of a word form given a stem and c s , c t are the first n characters of a word.",
"In this work, we set the weights α = 0.6 and β = 0.4 giving more importance to the roots.",
"We set a higher weight to roots on the L-R, which is language dependent, and compare to the L-C metric, which is language independent.",
"We transform the Levenshtein distances into similarity metrics by subtracting the normalised distance score from one.",
"The produced lists contain for each source word the possible n-best target words accordingly to the maximum scores with one of the previous measures.",
"The n-best list allows possible cognate translations to a given source word that share a part of the surface form.",
"Different from (Mann and Yarowsky, 2001) , we produce n-best cognate lists scored by edit distance instead of 1-best.",
"An important problem when comparing comparable corpora is the way of representing the search space, where an exhaustive method compares all the combinations of source and target words (Mann and Yarowsky, 2001) .",
"We constraint the search space by comparing each source word against the target words that belong to a frequency window around the frequency of the source word.",
"This constraint only applies for the L and L-R metrics.",
"We use Wikipedia dumps for the source and target side processed in the form frequency lists.",
"We order the target side list into bins of similar frequency and for the source side we filter words that appear only once.",
"We use the window approach given that the frequency between the corpora under study can not be directly comparable.",
"During testing we use a wide window of ±200 bins to minimise the loss of good candidate translations.",
"The second search space constraint heuristic is the L-C metric.",
"This metric only compares source words with the target words upto a given n prefix.",
"For c s , c t in L-C , we use the first four characters to compare groups of words as suggested in (Kondrak et al., 2003) .",
"Cognate Ranking Given that the n-best lists contain noise, we aim to prune them by an ML ranking model.",
"However, there is a lack of resources to train a classification model for cognates (i.e.",
"cognate vs. false friend), as mentioned in (Fišer and Ljubešić, 2013) .",
"Available data that can be used to judge the cognate lists are the alignment pairs extracted from parallel data.",
"We decide to use a ranking model to avoid data imbalance present in classification and to use the probability scores of the alignment pairs as ranks, as opposed to the classification model used by (Irvine and Callison-Burch, 2013) .",
"Moreover, we also use a popular domain adaptation technique (Daumé et al., 2010) given that we have access to different domains of parallel training data that might be compatible with our comparable corpora.",
"The training data are the alignments between pairs of words where we rank them accordingly to their correspondent alignment probability from the output of GIZA++ (Och and Ney, 2003) .",
"We then use a heuristic to prune training data in order to simulate cognate words.",
"Pairs of words scored below the Levenshtein similarity threshold of 0.5 are not considered as cognates given that they are likely to have a different surface form.",
"We represent the training and test data with features extracted from different edit distance scores and distributional measures.",
"The edit distances features are as follows: 1) Similarity measure L and 2) Number of times of each edit operation.",
"Thus, the model assigns a different importance to each operation.",
"The distributional feature is based on the cosine between the distributional vectors of a window of n words around the word currently under comparison.",
"We train distributional similarity models with word2vec (Mikolov et al., 2013a) for the source and target side separately.",
"We extract the continuous vector for each word in the window, concatenate it and then compute the cosine between the concatenated vectors of the source and the target.",
"We suspect that the vectors will have similar behaviour between the source and the target given that they are trained under parallel Wikipedia articles.",
"We develop two ML models: 1) Edit distance scores and 2) Edit distance scores and distributional similarity score.",
"We use SVMlight (Joachims, 1998) Results and Discussion In this section we describe the data used to produce the n-best lists and train the cognate ranking models.",
"We evaluate the ranking models with heldout data from each training domain.",
"We also provide manual evaluation over the ranked n-best lists for error analysis.",
"Data The n-best lists to detect cognates were extracted from the respective Wikipedias by using the method described in Section 2.1.",
"The training data for the ranking model consists of different types of parallel corpora.",
"The parallel corpora are as follows: 1) Wiki-titles we use the inter language links to create a parallel corpus from the tittles of the Wikipedia articles, with about 500K aligned links (i.e.",
"'sentences') per language pair (about 200k for bg-ru), giving us about 200K training instances per language pair 3 , 2) Opensubs is an open source corpus of subtitles built by the fan community, with 1M sentences, 6M tokens, 100K words, giving about 90K training instances (Tiedemann, 2012) and 3) Zoo is a proprietary corpus of subtitles produced by professional translators, with 100K sentences, 700K tokens, 40K words and giving about 20K training instances per language pair.",
"Our objective is to create MT dictionaries from the produced n-best lists and we use parallel data as a source of training to prune them.",
"We are interested in the corpora of subtitles because the chosen domain of our SMT experiments is subtitling, while the proposed ranking method can be used in other application domains as well.",
"We consider Zoo and Opensubs as two different domains given that they were built by different types of translators and they differ in size and quality.",
"The heldout data consists of 2K instances for each corpus.",
"We use Wikipedia documents and Opensusbs subtitles to train word2vec for the distributional similarity feature.",
"We use the continuous bag-ofwords algorithm for word2vec and set the parameters for training to 200 dimensions and a window of 8 words.",
"The Wikipedia documents with an average number of 70K documents for each language, and Opensubs subtitles with 1M sentences for each language.",
"In practice we only use the Wikipedia data given that for Opensubs the model is able to find relatively few vectors, for example a vector is found only for 20% of the words in the pt-es pair.",
"Evaluation of the Ranking Model We define two ranking models as: model E for edit distance features and model EC for both edit Table 1 shows the results of the ranking procedure.",
"For the Romance family language pairs the model EC with context features consistently reduces the error compared to the solely use of edit distance metrics.",
"The only exception is the it-es EC model with poor results for the domain of Wiki-titles.",
"The models for the Slavonic family behave similarly to the Romance family, where the use of context features reduces the ranking error.",
"The exception is the bg-ru model on the Opensubs domain.",
"A possible reason for the poor results on the ites and bg-ru models is that the model often assigns a high similarity score to unrelated words.",
"For example, in it-es, mortes 'deads' is treated as close to categoria 'category'.",
"A possible solution is to map the vectors form the source side into the space of the target side via a learned transformation matrix (Mikolov et al., 2013b) .",
"Preliminary Results on Comparable Corpora After we extracted the n-best lists for the Romance family comparable corpora, we applied one of the ranking models on these lists and we manually evaluated over a sample of 50 words 4 .",
"We set n to 10 for the n-best lists.",
"We use a frequency window of 200 for the n-best list search heuristic and the domain of the comparable corpora to Wiki-titles 4 The sample consists of words with a frequency between 2K and 5. for the domain adaptation technique.",
"The purpose of manual evaluation is to decide whether the ML setup is sensible on the objective task.",
"Each list is evaluated by accuracy at 1 and accuracy at 10.",
"We also show success and failure examples of the ranking and the n-best lists.",
"Table 2 shows the preliminary results of the ML model E on a sample of Wikipedia dumps.",
"The L and L-R lists consistently show poor results.",
"A possible reason is the amount of errors given the first step to extract the n-best lists.",
"For example, in pt-es, for the word vivem 'live' the 10-best list only contain one word with a similar meaning viva 'living' but it can be also translated as 'cheers'.",
"In the pt-es list for the word representação 'description' the correct translation representación is not among the 10-best in the L list.",
"However, it is present in the 10-best for the L-C list and the ML model EC ranks it in the first place.",
"The edit distance model E still makes mistakes like with the list L-C, the word vivem 'live' translates into viven 'living' and the correct translation is vivir.",
"However, given a certain context/sense the previous translation can be correct.",
"The ranking scores given by the SVM varies from each list version.",
"For the L-C lists the scores are more uniform in increasing order and with a small variance.",
"The L and L-R lists show the opposite behaviour.",
"We add the produced Wikipedia n-best lists with the L metric into a SMT training dataset for the ptes pair.",
"We use the Moses SMT toolkit (Koehn et al., 2007) to test the augmented datasets.",
"We compare the augmented model with a baseline both trained by using the Zoo corpus of subtitles.",
"We use a 1-best list consisting of 100K pairs.",
"Te dataset used for pt-es baseline is: 80K training sentences, 1K sentences for tuning and 2K sen- Lang Pairs acc@1 acc@10 acc@1 acc@10 acc@1 acc@10 pt-es 20 60 22 59 32 70 it-es 16 53 18 45 44 66 fr-es 10 48 12 51 29 59 A possible reason for low improvement in terms of the BLEU scores is because MT evaluation metrics, such as BLEU, compare the MT output with a human reference.",
"The human reference translations in our corpus have been done from English (e.g., En→Es), while the test translations come from a related language (En→Pt→Es), often resulting in different paraphrases of the same English source.",
"While our OOV rate improved, the evaluation scores did not reflected this, because our MT output was still far from the reference even in cases it was otherwise acceptable.",
"List L List L-R List L-C Conclusions and future Work We have presented work in progress for developing MT dictionaries extracted from comparable resources for related languages.",
"The extraction heuristic show positive results on the n-best lists that group words with the same starting char-5 https://github.com/clab/fast_align 6 https://kheafield.com/code/kenlm/ 7 The p-value for the uk-ru pair is 0.06 we do not consider this result as statistically significant.",
"acters, because the used comparable corpora consist of related languages that share a similar orthography.",
"However, the lists based on the frequency window heuristic show poor results to include the correct translations during the extraction step.",
"Our ML models based on similarity metrics over parallel corpora show rankings similar to heldout data.",
"However, we created our training data using simple heuristics that simulate cognate words (i.e.",
"pairs of words with a small surface difference).",
"The ML models are able to rank similar words on the top of the list and they give a reliable score to discriminate wrong translations.",
"Preliminary results on the addition of the n-best lists into an SMT system show modest improvements compare to the baseline.",
"However, the OOV rate shows improvements around 10% reduction on word types, because of the wide variety of lexical choices introduced by the MT dictionaries.",
"Future work involves the addition of unsupervised morphology features for the n-best list extraction, i.e.",
"first step, given that the use of starting characters shows to be an effective heuristic to prune the search space and language independent.",
"Finally, we will measure the contribution for all the produced cognate lists, where we can try different strategies to add the dictionaries into an SMT system (Irvine and Callison-Burch, 2014) ."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"3",
"3.1",
"3.2",
"3.3",
"4"
],
"paper_header_content": [
"Introduction",
"Methodology",
"Cognate detection",
"Cognate Ranking",
"Results and Discussion",
"Data",
"Evaluation of the Ranking Model",
"Preliminary Results on Comparable Corpora",
"Conclusions and future Work"
]
} | GEM-SciDuet-train-8#paper-972#slide-6 | Data description | n-best lists from Wikipedia dumps (frequency lists)
ML training Wiki-titles, parallel data from inter language links from the tittles of the Wikipedia articles 500K aligned links (i.e. sentences)
Opensubs, 90K training instances
Zoo proprietary corpus of subtitles produced by professional translators, 20K training instances
Ranking test Heldout data from training
Manual cognate test Wikipedia most frequent words
SMT test Zoo data | n-best lists from Wikipedia dumps (frequency lists)
ML training Wiki-titles, parallel data from inter language links from the tittles of the Wikipedia articles 500K aligned links (i.e. sentences)
Opensubs, 90K training instances
Zoo proprietary corpus of subtitles produced by professional translators, 20K training instances
Ranking test Heldout data from training
Manual cognate test Wikipedia most frequent words
SMT test Zoo data | [] |
GEM-SciDuet-train-8#paper-972#slide-7 | 972 | Obtaining SMT dictionaries for related languages | This study explores methods for developing Machine Translation dictionaries on the basis of word frequency lists coming from comparable corpora. We investigate (1) various methods to measure the similarity of cognates between related languages, (2) detection and removal of noisy cognate translations using SVM ranking. We show preliminary results on several Romance and Slavonic languages. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119
],
"paper_content_text": [
"Introduction Cognates are words having similarities in their spelling and meaning in two languages, either because the two languages are typologically related, e.g., maladie vs malattia ('disease'), or because they were both borrowed from the same source (informatique vs informatica).",
"The advantage of their use in Statistical Machine Translation (SMT) is that the procedure can be based on comparable corpora, i.e., similar corpora which are not translations of each other (Sharoff et al., 2013) .",
"Given that there are more sources of comparable corpora in comparison to parallel ones, the lexicon obtained from them is likely to be richer and more variable.",
"Detection of cognates is a well-known task, which has been explored for a range of languages using different methods.",
"The two main approaches applied to detection of the cognates are the generative and discriminative paradigms.",
"The first one is based on detection of the edit distance between potential candidate pairs.",
"The distance can be a simple Levenshtein distance, or a distance measure with the scores learned from an existing parallel set (Tiedemann, 1999; Mann and Yarowsky, 2001) .",
"The discriminative paradigm uses standard approaches to machine learning, which are based on (1) extracting features, e.g., character n-grams, and (2) learning to predict the transformations of the source word needed to (Jiampojamarn et al., 2010; Frunza and Inkpen, 2009) .",
"Given that SMT is usually based on a full-form lexicon, one of the possible issues in generation of cognates concerns the similarity of words in their root form vs the similarity in endings.",
"For example, the Ukrainian wordform áëèaeíüîãî 'near gen ' is cognate to Russian áëèaeíåãî, the root is identical, while the ending is considerably different (üîãî vs åãî).",
"Regular differences in the endings, which are shared across a large number of words, can be learned separately from the regular differences in the roots.",
"One also needs to take into account the false friends among cognates.",
"For example, diseñar means 'to design' in Spanish vs desenhar in Portuguese means 'to draw'.",
"There are also often cases of partial cognates, when the words share the meaning in some contexts, but not in others, e.g., aeåíà in Russian means 'wife', while its Bulgarian cognate aeåíà has two meanings: 'wife' and 'woman'.",
"Yet another complexity concerns a frequency mismatch.",
"Two cognates might differ in their frequency.",
"For example, dibujo in Spanish ('a drawing', rank 1779 in the Wikipedia frequency list) corresponds to a relatively rare cognate word debuxo in Portuguese (rank 104,514 in Wikipedia), while another Portuguese word desenho is more commonly used in this sense (rank 884 in the Portuguese Wikipedia).",
"For MT tasks we need translations that are equally appropriate in the source and target language, therefore cognates useful for a high-quality dictionary for SMT need to have roughly the same frequency in comparable corpora and they need to be used in similar contexts.",
"This study investigates the settings for extracting cognates for related languages in Romance and Slavonic language families for the task of reducing the number of unknown words for SMT.",
"This in-cludes the effects of having constraints for the cognates to be similar in their roots and in the endings, to occur in distributionally similar contexts and to have similar frequency.",
"Methodology The methodology for producing the list of cognates is based on the following steps: 1) Produce several lists of cognates using a family of distance measures, discussed in Section 2.1 from comparable corpora, 2) Prune the candidate lists by ranking items, this is done using a Machine Learning (ML) algorithm trained over parallel corpora for detecting the outliers, discussed in Section 2.2; The initial frequency lists for alignment are based Wikipedia dumps for the following languages: Romance (French, Italian, Spanish, Portuguese) and Slavonic (Bulgarian, Russian, Ukrainian), where the target languages are Spanish and Russian 1 .",
"Cognate detection We extract possible lists of cognates from comparable corpora by using a family of similarity measures: L direct matching between the languages using Levenshtein distance (Levenshtein, 1966) ; L(w s , w t ) = 1 − ed(w s , w t ) L-R Levenshtein distance with weights computed separately for the roots and for the endings; LR(r s , r t , e s , e t ) = α×ed(rs,rt)+β×ed(es,et) α+β L-C Levenshtein distance over word with similar number of starting characters (i.e.",
"prefix); LC(c s , c t ) = 1 − ed(c s , c t ), same prefix 0, otherwise where ed(., .)",
"is the normalised Levenshtein distance in characters between the source word w s and the target word w t .",
"The r s and r t are the stems produced by the Snowball stemmer 2 .",
"Since the Snowball stemmer does not support Ukrainian and Bulgarian, we used the Russian model for making the stem/ending split.",
"e s , e t are the characters at the end of a word form given a stem and c s , c t are the first n characters of a word.",
"In this work, we set the weights α = 0.6 and β = 0.4 giving more importance to the roots.",
"We set a higher weight to roots on the L-R, which is language dependent, and compare to the L-C metric, which is language independent.",
"We transform the Levenshtein distances into similarity metrics by subtracting the normalised distance score from one.",
"The produced lists contain for each source word the possible n-best target words accordingly to the maximum scores with one of the previous measures.",
"The n-best list allows possible cognate translations to a given source word that share a part of the surface form.",
"Different from (Mann and Yarowsky, 2001) , we produce n-best cognate lists scored by edit distance instead of 1-best.",
"An important problem when comparing comparable corpora is the way of representing the search space, where an exhaustive method compares all the combinations of source and target words (Mann and Yarowsky, 2001) .",
"We constraint the search space by comparing each source word against the target words that belong to a frequency window around the frequency of the source word.",
"This constraint only applies for the L and L-R metrics.",
"We use Wikipedia dumps for the source and target side processed in the form frequency lists.",
"We order the target side list into bins of similar frequency and for the source side we filter words that appear only once.",
"We use the window approach given that the frequency between the corpora under study can not be directly comparable.",
"During testing we use a wide window of ±200 bins to minimise the loss of good candidate translations.",
"The second search space constraint heuristic is the L-C metric.",
"This metric only compares source words with the target words upto a given n prefix.",
"For c s , c t in L-C , we use the first four characters to compare groups of words as suggested in (Kondrak et al., 2003) .",
"Cognate Ranking Given that the n-best lists contain noise, we aim to prune them by an ML ranking model.",
"However, there is a lack of resources to train a classification model for cognates (i.e.",
"cognate vs. false friend), as mentioned in (Fišer and Ljubešić, 2013) .",
"Available data that can be used to judge the cognate lists are the alignment pairs extracted from parallel data.",
"We decide to use a ranking model to avoid data imbalance present in classification and to use the probability scores of the alignment pairs as ranks, as opposed to the classification model used by (Irvine and Callison-Burch, 2013) .",
"Moreover, we also use a popular domain adaptation technique (Daumé et al., 2010) given that we have access to different domains of parallel training data that might be compatible with our comparable corpora.",
"The training data are the alignments between pairs of words where we rank them accordingly to their correspondent alignment probability from the output of GIZA++ (Och and Ney, 2003) .",
"We then use a heuristic to prune training data in order to simulate cognate words.",
"Pairs of words scored below the Levenshtein similarity threshold of 0.5 are not considered as cognates given that they are likely to have a different surface form.",
"We represent the training and test data with features extracted from different edit distance scores and distributional measures.",
"The edit distances features are as follows: 1) Similarity measure L and 2) Number of times of each edit operation.",
"Thus, the model assigns a different importance to each operation.",
"The distributional feature is based on the cosine between the distributional vectors of a window of n words around the word currently under comparison.",
"We train distributional similarity models with word2vec (Mikolov et al., 2013a) for the source and target side separately.",
"We extract the continuous vector for each word in the window, concatenate it and then compute the cosine between the concatenated vectors of the source and the target.",
"We suspect that the vectors will have similar behaviour between the source and the target given that they are trained under parallel Wikipedia articles.",
"We develop two ML models: 1) Edit distance scores and 2) Edit distance scores and distributional similarity score.",
"We use SVMlight (Joachims, 1998) Results and Discussion In this section we describe the data used to produce the n-best lists and train the cognate ranking models.",
"We evaluate the ranking models with heldout data from each training domain.",
"We also provide manual evaluation over the ranked n-best lists for error analysis.",
"Data The n-best lists to detect cognates were extracted from the respective Wikipedias by using the method described in Section 2.1.",
"The training data for the ranking model consists of different types of parallel corpora.",
"The parallel corpora are as follows: 1) Wiki-titles we use the inter language links to create a parallel corpus from the tittles of the Wikipedia articles, with about 500K aligned links (i.e.",
"'sentences') per language pair (about 200k for bg-ru), giving us about 200K training instances per language pair 3 , 2) Opensubs is an open source corpus of subtitles built by the fan community, with 1M sentences, 6M tokens, 100K words, giving about 90K training instances (Tiedemann, 2012) and 3) Zoo is a proprietary corpus of subtitles produced by professional translators, with 100K sentences, 700K tokens, 40K words and giving about 20K training instances per language pair.",
"Our objective is to create MT dictionaries from the produced n-best lists and we use parallel data as a source of training to prune them.",
"We are interested in the corpora of subtitles because the chosen domain of our SMT experiments is subtitling, while the proposed ranking method can be used in other application domains as well.",
"We consider Zoo and Opensubs as two different domains given that they were built by different types of translators and they differ in size and quality.",
"The heldout data consists of 2K instances for each corpus.",
"We use Wikipedia documents and Opensusbs subtitles to train word2vec for the distributional similarity feature.",
"We use the continuous bag-ofwords algorithm for word2vec and set the parameters for training to 200 dimensions and a window of 8 words.",
"The Wikipedia documents with an average number of 70K documents for each language, and Opensubs subtitles with 1M sentences for each language.",
"In practice we only use the Wikipedia data given that for Opensubs the model is able to find relatively few vectors, for example a vector is found only for 20% of the words in the pt-es pair.",
"Evaluation of the Ranking Model We define two ranking models as: model E for edit distance features and model EC for both edit Table 1 shows the results of the ranking procedure.",
"For the Romance family language pairs the model EC with context features consistently reduces the error compared to the solely use of edit distance metrics.",
"The only exception is the it-es EC model with poor results for the domain of Wiki-titles.",
"The models for the Slavonic family behave similarly to the Romance family, where the use of context features reduces the ranking error.",
"The exception is the bg-ru model on the Opensubs domain.",
"A possible reason for the poor results on the ites and bg-ru models is that the model often assigns a high similarity score to unrelated words.",
"For example, in it-es, mortes 'deads' is treated as close to categoria 'category'.",
"A possible solution is to map the vectors form the source side into the space of the target side via a learned transformation matrix (Mikolov et al., 2013b) .",
"Preliminary Results on Comparable Corpora After we extracted the n-best lists for the Romance family comparable corpora, we applied one of the ranking models on these lists and we manually evaluated over a sample of 50 words 4 .",
"We set n to 10 for the n-best lists.",
"We use a frequency window of 200 for the n-best list search heuristic and the domain of the comparable corpora to Wiki-titles 4 The sample consists of words with a frequency between 2K and 5. for the domain adaptation technique.",
"The purpose of manual evaluation is to decide whether the ML setup is sensible on the objective task.",
"Each list is evaluated by accuracy at 1 and accuracy at 10.",
"We also show success and failure examples of the ranking and the n-best lists.",
"Table 2 shows the preliminary results of the ML model E on a sample of Wikipedia dumps.",
"The L and L-R lists consistently show poor results.",
"A possible reason is the amount of errors given the first step to extract the n-best lists.",
"For example, in pt-es, for the word vivem 'live' the 10-best list only contain one word with a similar meaning viva 'living' but it can be also translated as 'cheers'.",
"In the pt-es list for the word representação 'description' the correct translation representación is not among the 10-best in the L list.",
"However, it is present in the 10-best for the L-C list and the ML model EC ranks it in the first place.",
"The edit distance model E still makes mistakes like with the list L-C, the word vivem 'live' translates into viven 'living' and the correct translation is vivir.",
"However, given a certain context/sense the previous translation can be correct.",
"The ranking scores given by the SVM varies from each list version.",
"For the L-C lists the scores are more uniform in increasing order and with a small variance.",
"The L and L-R lists show the opposite behaviour.",
"We add the produced Wikipedia n-best lists with the L metric into a SMT training dataset for the ptes pair.",
"We use the Moses SMT toolkit (Koehn et al., 2007) to test the augmented datasets.",
"We compare the augmented model with a baseline both trained by using the Zoo corpus of subtitles.",
"We use a 1-best list consisting of 100K pairs.",
"Te dataset used for pt-es baseline is: 80K training sentences, 1K sentences for tuning and 2K sen- Lang Pairs acc@1 acc@10 acc@1 acc@10 acc@1 acc@10 pt-es 20 60 22 59 32 70 it-es 16 53 18 45 44 66 fr-es 10 48 12 51 29 59 A possible reason for low improvement in terms of the BLEU scores is because MT evaluation metrics, such as BLEU, compare the MT output with a human reference.",
"The human reference translations in our corpus have been done from English (e.g., En→Es), while the test translations come from a related language (En→Pt→Es), often resulting in different paraphrases of the same English source.",
"While our OOV rate improved, the evaluation scores did not reflected this, because our MT output was still far from the reference even in cases it was otherwise acceptable.",
"List L List L-R List L-C Conclusions and future Work We have presented work in progress for developing MT dictionaries extracted from comparable resources for related languages.",
"The extraction heuristic show positive results on the n-best lists that group words with the same starting char-5 https://github.com/clab/fast_align 6 https://kheafield.com/code/kenlm/ 7 The p-value for the uk-ru pair is 0.06 we do not consider this result as statistically significant.",
"acters, because the used comparable corpora consist of related languages that share a similar orthography.",
"However, the lists based on the frequency window heuristic show poor results to include the correct translations during the extraction step.",
"Our ML models based on similarity metrics over parallel corpora show rankings similar to heldout data.",
"However, we created our training data using simple heuristics that simulate cognate words (i.e.",
"pairs of words with a small surface difference).",
"The ML models are able to rank similar words on the top of the list and they give a reliable score to discriminate wrong translations.",
"Preliminary results on the addition of the n-best lists into an SMT system show modest improvements compare to the baseline.",
"However, the OOV rate shows improvements around 10% reduction on word types, because of the wide variety of lexical choices introduced by the MT dictionaries.",
"Future work involves the addition of unsupervised morphology features for the n-best list extraction, i.e.",
"first step, given that the use of starting characters shows to be an effective heuristic to prune the search space and language independent.",
"Finally, we will measure the contribution for all the produced cognate lists, where we can try different strategies to add the dictionaries into an SMT system (Irvine and Callison-Burch, 2014) ."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"3",
"3.1",
"3.2",
"3.3",
"4"
],
"paper_header_content": [
"Introduction",
"Methodology",
"Cognate detection",
"Cognate Ranking",
"Results and Discussion",
"Data",
"Evaluation of the Ranking Model",
"Preliminary Results on Comparable Corpora",
"Conclusions and future Work"
]
} | GEM-SciDuet-train-8#paper-972#slide-7 | Language pairs | Romance Source: Portuguese, French, Italian Target: Spanish
Slavonic Source: Ukrainian, Bulgarian Target: Russian | Romance Source: Portuguese, French, Italian Target: Spanish
Slavonic Source: Ukrainian, Bulgarian Target: Russian | [] |
GEM-SciDuet-train-8#paper-972#slide-8 | 972 | Obtaining SMT dictionaries for related languages | This study explores methods for developing Machine Translation dictionaries on the basis of word frequency lists coming from comparable corpora. We investigate (1) various methods to measure the similarity of cognates between related languages, (2) detection and removal of noisy cognate translations using SVM ranking. We show preliminary results on several Romance and Slavonic languages. | {
"paper_content_id": [
0,
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69,
70,
71,
72,
73,
74,
75,
76,
77,
78,
79,
80,
81,
82,
83,
84,
85,
86,
87,
88,
89,
90,
91,
92,
93,
94,
95,
96,
97,
98,
99,
100,
101,
102,
103,
104,
105,
106,
107,
108,
109,
110,
111,
112,
113,
114,
115,
116,
117,
118,
119
],
"paper_content_text": [
"Introduction Cognates are words having similarities in their spelling and meaning in two languages, either because the two languages are typologically related, e.g., maladie vs malattia ('disease'), or because they were both borrowed from the same source (informatique vs informatica).",
"The advantage of their use in Statistical Machine Translation (SMT) is that the procedure can be based on comparable corpora, i.e., similar corpora which are not translations of each other (Sharoff et al., 2013) .",
"Given that there are more sources of comparable corpora in comparison to parallel ones, the lexicon obtained from them is likely to be richer and more variable.",
"Detection of cognates is a well-known task, which has been explored for a range of languages using different methods.",
"The two main approaches applied to detection of the cognates are the generative and discriminative paradigms.",
"The first one is based on detection of the edit distance between potential candidate pairs.",
"The distance can be a simple Levenshtein distance, or a distance measure with the scores learned from an existing parallel set (Tiedemann, 1999; Mann and Yarowsky, 2001) .",
"The discriminative paradigm uses standard approaches to machine learning, which are based on (1) extracting features, e.g., character n-grams, and (2) learning to predict the transformations of the source word needed to (Jiampojamarn et al., 2010; Frunza and Inkpen, 2009) .",
"Given that SMT is usually based on a full-form lexicon, one of the possible issues in generation of cognates concerns the similarity of words in their root form vs the similarity in endings.",
"For example, the Ukrainian wordform áëèaeíüîãî 'near gen ' is cognate to Russian áëèaeíåãî, the root is identical, while the ending is considerably different (üîãî vs åãî).",
"Regular differences in the endings, which are shared across a large number of words, can be learned separately from the regular differences in the roots.",
"One also needs to take into account the false friends among cognates.",
"For example, diseñar means 'to design' in Spanish vs desenhar in Portuguese means 'to draw'.",
"There are also often cases of partial cognates, when the words share the meaning in some contexts, but not in others, e.g., aeåíà in Russian means 'wife', while its Bulgarian cognate aeåíà has two meanings: 'wife' and 'woman'.",
"Yet another complexity concerns a frequency mismatch.",
"Two cognates might differ in their frequency.",
"For example, dibujo in Spanish ('a drawing', rank 1779 in the Wikipedia frequency list) corresponds to a relatively rare cognate word debuxo in Portuguese (rank 104,514 in Wikipedia), while another Portuguese word desenho is more commonly used in this sense (rank 884 in the Portuguese Wikipedia).",
"For MT tasks we need translations that are equally appropriate in the source and target language, therefore cognates useful for a high-quality dictionary for SMT need to have roughly the same frequency in comparable corpora and they need to be used in similar contexts.",
"This study investigates the settings for extracting cognates for related languages in Romance and Slavonic language families for the task of reducing the number of unknown words for SMT.",
"This in-cludes the effects of having constraints for the cognates to be similar in their roots and in the endings, to occur in distributionally similar contexts and to have similar frequency.",
"Methodology The methodology for producing the list of cognates is based on the following steps: 1) Produce several lists of cognates using a family of distance measures, discussed in Section 2.1 from comparable corpora, 2) Prune the candidate lists by ranking items, this is done using a Machine Learning (ML) algorithm trained over parallel corpora for detecting the outliers, discussed in Section 2.2; The initial frequency lists for alignment are based Wikipedia dumps for the following languages: Romance (French, Italian, Spanish, Portuguese) and Slavonic (Bulgarian, Russian, Ukrainian), where the target languages are Spanish and Russian 1 .",
"Cognate detection We extract possible lists of cognates from comparable corpora by using a family of similarity measures: L direct matching between the languages using Levenshtein distance (Levenshtein, 1966) ; L(w s , w t ) = 1 − ed(w s , w t ) L-R Levenshtein distance with weights computed separately for the roots and for the endings; LR(r s , r t , e s , e t ) = α×ed(rs,rt)+β×ed(es,et) α+β L-C Levenshtein distance over word with similar number of starting characters (i.e.",
"prefix); LC(c s , c t ) = 1 − ed(c s , c t ), same prefix 0, otherwise where ed(., .)",
"is the normalised Levenshtein distance in characters between the source word w s and the target word w t .",
"The r s and r t are the stems produced by the Snowball stemmer 2 .",
"Since the Snowball stemmer does not support Ukrainian and Bulgarian, we used the Russian model for making the stem/ending split.",
"e s , e t are the characters at the end of a word form given a stem and c s , c t are the first n characters of a word.",
"In this work, we set the weights α = 0.6 and β = 0.4 giving more importance to the roots.",
"We set a higher weight to roots on the L-R, which is language dependent, and compare to the L-C metric, which is language independent.",
"We transform the Levenshtein distances into similarity metrics by subtracting the normalised distance score from one.",
"The produced lists contain for each source word the possible n-best target words accordingly to the maximum scores with one of the previous measures.",
"The n-best list allows possible cognate translations to a given source word that share a part of the surface form.",
"Different from (Mann and Yarowsky, 2001) , we produce n-best cognate lists scored by edit distance instead of 1-best.",
"An important problem when comparing comparable corpora is the way of representing the search space, where an exhaustive method compares all the combinations of source and target words (Mann and Yarowsky, 2001) .",
"We constraint the search space by comparing each source word against the target words that belong to a frequency window around the frequency of the source word.",
"This constraint only applies for the L and L-R metrics.",
"We use Wikipedia dumps for the source and target side processed in the form frequency lists.",
"We order the target side list into bins of similar frequency and for the source side we filter words that appear only once.",
"We use the window approach given that the frequency between the corpora under study can not be directly comparable.",
"During testing we use a wide window of ±200 bins to minimise the loss of good candidate translations.",
"The second search space constraint heuristic is the L-C metric.",
"This metric only compares source words with the target words upto a given n prefix.",
"For c s , c t in L-C , we use the first four characters to compare groups of words as suggested in (Kondrak et al., 2003) .",
"Cognate Ranking Given that the n-best lists contain noise, we aim to prune them by an ML ranking model.",
"However, there is a lack of resources to train a classification model for cognates (i.e.",
"cognate vs. false friend), as mentioned in (Fišer and Ljubešić, 2013) .",
"Available data that can be used to judge the cognate lists are the alignment pairs extracted from parallel data.",
"We decide to use a ranking model to avoid data imbalance present in classification and to use the probability scores of the alignment pairs as ranks, as opposed to the classification model used by (Irvine and Callison-Burch, 2013) .",
"Moreover, we also use a popular domain adaptation technique (Daumé et al., 2010) given that we have access to different domains of parallel training data that might be compatible with our comparable corpora.",
"The training data are the alignments between pairs of words where we rank them accordingly to their correspondent alignment probability from the output of GIZA++ (Och and Ney, 2003) .",
"We then use a heuristic to prune training data in order to simulate cognate words.",
"Pairs of words scored below the Levenshtein similarity threshold of 0.5 are not considered as cognates given that they are likely to have a different surface form.",
"We represent the training and test data with features extracted from different edit distance scores and distributional measures.",
"The edit distances features are as follows: 1) Similarity measure L and 2) Number of times of each edit operation.",
"Thus, the model assigns a different importance to each operation.",
"The distributional feature is based on the cosine between the distributional vectors of a window of n words around the word currently under comparison.",
"We train distributional similarity models with word2vec (Mikolov et al., 2013a) for the source and target side separately.",
"We extract the continuous vector for each word in the window, concatenate it and then compute the cosine between the concatenated vectors of the source and the target.",
"We suspect that the vectors will have similar behaviour between the source and the target given that they are trained under parallel Wikipedia articles.",
"We develop two ML models: 1) Edit distance scores and 2) Edit distance scores and distributional similarity score.",
"We use SVMlight (Joachims, 1998) Results and Discussion In this section we describe the data used to produce the n-best lists and train the cognate ranking models.",
"We evaluate the ranking models with heldout data from each training domain.",
"We also provide manual evaluation over the ranked n-best lists for error analysis.",
"Data The n-best lists to detect cognates were extracted from the respective Wikipedias by using the method described in Section 2.1.",
"The training data for the ranking model consists of different types of parallel corpora.",
"The parallel corpora are as follows: 1) Wiki-titles we use the inter language links to create a parallel corpus from the tittles of the Wikipedia articles, with about 500K aligned links (i.e.",
"'sentences') per language pair (about 200k for bg-ru), giving us about 200K training instances per language pair 3 , 2) Opensubs is an open source corpus of subtitles built by the fan community, with 1M sentences, 6M tokens, 100K words, giving about 90K training instances (Tiedemann, 2012) and 3) Zoo is a proprietary corpus of subtitles produced by professional translators, with 100K sentences, 700K tokens, 40K words and giving about 20K training instances per language pair.",
"Our objective is to create MT dictionaries from the produced n-best lists and we use parallel data as a source of training to prune them.",
"We are interested in the corpora of subtitles because the chosen domain of our SMT experiments is subtitling, while the proposed ranking method can be used in other application domains as well.",
"We consider Zoo and Opensubs as two different domains given that they were built by different types of translators and they differ in size and quality.",
"The heldout data consists of 2K instances for each corpus.",
"We use Wikipedia documents and Opensusbs subtitles to train word2vec for the distributional similarity feature.",
"We use the continuous bag-ofwords algorithm for word2vec and set the parameters for training to 200 dimensions and a window of 8 words.",
"The Wikipedia documents with an average number of 70K documents for each language, and Opensubs subtitles with 1M sentences for each language.",
"In practice we only use the Wikipedia data given that for Opensubs the model is able to find relatively few vectors, for example a vector is found only for 20% of the words in the pt-es pair.",
"Evaluation of the Ranking Model We define two ranking models as: model E for edit distance features and model EC for both edit Table 1 shows the results of the ranking procedure.",
"For the Romance family language pairs the model EC with context features consistently reduces the error compared to the solely use of edit distance metrics.",
"The only exception is the it-es EC model with poor results for the domain of Wiki-titles.",
"The models for the Slavonic family behave similarly to the Romance family, where the use of context features reduces the ranking error.",
"The exception is the bg-ru model on the Opensubs domain.",
"A possible reason for the poor results on the ites and bg-ru models is that the model often assigns a high similarity score to unrelated words.",
"For example, in it-es, mortes 'deads' is treated as close to categoria 'category'.",
"A possible solution is to map the vectors form the source side into the space of the target side via a learned transformation matrix (Mikolov et al., 2013b) .",
"Preliminary Results on Comparable Corpora After we extracted the n-best lists for the Romance family comparable corpora, we applied one of the ranking models on these lists and we manually evaluated over a sample of 50 words 4 .",
"We set n to 10 for the n-best lists.",
"We use a frequency window of 200 for the n-best list search heuristic and the domain of the comparable corpora to Wiki-titles 4 The sample consists of words with a frequency between 2K and 5. for the domain adaptation technique.",
"The purpose of manual evaluation is to decide whether the ML setup is sensible on the objective task.",
"Each list is evaluated by accuracy at 1 and accuracy at 10.",
"We also show success and failure examples of the ranking and the n-best lists.",
"Table 2 shows the preliminary results of the ML model E on a sample of Wikipedia dumps.",
"The L and L-R lists consistently show poor results.",
"A possible reason is the amount of errors given the first step to extract the n-best lists.",
"For example, in pt-es, for the word vivem 'live' the 10-best list only contain one word with a similar meaning viva 'living' but it can be also translated as 'cheers'.",
"In the pt-es list for the word representação 'description' the correct translation representación is not among the 10-best in the L list.",
"However, it is present in the 10-best for the L-C list and the ML model EC ranks it in the first place.",
"The edit distance model E still makes mistakes like with the list L-C, the word vivem 'live' translates into viven 'living' and the correct translation is vivir.",
"However, given a certain context/sense the previous translation can be correct.",
"The ranking scores given by the SVM varies from each list version.",
"For the L-C lists the scores are more uniform in increasing order and with a small variance.",
"The L and L-R lists show the opposite behaviour.",
"We add the produced Wikipedia n-best lists with the L metric into a SMT training dataset for the ptes pair.",
"We use the Moses SMT toolkit (Koehn et al., 2007) to test the augmented datasets.",
"We compare the augmented model with a baseline both trained by using the Zoo corpus of subtitles.",
"We use a 1-best list consisting of 100K pairs.",
"Te dataset used for pt-es baseline is: 80K training sentences, 1K sentences for tuning and 2K sen- Lang Pairs acc@1 acc@10 acc@1 acc@10 acc@1 acc@10 pt-es 20 60 22 59 32 70 it-es 16 53 18 45 44 66 fr-es 10 48 12 51 29 59 A possible reason for low improvement in terms of the BLEU scores is because MT evaluation metrics, such as BLEU, compare the MT output with a human reference.",
"The human reference translations in our corpus have been done from English (e.g., En→Es), while the test translations come from a related language (En→Pt→Es), often resulting in different paraphrases of the same English source.",
"While our OOV rate improved, the evaluation scores did not reflected this, because our MT output was still far from the reference even in cases it was otherwise acceptable.",
"List L List L-R List L-C Conclusions and future Work We have presented work in progress for developing MT dictionaries extracted from comparable resources for related languages.",
"The extraction heuristic show positive results on the n-best lists that group words with the same starting char-5 https://github.com/clab/fast_align 6 https://kheafield.com/code/kenlm/ 7 The p-value for the uk-ru pair is 0.06 we do not consider this result as statistically significant.",
"acters, because the used comparable corpora consist of related languages that share a similar orthography.",
"However, the lists based on the frequency window heuristic show poor results to include the correct translations during the extraction step.",
"Our ML models based on similarity metrics over parallel corpora show rankings similar to heldout data.",
"However, we created our training data using simple heuristics that simulate cognate words (i.e.",
"pairs of words with a small surface difference).",
"The ML models are able to rank similar words on the top of the list and they give a reliable score to discriminate wrong translations.",
"Preliminary results on the addition of the n-best lists into an SMT system show modest improvements compare to the baseline.",
"However, the OOV rate shows improvements around 10% reduction on word types, because of the wide variety of lexical choices introduced by the MT dictionaries.",
"Future work involves the addition of unsupervised morphology features for the n-best list extraction, i.e.",
"first step, given that the use of starting characters shows to be an effective heuristic to prune the search space and language independent.",
"Finally, we will measure the contribution for all the produced cognate lists, where we can try different strategies to add the dictionaries into an SMT system (Irvine and Callison-Burch, 2014) ."
]
} | {
"paper_header_number": [
"1",
"2",
"2.1",
"2.2",
"3",
"3.1",
"3.2",
"3.3",
"4"
],
"paper_header_content": [
"Introduction",
"Methodology",
"Cognate detection",
"Cognate Ranking",
"Results and Discussion",
"Data",
"Evaluation of the Ranking Model",
"Preliminary Results on Comparable Corpora",
"Conclusions and future Work"
]
} | GEM-SciDuet-train-8#paper-972#slide-8 | Results on heldout data | Error score on heldout data
E Edit distance features
EC Edit distance plus distributed vectors features
Zoo error% Opensubs error% Wiki-titles error%
Romance pt-es it-es fr-es
Model E Model EC Model E Model EC Model E Model EC | Error score on heldout data
E Edit distance features
EC Edit distance plus distributed vectors features
Zoo error% Opensubs error% Wiki-titles error%
Romance pt-es it-es fr-es
Model E Model EC Model E Model EC Model E Model EC | [] |