{ "paper_id": "2020", "header": { "generated_with": "S2ORC 1.0.0", "date_generated": "2023-01-19T01:08:44.980287Z" }, "title": "Unsupervised Distillation of Syntactic Information from Contextualized Word Representations", "authors": [ { "first": "Shauli", "middle": [], "last": "Ravfogel", "suffix": "", "affiliation": { "laboratory": "", "institution": "Bar Ilan University", "location": {} }, "email": "shauli.ravfogel@gmail.com" }, { "first": "Yanai", "middle": [], "last": "Elazar", "suffix": "", "affiliation": { "laboratory": "", "institution": "Bar Ilan University", "location": {} }, "email": "yanaiela@gmail.com" }, { "first": "Jacob", "middle": [], "last": "Goldberger", "suffix": "", "affiliation": { "laboratory": "", "institution": "Bar Ilan University", "location": {} }, "email": "jacob.goldberger@biu.ac.il" }, { "first": "Yoav", "middle": [], "last": "Goldberg", "suffix": "", "affiliation": { "laboratory": "", "institution": "Bar Ilan University", "location": {} }, "email": "yoav.goldberg@gmail.com" } ], "year": "", "venue": null, "identifiers": {}, "abstract": "Contextualized word representations, such as ELMo and BERT, were shown to perform well on various semantic and syntactic task. In this work, we tackle the task of unsupervised disentanglement between semantics and structure in neural language representations: we aim to learn a transformation of the contextualized vectors, that discards the lexical semantics, but keeps the structural information. To this end, we automatically generate groups of sentences which are structurally similar but semantically different, and use metric-learning approach to learn a transformation that emphasizes the structural component that is encoded in the vectors. We demonstrate that our transformation clusters vectors in space by structural properties, rather than by lexical semantics. Finally, we demonstrate the utility of our distilled representations by showing that they outperform the original contextualized representations in a few-shot parsing setting. * Equal contribution 1 In this work we focus on English.", "pdf_parse": { "paper_id": "2020", "_pdf_hash": "", "abstract": [ { "text": "Contextualized word representations, such as ELMo and BERT, were shown to perform well on various semantic and syntactic task. In this work, we tackle the task of unsupervised disentanglement between semantics and structure in neural language representations: we aim to learn a transformation of the contextualized vectors, that discards the lexical semantics, but keeps the structural information. To this end, we automatically generate groups of sentences which are structurally similar but semantically different, and use metric-learning approach to learn a transformation that emphasizes the structural component that is encoded in the vectors. We demonstrate that our transformation clusters vectors in space by structural properties, rather than by lexical semantics. Finally, we demonstrate the utility of our distilled representations by showing that they outperform the original contextualized representations in a few-shot parsing setting. * Equal contribution 1 In this work we focus on English.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Abstract", "sec_num": null } ], "body_text": [ { "text": "Human language 1 is a complex system, involving an intricate interplay between meaning (semantics) and structural rules between words and phrases (syntax). Self-supervised neural sequence models for text trained with a language modeling objective, such as ELMo (Peters et al., 2018) , BERT (Devlin et al., 2019) , and RoBERTA (Liu et al., 2019b) , were shown to produce representations that excel in recovering both structure-related information (Gulordava et al., 2018; van Schijndel and Linzen, 2018; Wilcox et al., 2018; Goldberg, 2019) as well as in semantic information (Yang et al., 2019; Joshi et al., 2019) .", "cite_spans": [ { "start": 261, "end": 282, "text": "(Peters et al., 2018)", "ref_id": "BIBREF37" }, { "start": 290, "end": 311, "text": "(Devlin et al., 2019)", "ref_id": "BIBREF4" }, { "start": 326, "end": 345, "text": "(Liu et al., 2019b)", "ref_id": null }, { "start": 446, "end": 470, "text": "(Gulordava et al., 2018;", "ref_id": "BIBREF11" }, { "start": 471, "end": 502, "text": "van Schijndel and Linzen, 2018;", "ref_id": "BIBREF46" }, { "start": 503, "end": 523, "text": "Wilcox et al., 2018;", "ref_id": "BIBREF47" }, { "start": 524, "end": 539, "text": "Goldberg, 2019)", "ref_id": "BIBREF10" }, { "start": 575, "end": 594, "text": "(Yang et al., 2019;", "ref_id": "BIBREF48" }, { "start": 595, "end": 614, "text": "Joshi et al., 2019)", "ref_id": "BIBREF20" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "In this work, we study the problem of disentangling structure from semantics in neural language Pairs of words are represented by the difference between their transformation f , which is identical for all words. The pairs of words in the anchor and positive sentences are lexically different, but structurally similar. The negative example presented here is especially challenging, as it is lexically similar, but structurally different.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "representations: we aim to extract representations that capture the structural function of words and sentences, but which are not sensitive to their content. For example, consider the sentences:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "1. Neural networks are interesting.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "2. I study neural networks.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "3. Maple syrup is delicious.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "While (1) and (3) are different in content, they share a similar structure, the corresponding words in them, while unrelated in meaning, 2 serve the same function. Similarly for sentences (2) and (4). In contrast, sentence (1) shares the phrase neural networks with sentence (2), and maple syrup is shared between (3) and (4). 3 While the two occurrences of each phrase share the meaning, they are used in different structural (syntactic) configurations, serving different roles within the sentence (appearing in subject vs object position). 4 We seek a representation that will expose the similarity between \"networks\" in (1) and \"syrup\" in (2), while ignoring the similarity between \"syrup\" in (2) and \"syrup\" in (4).", "cite_spans": [ { "start": 327, "end": 328, "text": "3", "ref_id": null }, { "start": 542, "end": 543, "text": "4", "ref_id": null } ], "ref_spans": [], "eq_spans": [], "section": "John loves maple syrup.", "sec_num": "4." }, { "text": "We seek a function from contextualized word representations to a space that exposes these similarities. Crucially, we aim to do this in an unsupervised manner: we do not want to inform the process of the kind of structural information we want to obtain. We do this by learning a transformation that attempts to remove the lexical-semantic information in a sentence, while trying to preserve structural properties.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "John loves maple syrup.", "sec_num": "4." }, { "text": "Disentangling syntax from lexical semantics in word representations is a desired property for several reasons. From a purely scientific perspective, once disentanglement is achieved, one can better control for confounding factors and analyze the knowledge the model acquires, e.g. attributing the predictions of the model to one factor of variation while controlling for the other. In addition to explaining model predictions, such disentanglement can be useful for the comparison of the representations the model acquires to linguistic knowledge. From a more practical perspective, disentanglement can be a first step toward controlled generation/paraphrasing that considers only aspects of the structure, akin to the style-transfer works in computer vision, i.e., rewriting a sentence while preserving its structural properties while ignoring its meaning, or vice-versa. It can also inform searchbased application in which one can search for \"similar\" texts while controlling various aspects of the desired similarity.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "John loves maple syrup.", "sec_num": "4." }, { "text": "To achieve this goal, we begin with the intuition that the structural component in the representation (capturing the form) should remain the same regardless of the lexical semantics of the sentence (the meaning). Rather than beginning with a parsed corpus, we automatically generate a large number of structurally-similar sentences, without presupposing their formal structure ( \u00a73.1). This allows us to pose the disentanglement problem as a metriclearning problem: we aim to learn a transformation of the contextualized representation, which is invariant to changes in the lexical semantics within each group of structurally-similar sentences ( \u00a73.3). We demonstrate the structural properties captured by the resulting representations in multiple experiments ( \u00a74), among them automatic identification of structurally-similar words and few-shot parsing.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "John loves maple syrup.", "sec_num": "4." }, { "text": "We release our code at https://github.com/ shauli-ravfogel/NeuralDecomposition.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "John loves maple syrup.", "sec_num": "4." }, { "text": "The problem of disentangling different sources of variation has long been studied in computer vision, and was recently applied to neural models (Bengio et al., 2013; Mathieu et al., 2016; Hadad et al., 2018) . Such disentanglement can assist in learning representations that are invariant to specific factors, such as pose-invariant face-recognition (Peng et al., 2017) or style-invariant digit recognition (Narayanaswamy et al., 2017) . From a generative point of view, disentanglement can be used to modify one aspect of the input (e.g., \"style\"), while keeping the other factors (e.g., \"content\") intact, as done in neural image style-transfer (Gatys, 2017).", "cite_spans": [ { "start": 144, "end": 165, "text": "(Bengio et al., 2013;", "ref_id": "BIBREF2" }, { "start": 166, "end": 187, "text": "Mathieu et al., 2016;", "ref_id": "BIBREF30" }, { "start": 188, "end": 207, "text": "Hadad et al., 2018)", "ref_id": "BIBREF12" }, { "start": 350, "end": 369, "text": "(Peng et al., 2017)", "ref_id": "BIBREF35" }, { "start": 407, "end": 435, "text": "(Narayanaswamy et al., 2017)", "ref_id": "BIBREF34" } ], "ref_spans": [], "eq_spans": [], "section": "Related Work", "sec_num": "2" }, { "text": "In NLP, disentanglement is much less researched. In controlled natural language generation and style transfer, several works attempted to disentangle factors of variation such as sentiment or age of the writer, with the intention to control for those factors and generate new sentences with specific properties (Sohn et al., 2015; Ficler and Goldberg, 2017; , or transfer existing sentences to similar sentences that differ only in the those properties. The latter goal of style transfer is often realized by learning representations which are invariant to the controlled attributes (Fu et al., 2018; Hu et al., 2017) .", "cite_spans": [ { "start": 311, "end": 330, "text": "(Sohn et al., 2015;", "ref_id": "BIBREF42" }, { "start": 331, "end": 357, "text": "Ficler and Goldberg, 2017;", "ref_id": "BIBREF7" }, { "start": 583, "end": 600, "text": "(Fu et al., 2018;", "ref_id": "BIBREF8" }, { "start": 601, "end": 617, "text": "Hu et al., 2017)", "ref_id": "BIBREF18" } ], "ref_spans": [], "eq_spans": [], "section": "Related Work", "sec_num": "2" }, { "text": "Another main line of work which is relevant to our approach is that of probing. The concept, originally introduced by Adi et al. (2016) and Hupkes et al. (2018) , relies on training classifiers (probes) to expose symbolic linguistic information that is encoded in the model. A large body of works have shown sensitivity to both semantic (Tenney et al., 2019a; Richardson et al., 2019) and syntactic (Tenney et al., 2019b; Reif et al., 2019; Hewitt and Manning, 2019; Liu et al., 2019a) information. Hewitt and Manning (2019) demonstrated that it is possible to train a linear transformation, under which squared euclidean distance between transformed contextualized word vectors correspond to the distances between the respective words in the syntactic tree. Li and Eisner (2019) have used a variational estimation method (Alemi et al., 2016) of the information-bottleneck principle (Tishby et al., 1999) to extract word embeddings that are useful to the end task of parsing.", "cite_spans": [ { "start": 118, "end": 135, "text": "Adi et al. (2016)", "ref_id": "BIBREF0" }, { "start": 140, "end": 160, "text": "Hupkes et al. (2018)", "ref_id": "BIBREF19" }, { "start": 337, "end": 359, "text": "(Tenney et al., 2019a;", "ref_id": "BIBREF43" }, { "start": 360, "end": 384, "text": "Richardson et al., 2019)", "ref_id": "BIBREF40" }, { "start": 399, "end": 421, "text": "(Tenney et al., 2019b;", "ref_id": "BIBREF44" }, { "start": 422, "end": 440, "text": "Reif et al., 2019;", "ref_id": "BIBREF39" }, { "start": 441, "end": 466, "text": "Hewitt and Manning, 2019;", "ref_id": "BIBREF14" }, { "start": 467, "end": 485, "text": "Liu et al., 2019a)", "ref_id": "BIBREF27" }, { "start": 499, "end": 524, "text": "Hewitt and Manning (2019)", "ref_id": "BIBREF14" }, { "start": 759, "end": 779, "text": "Li and Eisner (2019)", "ref_id": "BIBREF25" }, { "start": 822, "end": 842, "text": "(Alemi et al., 2016)", "ref_id": "BIBREF1" }, { "start": 883, "end": 904, "text": "(Tishby et al., 1999)", "ref_id": "BIBREF45" } ], "ref_spans": [], "eq_spans": [], "section": "Related Work", "sec_num": "2" }, { "text": "While impressive, those works presuppose a specific syntactic structure (e.g. annotated parse tree) and use this linguistic signal to learn the probe in a supervised manner. This approach can introduce confounding between extracting information and learning it by the probe (Hewitt and Liang, 2019; Ravichander et al., 2020; Maudslay et al., 2020; Elazar et al., 2020) . In contrast, we aim to expose the structural information encoded in the network in an unsupervised manner, without pre-supposing an existing syntactic annotation scheme.", "cite_spans": [ { "start": 274, "end": 298, "text": "(Hewitt and Liang, 2019;", "ref_id": "BIBREF13" }, { "start": 299, "end": 324, "text": "Ravichander et al., 2020;", "ref_id": "BIBREF38" }, { "start": 325, "end": 347, "text": "Maudslay et al., 2020;", "ref_id": null }, { "start": 348, "end": 368, "text": "Elazar et al., 2020)", "ref_id": "BIBREF6" } ], "ref_spans": [], "eq_spans": [], "section": "Related Work", "sec_num": "2" }, { "text": "Our goal is to learn a function f : R n \u2192 R m , which operates on contextualized word representations x and extracts vectors f (x) which make the structural information encoded in x more salient, while discarding as much lexical information as possible. In the sentences \"Maple syrup is delicious\" and \"Neural networks are interesting\", we want to learn a function f such that f (", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Method", "sec_num": "3" }, { "text": "v 3 syrup ) \u2248 f (v 1 networks ), where v i", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Method", "sec_num": "3" }, { "text": "word is the contextualized vector representation of the word in sentence i. We also want f (", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Method", "sec_num": "3" }, { "text": "v 4 syrup ) \u2248 f (v 2 networks ), while keeping f (v 1 networks ) \u2248 f (v 2 networks )", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Method", "sec_num": "3" }, { "text": ". Moreover, we would like the relation between the words \"maple\" and \"delicious\" in the third sentence, to be similar to the relation between \"neural\" and \"interesting\" in the first sentence:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Method", "sec_num": "3" }, { "text": "pair(v 3 maple , v 3 delicious ) \u2248 pair(v 1 neural , v 1 interesting )", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Method", "sec_num": "3" }, { "text": ". Operatively, we represent pairs of words (x, y) by the difference between their transformation f (x) \u2212 f (y), and aim to learn a function f that preserves:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Method", "sec_num": "3" }, { "text": "f (v 3 maple )\u2212f (v 3 delicious ) \u2248 f (v 1 neural )\u2212f (v 1 interesting )", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Method", "sec_num": "3" }, { "text": ". The choice to represent pairs this way was inspired by several works that demonstrated that nontrivial semantic and syntactic relations between uncontextualized word representations can be approximated by simple vector arithmetic (Mikolov et al., 2013a,b; Levy and Goldberg, 2014) .", "cite_spans": [ { "start": 232, "end": 257, "text": "(Mikolov et al., 2013a,b;", "ref_id": null }, { "start": 258, "end": 282, "text": "Levy and Goldberg, 2014)", "ref_id": "BIBREF24" } ], "ref_spans": [], "eq_spans": [], "section": "Method", "sec_num": "3" }, { "text": "To learn f , we start with groups of sentences such that the sentences within each group are known to share structure but differ in lexical semantics. We call the sentences in each group structurally equivalent. Figure 2 shows an example of two structurally equivalent sets. Acquiring such sets is challenging, especially if we do not assume a known syntactic formalism and cannot mine for sentences based on their observed tree structures.", "cite_spans": [], "ref_spans": [ { "start": 212, "end": 220, "text": "Figure 2", "ref_id": null } ], "eq_spans": [], "section": "Method", "sec_num": "3" }, { "text": "To this end, we automatically generate the sets starting with known sentences and sampling variants from a language model ( \u00a73.1). Our sentence-set generation procedure ensures that words from the same set that share an index also share their structural function. We call such words corresponding.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Method", "sec_num": "3" }, { "text": "We now proceed to learn a function f to map contextualized vectors of corresponding words (and the relations between them, as described above) to neighbouring points in the space.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Method", "sec_num": "3" }, { "text": "We train f such that the representation assigned to positive pairs -pairs that share indices and come from the same equivalent set -is distinguished from the representations of negative pairs -challenging pairs that come from different sentences, and thus do not share the structure of the original pair, but can, potentially, share their lexical meaning. We do so using Triplet loss, which pushes the representations of pairs coming from the same group closer together ( \u00a73.3). Figure 1 sketches the network.", "cite_spans": [], "ref_spans": [ { "start": 479, "end": 487, "text": "Figure 1", "ref_id": "FIGREF0" } ], "eq_spans": [], "section": "Method", "sec_num": "3" }, { "text": "In order to generate sentences that approximately share their structure, we sequentially replace content words in the sentence with other content words, while aiming to maintain the grammatically of the sentence, and keep its structure intact. Since we do not want to rely on syntactic annotation when performing this replacement, we opted to use a pre-trained language model -BERT -under the assumption that strong neural language models do implicitly encode many of the syntactic restrictions that apply to words in different grammatical functions (e.g., we assume that BERT would not predict a transitive verb in the place of an intransitive verb, or a verb that accepts a complement in the place of a verb that does not accept a complement). While this assumption seems to hold with regard to basic distinctions such as transitive vs. intransitive verbs, its validity is less clear in the more nuanced cases, in which small differences in the surface level can translate to substantial differences in abstract syntactic structure -such as replacing a control verb with a raising verb. This is a limitation of the current approach, although we find that the average sentence we generate is grammatical and similar in structure to the original sentence. Moreover, as our goal is to expose the structural similarity encoded in neural language models, we Figure 2 : Two groups of structurally-equivalent sentences. In each group, the first sentence is original sentence from Wikipedia, and the sentences below it were generated by the process of repeated BERT substitution. Some sets of corresponding words-that is, words that share the same structural function-are highlighted in the same color.", "cite_spans": [], "ref_spans": [ { "start": 1355, "end": 1363, "text": "Figure 2", "ref_id": null } ], "eq_spans": [], "section": "Generating Structurally-similar Sentences", "sec_num": "3.1" }, { "text": "find it reasonable to only capture the distinctions that are captured by modern language models.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Generating Structurally-similar Sentences", "sec_num": "3.1" }, { "text": "Implementation We start each group with a Wikipedia sentence, for which we generate k = 6 equivalent sentences by iterating over the sentence from left to right sequentially, masking the ith word, and replacing it with one of BERT's top-30 predictions. To increase semantic variability, we perform the replacement in place (online): after randomly choosing a guess w, we insert w to the sentence at index i, and continue guessing the i + 1 word based on the modified sentence. 5 We exclude a closed set of a few dozens of words (mostly function words) and keep them unchanged in all k variations of a sentence. We further maintain structural correctness by maintaining the POS 6 , and encourage semantic diversity by the auto-regressive replacement process. In Table 6 in the Appendix we show some additional generated groups. The sets in Figure 2 were generated using this method.", "cite_spans": [ { "start": 477, "end": 478, "text": "5", "ref_id": null } ], "ref_spans": [ { "start": 761, "end": 768, "text": "Table 6", "ref_id": "TABREF1" }, { "start": 839, "end": 847, "text": "Figure 2", "ref_id": null } ], "eq_spans": [], "section": "Generating Structurally-similar Sentences", "sec_num": "3.1" }, { "text": "We sample N = 150, 000 random sentences and use the our method to generate 900, 000 equivalent sets E of structurally equivalent sentences. Then, we encode the sentences and randomly collect 1, 500, 000 contextualized vector representations of words from these sets, resulting in 1,500,000 training pairs and 200,000 evaluation pairs for the training process of f . We experiment with both ELMo and BERT language models. In average, we sample 11 word-pairs from each group of equivalent sentences. For ELMo, we represent each word in context as a concatenation of the last two ELMo layers (excluding the word embedding layer, which is not contextualized and therefore irrelevant for structure), resulting in representations of dimension 2048. For BERT, we concatenate the mean of the words' representation 7 across all contextualized layers of BERT-Large, with the representation of layer 16, which was found by Hewitt and Manning (2019) most indicative of syntax.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Word Representation", "sec_num": "3.2" }, { "text": "We learn the mapping function f using triplet loss (Figure 1 ). Given a group of equivalent sentences E i , we randomly choose two sentences to be the anchor sentence S A and the positive sentence S P , and sample two different word indices", "cite_spans": [], "ref_spans": [ { "start": 51, "end": 60, "text": "(Figure 1", "ref_id": "FIGREF0" } ], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "{i 1 , i 2 }. Let S A [i 1 ] be the contextualized representation of the i 1 th word in sentence S A . The words S A [i 1 ] and S A [i 2 ]", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "from the anchor sentence would form a representation of a pair of words, which should be close to the pair S P [i 1 ], S P [i 2 ] from the positive sentence.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "We represent pairs as their differences after transformation, resulting in the anchor pair V A and positive pair V P :", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "V A = f (S A [i 1 ]) \u2212 f (S A [i 2 ]) S A \u2208 E i (1) V P = f (S P [i 1 ]) \u2212 f (S P [i 2 ]) S P \u2208 E i (2)", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "where f is the parameterized syntactic transformation we aim to learn. We also consider a negative pair:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "V N = f (S N [j 1 ]) \u2212 f (S N [j 2 ]) S N \u2208 E i (3)", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "coming from sentence S N which is not in the equivalent set.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "As f has shared parameters for both words in the pair, it can be considered a part of a Siamese network, making our learning procedure an instance of a triplet Siamese network (Schroff et al., 2015) . We choose f to be a simple model: a single linear layer that maps from dimensionality 2048 to 75.", "cite_spans": [ { "start": 176, "end": 198, "text": "(Schroff et al., 2015)", "ref_id": "BIBREF41" } ], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "The dimensions of the transformation were chosen according to development set performance.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "We use triplet loss (Schroff et al., 2015) to move the representation of the anchor vector V A closer to the representation of the positive vector V P and farther apart from the representation of the negative vector V N . Following Hoffer and Ailon (2015), we calculate the softmax version of the triplet loss:", "cite_spans": [ { "start": 20, "end": 42, "text": "(Schroff et al., 2015)", "ref_id": "BIBREF41" } ], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "L triplet (V A , V P , V N ) = e d(V A ,V P ) e d(V A ,V P ) + e d(V A ,V N ) (4) where d(x, y) = 1 \u2212 x y", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "x y is the cosinedistance between the vectors x and y. Note that", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "L triplet \u2192 0 as d(V A ,V P ) d(V A ,V N ) \u2192 0, as expected.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "The triplet objective is optimized end-to-end using the Adam optimizer (Kingma and Ba, 2015). We train for 5 epochs with a mini-batch of size 500 8 , and take the last model as the final syntactic extractor. During training, the gradient backpropagates through the pair vectors to the parameters f of the Siamese model, to get representations of individual words that are similar for corresponding words in equivalent sentences. We note that we do not back-propagate the gradient to the contextualized vectors: we keep them intact, and only adjust the learned transformation.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "Hard negative sampling We obtain the negative vectors V N using hard negative sampling. For each mini-batch B, we collect 500 {V A i , V P i } pairs, each pair taken from an equivalent set E i . The negative instances V N i are obtained by searching the batch for a vector that is closest to the anchor and comes from a different set:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "V N i = arg min V A j =i \u2208B d(V A i , V A j ).", "eq_num": "(5)" } ], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "In addition, we enforce a symmetry between the anchor and positive vectors, by adding a pair (positive, anchor) for each pair (anchor, positive) in B. That is, V N i is the \"most misleading\" word-pair vector: it comes from a sentence that has a different structure than the structure of V A i sentence, but is the closest to V A i in the mini-batch.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Triplet Loss", "sec_num": "3.3" }, { "text": "We have trained the syntactic transformation f in a way that should encourage it to retain the structural information encoded in contextualized vectors, but discard other information. We assess the representations the model acquired in an unsupervised manner, by evaluating the extent to which the local neighbors of each transformed contextualized vector f (x) share known structural properties, such as grammatical function within the sentence. For the baseline, we expect the neighbors of each vector to share a mix of semantic and syntactic properties. For the transformed vectors, we expect the neighbors to share mainly syntactic properties. Finally, we demonstrate that in a few-shot setting, our representations outperform the original ELMO representation, indicating they are indeed distilled from syntax, and discard other information that is encoded in ELMO vectors but is irrelevant for the extraction of the structure of a sentence.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experiments and Analysis", "sec_num": "4" }, { "text": "Corpus For training the transformation f , we rely on 150,000 sentences from Wikipedia, tokenized and POS-tagged by spaCy (Honnibal and Johnson, 2015; Honnibal and Montani, 2017) . The POS tags are used in the equivalent set generation to filter replacement words. Apart from POS tagging, we do not rely on any syntactic annotation during training. The evaluation sentences for the experiments mentioned below are sampled from a collection of 1,000,000 original and unmodified Wikipedia sentences (different from those used in the model training). Figure 3 shows a 2dimensional t-SNE projection (Maaten and Hinton, 2008) of 15,000 random content words. The left panel projects the original ELMo states, while the right panel is the syntactically transformed ones. The points are colored according to the dependency label (relation to parent) of the corresponding word, predicted by the parser.", "cite_spans": [ { "start": 122, "end": 150, "text": "(Honnibal and Johnson, 2015;", "ref_id": "BIBREF16" }, { "start": 151, "end": 178, "text": "Honnibal and Montani, 2017)", "ref_id": "BIBREF17" } ], "ref_spans": [ { "start": 548, "end": 556, "text": "Figure 3", "ref_id": "FIGREF1" } ], "eq_spans": [], "section": "Experiments and Analysis", "sec_num": "4" }, { "text": "In the original ELMo representation most states -apart from those characterized by a specific partof-speech, such as amod (adjectives, in orange) or nummod (numbers, in light green) -do not fit well into a single cluster. In contrast, the syntactically transformed vectors are more neatly clustered, with some clusters, such as direct objects (brown) and prepositional-objects (blue), that are relatively separated after, but not before, the transformation. Interestingly, some functions that used to be a single group in ELMo (like the adjectives in orange, or the noun-compounds in green) are Type Text Q1 in this way of thinking, an impacting projectile goes into an ice-rich layer -but no further. N they generally have a pre-engraved rifling band to engage the rifled launch tube, spin-stabilizing the projectile, hence the term \"rifle\". NT to achieve a large explosive yield, a linear implosion weapon needs more material, about 13 kgs.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Qualitative Analysis t-SNE Visualization", "sec_num": "4.1" }, { "text": "Q2 the mint's director at the time, nicolas peinado, was also an architect and made the initial plans. N the director is angry at crazy loop and glares at him, even trying to get a woman to kick crazy loop out of the show (which goes unsuccessfully). NT jetley's mother, kaushaliya rani, was the daughter of high court advocate shivram jhingan.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Qualitative Analysis t-SNE Visualization", "sec_num": "4.1" }, { "text": "Q3 their first project is software that lets players connect the company's controller to their device. N you could try use norton safe web, which lets you enter a website and show whether there seems to be anything bad in it. NT the city offers a route-finding website that allows users to map personalized bike routes. Table 1 : Text examples for a few query words (in the Q rows, in bold), and their closest neighbours before (N) and after (NT) the transformation. now split into several clusters, corresponding to their use in different sentence positions, separating for examples adjectives that are used in subject positions from those in object position or within prepositional phrases. Additionally, as noun compounds (\"maple\" in \"maple syrup\") and adjectival modifiers (\"tasty\" in \"tasty syrup\") are relatively structurally similar (they appear between determiners and nouns within noun phrases, and can move with the noun phrase to different positions), they are split and grouped together in the representation (the green and orange clouds).", "cite_spans": [], "ref_spans": [ { "start": 320, "end": 327, "text": "Table 1", "ref_id": null } ], "eq_spans": [], "section": "Qualitative Analysis t-SNE Visualization", "sec_num": "4.1" }, { "text": "To quantify the difference, we run K-means clustering on the projected vectors, and calculate the average cluster purity score as the relative proportion of the most common dependency label in each cluster. The higher this value is, the more the division to clusters reflect division to grammatical functions (dependency labels). We run the clustering with different K values: 10, 20, 40, 80. We find an increase in class purity following our transformation: from scores of 22.6%, 26.8%, 32.6% and 36.4% (respectively) for the original vectors, to scores of 24.3%, 33.4%, 42.1% and 48.0% (respectively) for the transformed vectors.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Qualitative Analysis t-SNE Visualization", "sec_num": "4.1" }, { "text": "Examples In Table 1 we present a few query words (Q) and their closest neighbours before (N) and after (NT) the transformation. Note the high structural similarity of the entire sentence, as well as the function of the word within it (Q1: last word of subject NP in a middle clause, Q2: possessed noun in sentence initial subject NP, Q3: head of relative clause of a direct object).", "cite_spans": [], "ref_spans": [ { "start": 12, "end": 19, "text": "Table 1", "ref_id": null } ], "eq_spans": [], "section": "Qualitative Analysis t-SNE Visualization", "sec_num": "4.1" }, { "text": "Additional examples (including cases in which the retrieved vector does not share the dependency edge with the query vector) are supplied in Appendix \u00a7A.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Qualitative Analysis t-SNE Visualization", "sec_num": "4.1" }, { "text": "We expect the transformed vectors to capture more structural and less lexical similarities than the source vectors. We expect each vectors' neighbors in space to share the structural function of the word over which the vector was collected, but not necessarily share its lexical meaning. We focus on the following structural properties: (1) Dependencytree edge of a given word (dep-edge), that represents its function (subject, object etc.). (2) The dependency edge of the word parent's (head's depedge) in the tree -to represent higher level structure, such as a subject that resides within a relative clause, as in the word \"man\" in the phrase \"the child that the man saw\". (3) Depth in the dependency tree (distance from the root of the sentence tree). (4) Constituency-parse paths: consider, for example, the sentence \"They saw the moon with the telescope\". The word \"telescope\" is a part of a Table 2 : Closest-word queries, before and after the application of the syntactic transformation. \"Basline\" refers to unmodified ELMo vectors, \"Transformed\" refers to ELMo vectors after the learned syntactic transformation f , and \"Transformed-untrained\" refers to ElMo vectors, after a transformation that was trained on a randomly-initialized ELMo. \"hard\" denotes results on the subset of POS tags which are most structurally diverse.", "cite_spans": [], "ref_spans": [ { "start": 898, "end": 905, "text": "Table 2", "ref_id": null } ], "eq_spans": [], "section": "Quantitative Evaluation", "sec_num": "4.2" }, { "text": "noun-phrase \"the telescope\", which resides inside a prepositional phrase \"with the telescope\", which is part of the Verbal phrase \"saw with the telescope\". The complete constituency path for this word is therefore \"NP-PP-VP\". We calculate the complete tree path to the root (Tree-path-complete), as well as paths limited to lengths 2 and 3. For this evaluation, we parse 400,000 random sentences taken from the 1-million-sentences Wikipedia sample, run ELMo and BERT to collect the contextualized representations of the sentences, and randomly choose 400,000 query word vectors (excluding function words). We then retrieve, for each query vector x, the value vector y that is closest to x in cosine-distance, and record the percentage of closest-vector pairs (x, y) that share each of the structural properties listed above. For the tree depth property, we calculate the Pearson correlation between the depths of the queries and the retrieved values. We use the Berkeley Neural Parser (Kitaev and Klein, 2018) for constituency parsing. We exclude function words from the evaluation.", "cite_spans": [ { "start": 985, "end": 1009, "text": "(Kitaev and Klein, 2018)", "ref_id": "BIBREF22" } ], "ref_spans": [], "eq_spans": [], "section": "Quantitative Evaluation", "sec_num": "4.2" }, { "text": "Easier and Harder cases The baseline models tend to retrieve words that are lexically similar. Since certain words tend to appear at above-chance probability in certain structural functions, this can make the baseline be \"right for the wrong reason\", as the success in the closest-word test reflects lexical similarity, rather than grammatical generalization. To control for this confounding, we sort the different POS tags according to the entropy of their dependency-labels distribution, and repeat the evaluation only for words belonging to those POS tags having the highest entropy (those are the most structurally variant, and tend to appear in different structural functions). The performance of the baselines (ELMo, BERT models) on those words drops significantly, while the performance of our model is only mildly influenced, indicating the superiority of the model in capturing structural rather than lexical information.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Quantitative Evaluation", "sec_num": "4.2" }, { "text": "Results The results for ELMo are presented in Table 2 . For BERT, we witnessed similar, but somewhat lower, accuracy: for example, 68.1% dependency-edge accuracy, 56.5% head's dependency-edge accuracy, and 22.1% complete constituency-path accuracy. The results for BERT are available in Appendix \u00a7B, and for the reminder of the paper, we focus in ELMo. We observe significant improvement over the baseline for all tests. The correlation between the depth in tree of the query and the value words, for examples, rises from 44.8% to 56.1%, indicating that our model encourages the structural property of the depth of the word to be more saliently encoded in its representation compared with the baseline. The most notable relative improvement is recorded with regard to full constituency-path to the root: from 16.6% before the structural transformation, to 25.3% after it -an improvement of 52%. In addition to the increase in syntax-related properties, we observe a sharp drop -from 73.6% to 28.4% -in the proportion of query-value pairs that are lexically identical (lexical match, Table 2 ). This indicates our transformation f removes much of the lexical information, which is irrelevant for structure. To assess to what extent the improvements stems from the information encoded in ELMo, rather than being an artifact of the triplet-loss training, we also evaluate on a transformation f that was trained on a randomlyinitialized ELMo, a surprisingly strong baseline (Conneau et al., 2018) . We find this model performs substantially worse than the baseline (Table 2 , \"Transformed-untrained (all)\").", "cite_spans": [ { "start": 1470, "end": 1492, "text": "(Conneau et al., 2018)", "ref_id": "BIBREF3" } ], "ref_spans": [ { "start": 46, "end": 53, "text": "Table 2", "ref_id": null }, { "start": 1083, "end": 1090, "text": "Table 2", "ref_id": null }, { "start": 1561, "end": 1570, "text": "(Table 2", "ref_id": null } ], "eq_spans": [], "section": "Quantitative Evaluation", "sec_num": "4.2" }, { "text": "Distillation: Few-Shot Parsing", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Minimal Supervision for Structure", "sec_num": "4.3" }, { "text": "The absolute nearest-neighbour accuracy values may appear to be relatively low: for example, only 67.6% of the (query, value) pairs share the same dependency edge.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Minimal Supervision for Structure", "sec_num": "4.3" }, { "text": "As the model acquires its representation without being exposed to human-mandated syntactic convention, some of the apparent discrepancies in nearest neighbours may be due to the fact the model acquires different kind of generalization, or learned a representation that emphasizes different kinds of similarities. Still, we expect the resulting (75 dimensional) representations to contain distilled structure information that is mappable to human notions of syntax. To test this, we compare dependency-parsers trained on our representation and on the source representation. If our representation indeed captures structural information, we expect it to excel on a low data setting. To this end, we test our hypothesis with few-shot dependency parsing setup, where we train a model to predict syntactic trees representation with only a few hundred labeled examples.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Minimal Supervision for Structure", "sec_num": "4.3" }, { "text": "We use an off-the-shelf dependency parser model (Dozat and Manning, 2016) and swap the pre-trained Glove embeddings (Pennington et al., 2014) with ELMo contextualized embeddings (Peters et al., 2018) . In order to have a fair comparison with our method, we use the concatenation of the two last layers of Elmo; we refer to this experiment as elmo. As our representation is much smaller than ELMo's (75 as opposed to 2048), a potential issue for a low data setting is the higher number of parameters to optimize in the later case, therefore a lower dimension may achieve better results. We design two additional baselines to remedy this potential issue: (1) Using PCA in order to reduce the representation dimensionality. We randomly chose 1M words from Wikipedia, calculated their representation with ELMo embeddings and performed PCA. This transformation is applied during training on top of ELMo representation while keeping the 75 first components. This experiment is referred to as elmo-pca. This representation should perform well if the most salient information in the ELMo representations are structural. We exepct it to not be the case. (2) Automatically learning a matrix that reduces the embedding dimension. This matrix is learned during training and can potentially extract the relevant structural information from the representations. We refer to this experiment as elmo-reduced. Additionally, we also compare to a baseline where we use the gold-POS labels as the sole input to the model, by initializing an embedding matrix of the same size for each POS. We refer to this experiment as pos. Lastly, we examine the performance of our representation, where we apply our structural extraction method on top of ELMo representation. We refer to this experiment as syntax.", "cite_spans": [ { "start": 48, "end": 73, "text": "(Dozat and Manning, 2016)", "ref_id": "BIBREF5" }, { "start": 116, "end": 141, "text": "(Pennington et al., 2014)", "ref_id": "BIBREF36" }, { "start": 178, "end": 199, "text": "(Peters et al., 2018)", "ref_id": "BIBREF37" } ], "ref_spans": [], "eq_spans": [], "section": "Minimal Supervision for Structure", "sec_num": "4.3" }, { "text": "We run the few-shot setup with multiple training size values: 50, 100, 200, 500. The results-for both labeled (LAS) and unlabeled (UAS) attachment scores-are presented in Figure 4 , and the numerical results are available in the Appendix \u00a7C. In the lower training size setting, we obtain the best performances compared to all baselines. The more training data is used, the gap between our representation and the baselines reduced, but the syntax representation still outperforms elmo. Using gold POS labels as inputs works relatively well with 50 training examples, but it quickly reaches a plato in performance and remains behind the other baselines. Reducing the dimensions with PCA (elmo-pca) works considerably worse than ELMo, indicating PCA loses important information. Reducing the dimensions with a learned matrix (elmo-reduced) works substantially better than ELMo, and achieve the same UAS as our representation from 200 training sentences onward. However, our transformation was learned in an unsupervised fashion, without access to the syntactic trees.", "cite_spans": [], "ref_spans": [ { "start": 171, "end": 179, "text": "Figure 4", "ref_id": "FIGREF2" } ], "eq_spans": [], "section": "Minimal Supervision for Structure", "sec_num": "4.3" }, { "text": "Finally, when considering the labeled attachment score, where the model is tasked at predicting not only the child-parent relation but also its label, our syntax representation outperforms elmo-reduced.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Minimal Supervision for Structure", "sec_num": "4.3" }, { "text": "We propose an unsupervised method for the distillation of structural information from neural contextualized word representations. We used a process of sequential BERT-based substitution to create a large number of sentences which are structurally similar, but semantically different. By controlling for structure while changing lexical choice, we learn a metric under which pairs of words that come from structurally-similar sentences are close in space. We demonstrated that the representations acquired by this method share structural properties with their neighbors in space, and show that with a minimal supervision, those representations outperform ELMo in the task of few-shots parsing. The method is a first step towards a better disentanglement between various kinds of information that is represented in neural sequence models.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Conclusion", "sec_num": "5" }, { "text": "The method used to create the structurally equivalent sentences can be useful by its own as a dataaugmentation technique. In future work, we aim to extend this method to allow for a more soft alignment between structurally-equivalent sentences.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Conclusion", "sec_num": "5" }, { "text": "\u2022 Q: as they did , the probability of an impact event temporarily climbed , peaking at 2 . N: however , the probability of flipping a head after having already flipped 20 heads in a row is simply NT: during the first year , the scope of red terror expanded significantly and the number of executions grew into the thousands .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: the celtics honored his memory during the following season by retiring his number 35 . N: the beatles performed the song at the 1969 let it be sessions . NT: the warriors dedicated their round five home match to fai 's memory .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: in the old zurich war , the swiss confederation plundered the monastery , whose monks had fled to zurich . N: the hridaya stra and the \" five meditations \" are recited , after which monks will be served with the gruel and vegetables . NT: other commanders were killed and later rooplo kolhi was arrested near pag wool well , where his troops were fetching water.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: the main cause of the punic wars was the conflict of interests between the existing carthaginian empire and the expanding roman republic . N: the main issue was whether or not something had to be directly perceptible ( meaning intelligible to an ordinary human being ) for it to be a \" copy . NT: the main enemy of the game is a sadistic but intelligent arms-dealer known as the jackal , whose guns are fueling the violence in the country .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: jones maintained lifelong links with his native county , where he had a home , bron menai , dwyran . N: his association with the bbc ended in 1981 with a move back to his native county and itv company yorkshire television , replacing martin tyler as the regional station 's football commentator . NT: he leaves again for his native england , moving to a place near bath , where he works with a powerful local coven .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: silver iodate can be obtained by reacting silver nitrate ( agno3 ) with sodium iodate . N: best mechanical strength is obtained if both sides of the disc are fused to the same type of glass tube and both tubes are under vacuum .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "NT: each of these options can be obtained with a master degree from the university along with the master of engineering degree .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: it confirmed that thomas medwin was a thoroughly learned man , if occasionally imprecise and careless N: it was confirmed that the truth about heather 's murder would be revealed which ultimately led to ben 's departure .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "NT: it proclaimed that the entire movement of plastic art of our time had been thrown into confusion by the discoveries above-mentioned .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: after the death of nadab and abihu , moses dictated what was to be done with their bodies . N: most sources indicate that while no marriage took place between haile melekot and woizero ijigayehu , sahle selassie ordered his grandson legitimized . NT: vvkj pilots who flew the hurricane conversion considered it to be superior to the standard model .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: letters were delivered to sorters who examined the address and placed it in one of a number of \" pigeon holes \" . N: i examined and reported on the thread called transcendental meditation which appears on the page you linked to . NT: ronson visits purported psychopaths , as well as psychologists and psychiatrists who have studied them , and meets with robert d .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: slowboat to hades is a compilation dvd by gorillaz , released in october 2006 . N: the album was released in may 2003 as a single album with a bonus dvd . NT: master series is a compilation album by the british synthpop band visage released in 1997 .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: however , there are also many theories and conspiracies that describe the basis of the plot .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "N: the name tabasco is not definitively known with a number of theories debated among linguists . NT: it is likely that to this day there are some harrisons and harrises that are related .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: nne , married first , to richard , eldest son of sir richard nagle , secretary of state for ireland , temp . N: in the early 1960s , profumo was the secretary of state for war in harold macmillan 's conservative government and was married to actress valerie hobson . NT: he was born in edinburgh , the son of william simpson , minister of the tron church , edinburgh , by his wife jean douglas balderston .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: battle of stoke field , the final engagement of the wars of the roses . N: among others , hogan announced the \" engagement \" of utah-born pitcher roy castleton .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "NT: song of susannah , the sixth installment in the dark tower series .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: it vies for control with its host , causing physiological changes that will eventually cause the host 's internal organs to explode . N: hurtig and loewen developed rival factions within the party , and battled for control . NT: players take control of each of the four main characters at different times throughout the game , which enables multilateral perspective on the storyline .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: as such , radio tirana kept close to the official policy of the people 's republic of china , which was also both anti-west and anti-soviet whilst still being socialist in tone . N: this was in line with the policy outlined by constantine vii porphyrogenitus in de administrando imperio of fomenting strife between the rus ' and the pechenegs . NT: april 2006 , the upr periodically examines the human rights performance of all 193 un member states .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: the engine was designed to accept either regular grade , 87 octane gasoline or premium grade , 91 octane gasoline . N: for example , an advanced html editing field could accept a pasted or inserted image and convert it to a data uri to hide the complexity of external resources from the user . NT: it uses plug-ins ( html parsing technology ) to collect bibliographic information , videos and patents from webpages .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: one such decree was the notorious 1876 ems ukaz , which banned the kulishivka and imposed a russian orthography until 1905 ( called the yaryzhka , after the russian letter yery ) . N: fin 1612 , the shogun declared a decree that specifically banned the killing of cattle . NT: tannis has eliminated the other time lords and set the doctor and the minister against each other .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: a 25 degree list was reduced to 15 degrees ; men had abandoned ship prematurely -hence the pow . N: i suggest the article be reduced to something over half the size . NT: the old high school was converted into a middle school , until in 1971 the 5 .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: the library catalog is maintained on a database that is made accessible to users through the internet. N: this screenshot is made for educational use and used for identification purposes in the article on nba on abc . NT: hpc is the main ingredient in cellugel which is used in book conservation .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: although he lost , he was evaluated highly by kazuyoshi ishii , and he was invited to seidokaikan . N: he attended suny fredonia for one year and in 1976 received a b . NT: played primarily as a small forward , he showed some opportunist play and in his 18 games managed a creditable 12 goals .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "\u2022 Q: for each round won , you gain one point towards winning the match . N: in the fourth round , federer beat tommy robredo and equalled jimmy connors ' record of 27 consecutive grand slam quarterfinals . NT: at the beginning of each mission , as well as the end of the last mission , a cutscene is played that helps develop the story .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A Additional Query-Value Examples", "sec_num": null }, { "text": "In Table 3 , we present the full quantitative results when using BERT as the encoder. \"Baseline\" refers to unmodified vectors derived from BERT, and \"Transformed\" refers to the vectors after the learned syntactic transformation f . \"hard\" refers to evaluation on the subset of POS tags which are most structurally diverse.", "cite_spans": [], "ref_spans": [ { "start": 3, "end": 10, "text": "Table 3", "ref_id": null } ], "eq_spans": [], "section": "B BERT Closest-Word Results", "sec_num": null }, { "text": "Below are the LAS and UAS scores for the experiments described in \u00a74.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "C Complete Parsing Results", "sec_num": null }, { "text": "In Table 6 we present randomly selected examples of groups of structurally-similar sentences ( \u00a73.1).", "cite_spans": [], "ref_spans": [ { "start": 3, "end": 10, "text": "Table 6", "ref_id": "TABREF1" } ], "eq_spans": [], "section": "D Examples of Equivalent Sentences", "sec_num": null }, { "text": "Original the structure is privately owned by the lake-hanford family of aurora , indiana and is not open to the public . 1 the preserve is generally enjoyed by the ecological department of warren , california and is not free to the staff . 2 the park is presently covered by the lake-hanford west of shrewsbury , italy and is not broken to the landscape . 3 the festival is wholly offered by the west club of liberty , arkansas and is not central to the tradition . 4 the pool is mostly administered by the shell town of greenville , maryland and is not navigable to the water . 5 the house is geographically managed by the lake-hanford foundation of ferguson , fl and is not open to the sun .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "#Version Sentence", "sec_num": null }, { "text": "Original on november 18th , 2011 , sllner released the studio album mei zuastand which features re-recorded songs from his entire career . 1 on thursday 9th , 1975 , wolf dedicated the label das en imprint which comprises mixed albums from his golden series . 2 on year 13th , 1985 , hoffmann wrote the vinyl mix von deutschland which plays imagined samples from his bible canon . 3 on circa christmas , 2000 , press signed the lp debut re work which involves created phrases from his bible quote . 4 on january 15th , 1995 , sllner wrote the camera y se theory which mixes cast phrases from his experimental archive . 5 on oct 13th , 1983 , hansen organised the compilation concert ha radio which gives launched clips from his small film .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "#Version Sentence", "sec_num": null }, { "text": "Original uhm ; we 're not proposing to give rollbackers the reviewer right . 1 ah ; we 're not calling to quote comics the way hello . 2 hi ; we 're not preparing to hear hits the dirt lady . 3", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "#Version Sentence", "sec_num": null }, { "text": "shi ; we 're not asking to put rollbackers the board die . 4 ar ; we 're not expecting to face rollbackers the place fell . 5 whoa ; we 're not getting to detroit wants the boat paid .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "#Version Sentence", "sec_num": null }, { "text": "Original coniston water is an example of a ribbon lake formed by glaciation . 1 floating town is an artwork of a concrete area contaminated by mud . 2 vista florida is an isle of a seaside lagoon fed by watershed . 3 pit process is an occurrence of a hollow underground caused by settlement . 4 union pass is an explanation of a highland section developed by anderson . 5 ball phase is an exploration of a basalt basalt influenced by creep .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "#Version Sentence", "sec_num": null }, { "text": "Original the highest lookout point , at above sea level , is trimble mountain , off brewer road . 1 the greatest steep elevation , at above east cliff , is green rock , off little neck . 2 the greatest lake club , at above east summit , is swiss cut , off northern pike . 3 the biggest missing asset , at above single count , is local motel , off washington plaza . 4 the smallest public surfing , at above virgin point , is grant lagoon , off white strait . 5 the southwest east boundary , at above water flow , is trim hollow , off east town .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "#Version Sentence", "sec_num": null }, { "text": "Original ample sdk is a lightweight javascript library intended to simplify cross-browser web application development . 1 rapid editor is a popular editorial script suited to manage multi domain book edition . 2 free id is a mandatory public implementation written to manage repository generic server environment . 3 solar platform is a native developed stack written to ease regional complex sensing analysis . 4 standard library is a complete python interface required to provide cellular mesh construction engine . 5 flex module is a standardized foundry block applied to facilitate component development common work .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "#Version Sentence", "sec_num": null }, { "text": "Original she wore a pale pink gown , silver crown and had pale pink wings . 1 she boasted a large halt purple , fuzzy lip and had twin firm wrists . 2 she spun a thin olive jelly , joined yarn and had large silver bubbles . 3 she flared a high frequency yellow , reddish rose and had fried like moses . 4", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "#Version Sentence", "sec_num": null }, { "text": "she exhibited a small frame overall , broad head and had oval eyed curves . 5", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "#Version Sentence", "sec_num": null }, { "text": "she wrapped a silky ga yellow , moth hide and had homemade gold roses .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "#Version Sentence", "sec_num": null }, { "text": "Original tegan is somewhat quiet and is rather scared , but kamryn reasures her everything will be ok . 1 man is slightly pissed and is rather awkward , but kamryn protests her night will be ok . 2 lao is real sad and is rather disappointed , but san figures her story will be ok . 3 daughter is increasingly pregnant and is rather uncomfortable , but ni confirms her birth will be ok . 4 mai is strangely warm and is rather short , but papa wishes her day will be ok . 5 mare is slowly back and is rather upset , but pa asserts her sister will be ok .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "#Version Sentence", "sec_num": null }, { "text": "Original shapley participated in the \" great debate \" with heber d . 1 morris put in the \" heroic speech \" with heber energy . 2 hall met in the \" ninth season \" with walton moore . 3 patel helped in the \" double coup \" with ibn salem . 4 chu sent in the \" universal text \" with u z . 5 smith exhibited in the \" red year \" with william james .", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "#Version Sentence", "sec_num": null }, { "text": "Original the added english voice-over narration by the vampire ancestor removes any ambiguity . 1 the untitled thai adventure script by the light corps includes any future . 2 the improved industrial hole tool by the freeman workshop touches any resistance . 3 the arched robotic interference use by the computer computer checks any message . 4 the fixed regular speech described by the german army encompasses any type . 5 the combined complete phone acquisition by the surround computer marks any microphone . ", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "#Version Sentence", "sec_num": null }, { "text": "We focus on lexical semantics.3 There is a syntactic distinction between the two, with \"maple\" being part of a noun compound and \"neural\" being an adjective. However, we focus in their similarity as noun modifiers in both phrases.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "These differences in syntactic position are also of relevance to language modeling, as different positions may pose different restrictions on the words that can appear in them.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "We note that this process bears some similarity to Gibbs sampling from BERT conditioned LM.6 We maintain the same POS so that the dataset will be valid for other tasks that require structure-preserving variants. However, In practice, we did not observe major differences when repeating the experiments reported here without the POS-preserving constraint when generating the data.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "Since BERT uses word-piece tokenization, we take the first token to represent each word.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "A large enough mini-batch is necessary to find challenging negative examples.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "Dep. edge Head's dep. edge Tree path Tree path Tree path", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null } ], "back_matter": [ { "text": "We would like to thank Gal Chechik for providing valuable feedback on early version of this work. This project has received funding from the Europoean Research Council (ERC) under the Europoean Union's Horizon 2020 research and innovation programme, grant agreement No. 802774 (iEXTRACT). Yanai Elazar is grateful to be partially supported by the PBC fellowship for outstanding PhD candidates in Data Science.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Acknowledgments", "sec_num": null }, { "text": " Table 3 : Full quantitative results when using BERT as the encoder. \"Baseline\" refers to unmodified vectors derived from BERT, and \"Transformed\" refers to the vectors after the learned syntactic transformation f . \"hard\" refers to evaluation on the subset of POS tags which are most structurally diverse. ", "cite_spans": [], "ref_spans": [ { "start": 1, "end": 8, "text": "Table 3", "ref_id": null } ], "eq_spans": [], "section": "annex", "sec_num": null } ], "bib_entries": { "BIBREF0": { "ref_id": "b0", "title": "Fine-grained analysis of sentence embeddings using auxiliary prediction tasks", "authors": [ { "first": "Yossi", "middle": [], "last": "Adi", "suffix": "" }, { "first": "Einat", "middle": [], "last": "Kermany", "suffix": "" }, { "first": "Yonatan", "middle": [], "last": "Belinkov", "suffix": "" }, { "first": "Ofer", "middle": [], "last": "Lavi", "suffix": "" }, { "first": "Yoav", "middle": [], "last": "Goldberg", "suffix": "" } ], "year": 2016, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Yossi Adi, Einat Kermany, Yonatan Belinkov, Ofer Lavi, and Yoav Goldberg. 2016. Fine-grained anal- ysis of sentence embeddings using auxiliary predic- tion tasks. CoRR, abs/1608.04207.", "links": null }, "BIBREF1": { "ref_id": "b1", "title": "Deep variational information bottleneck", "authors": [ { "first": "Alexander", "middle": [], "last": "Alemi", "suffix": "" }, { "first": "Ian", "middle": [], "last": "Fischer", "suffix": "" }, { "first": "Joshua", "middle": [ "V" ], "last": "Dillon", "suffix": "" }, { "first": "Murphy", "middle": [], "last": "Murphy", "suffix": "" } ], "year": 2016, "venue": "Proceedings of the International Conference on Learning Representations (ICLR", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Alexander Alemi, Ian Fischer, Joshua V. Dillon, and Murphy Murphy. 2016. Deep variational informa- tion bottleneck. In Proceedings of the International Conference on Learning Representations (ICLR).", "links": null }, "BIBREF2": { "ref_id": "b2", "title": "Representation learning: A review and new perspectives", "authors": [ { "first": "Yoshua", "middle": [], "last": "Bengio", "suffix": "" }, { "first": "Aaron", "middle": [ "C" ], "last": "Courville", "suffix": "" }, { "first": "Pascal", "middle": [], "last": "Vincent", "suffix": "" } ], "year": 2013, "venue": "IEEE Trans. Pattern Anal. Mach. Intell", "volume": "35", "issue": "8", "pages": "1798--1828", "other_ids": { "DOI": [ "10.1109/TPAMI.2013.50" ] }, "num": null, "urls": [], "raw_text": "Yoshua Bengio, Aaron C. Courville, and Pascal Vin- cent. 2013. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell., 35(8):1798-1828.", "links": null }, "BIBREF3": { "ref_id": "b3", "title": "What you can cram into a single \\$&!#* vector: Probing sentence embeddings for linguistic properties", "authors": [ { "first": "Alexis", "middle": [], "last": "Conneau", "suffix": "" }, { "first": "Germ\u00e1n", "middle": [], "last": "Kruszewski", "suffix": "" }, { "first": "Guillaume", "middle": [], "last": "Lample", "suffix": "" }, { "first": "Lo\u00efc", "middle": [], "last": "Barrault", "suffix": "" }, { "first": "Marco", "middle": [], "last": "Baroni", "suffix": "" } ], "year": 2018, "venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, ACL 2018", "volume": "1", "issue": "", "pages": "2126--2136", "other_ids": { "DOI": [ "10.18653/v1/P18-1198" ] }, "num": null, "urls": [], "raw_text": "Alexis Conneau, Germ\u00e1n Kruszewski, Guillaume Lam- ple, Lo\u00efc Barrault, and Marco Baroni. 2018. What you can cram into a single \\$&!#* vector: Probing sentence embeddings for linguistic properties. In Proceedings of the 56th Annual Meeting of the As- sociation for Computational Linguistics, ACL 2018, Melbourne, Australia, July 15-20, 2018, Volume 1: Long Papers, pages 2126-2136.", "links": null }, "BIBREF4": { "ref_id": "b4", "title": "BERT: pre-training of deep bidirectional transformers for language understanding", "authors": [ { "first": "Jacob", "middle": [], "last": "Devlin", "suffix": "" }, { "first": "Ming-Wei", "middle": [], "last": "Chang", "suffix": "" }, { "first": "Kenton", "middle": [], "last": "Lee", "suffix": "" }, { "first": "Kristina", "middle": [], "last": "Toutanova", "suffix": "" } ], "year": 2019, "venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019", "volume": "1", "issue": "", "pages": "4171--4186", "other_ids": { "DOI": [ "10.18653/v1/n19-1423" ] }, "num": null, "urls": [], "raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: pre-training of deep bidirectional transformers for language under- standing. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2-7, 2019, Volume 1 (Long and Short Pa- pers), pages 4171-4186. Association for Computa- tional Linguistics.", "links": null }, "BIBREF5": { "ref_id": "b5", "title": "Deep biaffine attention for neural dependency parsing", "authors": [ { "first": "Timothy", "middle": [], "last": "Dozat", "suffix": "" }, { "first": "D", "middle": [], "last": "Christopher", "suffix": "" }, { "first": "", "middle": [], "last": "Manning", "suffix": "" } ], "year": 2016, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": { "arXiv": [ "arXiv:1611.01734" ] }, "num": null, "urls": [], "raw_text": "Timothy Dozat and Christopher D Manning. 2016. Deep biaffine attention for neural dependency pars- ing. arXiv preprint arXiv:1611.01734.", "links": null }, "BIBREF6": { "ref_id": "b6", "title": "When bert forgets how to pos: Amnesic probing of linguistic properties and mlm predictions", "authors": [ { "first": "Yanai", "middle": [], "last": "Elazar", "suffix": "" }, { "first": "Shauli", "middle": [], "last": "Ravfogel", "suffix": "" }, { "first": "Alon", "middle": [], "last": "Jacovi", "suffix": "" }, { "first": "Yoav", "middle": [], "last": "Goldberg", "suffix": "" } ], "year": 2020, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Yanai Elazar, Shauli Ravfogel, Alon Jacovi, and Yoav Goldberg. 2020. When bert forgets how to pos: Am- nesic probing of linguistic properties and mlm pre- dictions.", "links": null }, "BIBREF7": { "ref_id": "b7", "title": "Controlling linguistic style aspects in neural language generation", "authors": [ { "first": "Jessica", "middle": [], "last": "Ficler", "suffix": "" }, { "first": "Yoav", "middle": [], "last": "Goldberg", "suffix": "" } ], "year": 2017, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": { "arXiv": [ "arXiv:1707.02633" ] }, "num": null, "urls": [], "raw_text": "Jessica Ficler and Yoav Goldberg. 2017. Controlling linguistic style aspects in neural language genera- tion. arXiv preprint arXiv:1707.02633.", "links": null }, "BIBREF8": { "ref_id": "b8", "title": "Style transfer in text: Exploration and evaluation", "authors": [ { "first": "Zhenxin", "middle": [], "last": "Fu", "suffix": "" }, { "first": "Xiaoye", "middle": [], "last": "Tan", "suffix": "" }, { "first": "Nanyun", "middle": [], "last": "Peng", "suffix": "" }, { "first": "Dongyan", "middle": [], "last": "Zhao", "suffix": "" }, { "first": "Rui", "middle": [], "last": "Yan", "suffix": "" } ], "year": 2018, "venue": "Thirty-Second AAAI Conference on Artificial Intelligence", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Zhenxin Fu, Xiaoye Tan, Nanyun Peng, Dongyan Zhao, and Rui Yan. 2018. Style transfer in text: Explo- ration and evaluation. In Thirty-Second AAAI Con- ference on Artificial Intelligence.", "links": null }, "BIBREF9": { "ref_id": "b9", "title": "Texture synthesis and style transfer using perceptual image representations from convolutional neural networks", "authors": [ { "first": "A", "middle": [], "last": "Leon", "suffix": "" }, { "first": "", "middle": [], "last": "Gatys", "suffix": "" } ], "year": 2017, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Leon A. Gatys. 2017. Texture synthesis and style trans- fer using perceptual image representations from con- volutional neural networks. Ph.D. thesis, University of T\u00fcbingen, Germany.", "links": null }, "BIBREF10": { "ref_id": "b10", "title": "Assessing BERT's syntactic abilities", "authors": [ { "first": "Yoav", "middle": [], "last": "Goldberg", "suffix": "" } ], "year": 2019, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Yoav Goldberg. 2019. Assessing BERT's syntactic abilities. CoRR, abs/1901.05287.", "links": null }, "BIBREF11": { "ref_id": "b11", "title": "Colorless green recurrent networks dream hierarchically", "authors": [ { "first": "Kristina", "middle": [], "last": "Gulordava", "suffix": "" }, { "first": "Piotr", "middle": [], "last": "Bojanowski", "suffix": "" }, { "first": "Edouard", "middle": [], "last": "Grave", "suffix": "" }, { "first": "Tal", "middle": [], "last": "Linzen", "suffix": "" }, { "first": "Marco", "middle": [], "last": "Baroni", "suffix": "" } ], "year": 2018, "venue": "Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT", "volume": "", "issue": "", "pages": "1195--1205", "other_ids": {}, "num": null, "urls": [], "raw_text": "Kristina Gulordava, Piotr Bojanowski, Edouard Grave, Tal Linzen, and Marco Baroni. 2018. Color- less green recurrent networks dream hierarchically. In Proceedings of the Conference of the North American Chapter of the Association for Computa- tional Linguistics: Human Language Technologies, NAACL-HLT, pages 1195-1205.", "links": null }, "BIBREF12": { "ref_id": "b12", "title": "A two-step disentanglement method", "authors": [ { "first": "Naama", "middle": [], "last": "Hadad", "suffix": "" }, { "first": "Lior", "middle": [], "last": "Wolf", "suffix": "" }, { "first": "Moni", "middle": [], "last": "Shahar", "suffix": "" } ], "year": 2018, "venue": "IEEE Conference on Computer Vision and Pattern Recognition, (CVPR)", "volume": "", "issue": "", "pages": "772--780", "other_ids": { "DOI": [ "10.1109/CVPR.2018.00087" ] }, "num": null, "urls": [], "raw_text": "Naama Hadad, Lior Wolf, and Moni Shahar. 2018. A two-step disentanglement method. In IEEE Confer- ence on Computer Vision and Pattern Recognition, (CVPR), pages 772-780.", "links": null }, "BIBREF13": { "ref_id": "b13", "title": "Designing and interpreting probes with control tasks", "authors": [ { "first": "John", "middle": [], "last": "Hewitt", "suffix": "" }, { "first": "Percy", "middle": [], "last": "Liang", "suffix": "" } ], "year": 2019, "venue": "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019", "volume": "", "issue": "", "pages": "2733--2743", "other_ids": { "DOI": [ "10.18653/v1/D19-1275" ] }, "num": null, "urls": [], "raw_text": "John Hewitt and Percy Liang. 2019. Designing and interpreting probes with control tasks. In Proceed- ings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th Inter- national Joint Conference on Natural Language Pro- cessing, EMNLP-IJCNLP 2019, Hong Kong, China, November 3-7, 2019, pages 2733-2743. Association for Computational Linguistics.", "links": null }, "BIBREF14": { "ref_id": "b14", "title": "A structural probe for finding syntax in word representations", "authors": [ { "first": "John", "middle": [], "last": "Hewitt", "suffix": "" }, { "first": "D", "middle": [], "last": "Christopher", "suffix": "" }, { "first": "", "middle": [], "last": "Manning", "suffix": "" } ], "year": 2019, "venue": "Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT", "volume": "", "issue": "", "pages": "4129--4138", "other_ids": {}, "num": null, "urls": [], "raw_text": "John Hewitt and Christopher D. Manning. 2019. A structural probe for finding syntax in word repre- sentations. In Proceedings of the Conference of the North American Chapter of the Association for Com- putational Linguistics: Human Language Technolo- gies, NAACL-HLT, pages 4129-4138.", "links": null }, "BIBREF15": { "ref_id": "b15", "title": "Deep metric learning using triplet network", "authors": [ { "first": "Elad", "middle": [], "last": "Hoffer", "suffix": "" }, { "first": "Nir", "middle": [], "last": "Ailon", "suffix": "" } ], "year": 2015, "venue": "Similarity-Based Pattern Recognition -Third International Workshop, SIM-BAD", "volume": "", "issue": "", "pages": "84--92", "other_ids": { "DOI": [ "10.1007/978-3-319-24261-3_7" ] }, "num": null, "urls": [], "raw_text": "Elad Hoffer and Nir Ailon. 2015. Deep metric learning using triplet network. In Similarity-Based Pattern Recognition -Third International Workshop, SIM- BAD, pages 84-92.", "links": null }, "BIBREF16": { "ref_id": "b16", "title": "An improved non-monotonic transition system for dependency parsing", "authors": [ { "first": "Matthew", "middle": [], "last": "Honnibal", "suffix": "" }, { "first": "Mark", "middle": [], "last": "Johnson", "suffix": "" } ], "year": 2015, "venue": "Proceedings of the 2015 conference on empirical methods in natural language processing", "volume": "", "issue": "", "pages": "1373--1378", "other_ids": {}, "num": null, "urls": [], "raw_text": "Matthew Honnibal and Mark Johnson. 2015. An im- proved non-monotonic transition system for depen- dency parsing. In Proceedings of the 2015 confer- ence on empirical methods in natural language pro- cessing, pages 1373-1378.", "links": null }, "BIBREF17": { "ref_id": "b17", "title": "spacy 2: Natural language understanding with bloom embeddings, convolutional neural networks and incremental parsing", "authors": [ { "first": "Matthew", "middle": [], "last": "Honnibal", "suffix": "" }, { "first": "Ines", "middle": [], "last": "Montani", "suffix": "" } ], "year": 2017, "venue": "", "volume": "7", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Matthew Honnibal and Ines Montani. 2017. spacy 2: Natural language understanding with bloom embed- dings, convolutional neural networks and incremen- tal parsing. To appear, 7(1).", "links": null }, "BIBREF18": { "ref_id": "b18", "title": "Toward controlled generation of text", "authors": [ { "first": "Zhiting", "middle": [], "last": "Hu", "suffix": "" }, { "first": "Zichao", "middle": [], "last": "Yang", "suffix": "" }, { "first": "Xiaodan", "middle": [], "last": "Liang", "suffix": "" }, { "first": "Ruslan", "middle": [], "last": "Salakhutdinov", "suffix": "" }, { "first": "Eric", "middle": [ "P" ], "last": "Xing", "suffix": "" } ], "year": 2017, "venue": "Proceedings of the 34th International Conference on Machine Learning", "volume": "", "issue": "", "pages": "1587--1596", "other_ids": {}, "num": null, "urls": [], "raw_text": "Zhiting Hu, Zichao Yang, Xiaodan Liang, Ruslan Salakhutdinov, and Eric P. Xing. 2017. Toward con- trolled generation of text. In Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017, pages 1587-1596.", "links": null }, "BIBREF19": { "ref_id": "b19", "title": "Visualisation and 'diagnostic classifiers' reveal how recurrent and recursive neural networks process hierarchical structure", "authors": [ { "first": "Dieuwke", "middle": [], "last": "Hupkes", "suffix": "" }, { "first": "Sara", "middle": [], "last": "Veldhoen", "suffix": "" }, { "first": "Willem", "middle": [ "H" ], "last": "Zuidema", "suffix": "" } ], "year": 2018, "venue": "J. Artif. Intell. Res", "volume": "61", "issue": "", "pages": "907--926", "other_ids": { "DOI": [ "10.1613/jair.1.11196" ] }, "num": null, "urls": [], "raw_text": "Dieuwke Hupkes, Sara Veldhoen, and Willem H. Zuidema. 2018. Visualisation and 'diagnostic classi- fiers' reveal how recurrent and recursive neural net- works process hierarchical structure. J. Artif. Intell. Res., 61:907-926.", "links": null }, "BIBREF20": { "ref_id": "b20", "title": "BERT for coreference resolution: Baselines and analysis", "authors": [ { "first": "Mandar", "middle": [], "last": "Joshi", "suffix": "" }, { "first": "Omer", "middle": [], "last": "Levy", "suffix": "" }, { "first": "S", "middle": [], "last": "Daniel", "suffix": "" }, { "first": "Luke", "middle": [], "last": "Weld", "suffix": "" }, { "first": "", "middle": [], "last": "Zettlemoyer", "suffix": "" } ], "year": 2019, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": { "arXiv": [ "arXiv:1908.09091" ] }, "num": null, "urls": [], "raw_text": "Mandar Joshi, Omer Levy, Daniel S Weld, and Luke Zettlemoyer. 2019. BERT for coreference reso- lution: Baselines and analysis. arXiv preprint arXiv:1908.09091.", "links": null }, "BIBREF21": { "ref_id": "b21", "title": "Adam: A method for stochastic optimization", "authors": [ { "first": "P", "middle": [], "last": "Diederik", "suffix": "" }, { "first": "Jimmy", "middle": [], "last": "Kingma", "suffix": "" }, { "first": "", "middle": [], "last": "Ba", "suffix": "" } ], "year": 2015, "venue": "International Conference on Learning Representations, ICLR", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Diederik P. Kingma and Jimmy Ba. 2015. Adam: A method for stochastic optimization. In International Conference on Learning Representations, ICLR.", "links": null }, "BIBREF22": { "ref_id": "b22", "title": "Constituency parsing with a self-attentive encoder", "authors": [ { "first": "Nikita", "middle": [], "last": "Kitaev", "suffix": "" }, { "first": "Dan", "middle": [], "last": "Klein", "suffix": "" } ], "year": 2018, "venue": "Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL)", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Nikita Kitaev and Dan Klein. 2018. Constituency pars- ing with a self-attentive encoder. In Proceedings of the Annual Meeting of the Association for Computa- tional Linguistics (ACL).", "links": null }, "BIBREF23": { "ref_id": "b23", "title": "Multiple-attribute text rewriting", "authors": [ { "first": "Guillaume", "middle": [], "last": "Lample", "suffix": "" }, { "first": "Sandeep", "middle": [], "last": "Subramanian", "suffix": "" }, { "first": "Eric", "middle": [], "last": "Smith", "suffix": "" }, { "first": "Ludovic", "middle": [], "last": "Denoyer", "suffix": "" }, { "first": "Marc'aurelio", "middle": [], "last": "Ranzato", "suffix": "" }, { "first": "Y-Lan", "middle": [], "last": "Boureau", "suffix": "" } ], "year": 2018, "venue": "International Conference on Learning Representations", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Guillaume Lample, Sandeep Subramanian, Eric Smith, Ludovic Denoyer, Marc'Aurelio Ranzato, and Y- Lan Boureau. 2018. Multiple-attribute text rewrit- ing. In International Conference on Learning Rep- resentations.", "links": null }, "BIBREF24": { "ref_id": "b24", "title": "Linguistic regularities in sparse and explicit word representations", "authors": [ { "first": "Omer", "middle": [], "last": "Levy", "suffix": "" }, { "first": "Yoav", "middle": [], "last": "Goldberg", "suffix": "" } ], "year": 2014, "venue": "Proceedings of the eighteenth conference on computational natural language learning", "volume": "", "issue": "", "pages": "171--180", "other_ids": {}, "num": null, "urls": [], "raw_text": "Omer Levy and Yoav Goldberg. 2014. Linguistic reg- ularities in sparse and explicit word representations. In Proceedings of the eighteenth conference on com- putational natural language learning, pages 171- 180.", "links": null }, "BIBREF25": { "ref_id": "b25", "title": "Specializing word embeddings (for parsing) by information bottleneck", "authors": [ { "first": "Lisa", "middle": [], "last": "Xiang", "suffix": "" }, { "first": "Jason", "middle": [], "last": "Li", "suffix": "" }, { "first": "", "middle": [], "last": "Eisner", "suffix": "" } ], "year": 2019, "venue": "Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP)", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Xiang Lisa Li and Jason Eisner. 2019. Specializing word embeddings (for parsing) by information bot- tleneck. In Proceedings of the Conference on Em- pirical Methods in Natural Language Processing (EMNLP).", "links": null }, "BIBREF26": { "ref_id": "b26", "title": "Open sesame: Getting inside berts linguistic knowledge", "authors": [ { "first": "Yongjie", "middle": [], "last": "Lin", "suffix": "" }, { "first": "Yi", "middle": [], "last": "Chern Tan", "suffix": "" }, { "first": "Robert", "middle": [], "last": "Frank", "suffix": "" } ], "year": 2019, "venue": "Proceedings of the 2019 ACL Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP", "volume": "", "issue": "", "pages": "241--253", "other_ids": {}, "num": null, "urls": [], "raw_text": "Yongjie Lin, Yi Chern Tan, and Robert Frank. 2019. Open sesame: Getting inside berts linguistic knowl- edge. In Proceedings of the 2019 ACL Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, pages 241-253.", "links": null }, "BIBREF27": { "ref_id": "b27", "title": "Linguistic knowledge and transferability of contextual representations", "authors": [ { "first": "F", "middle": [], "last": "Nelson", "suffix": "" }, { "first": "Matt", "middle": [], "last": "Liu", "suffix": "" }, { "first": "Yonatan", "middle": [], "last": "Gardner", "suffix": "" }, { "first": "Matthew", "middle": [], "last": "Belinkov", "suffix": "" }, { "first": "Noah A", "middle": [], "last": "Peters", "suffix": "" }, { "first": "", "middle": [], "last": "Smith", "suffix": "" } ], "year": 2019, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": { "arXiv": [ "arXiv:1903.08855" ] }, "num": null, "urls": [], "raw_text": "Nelson F Liu, Matt Gardner, Yonatan Belinkov, Matthew Peters, and Noah A Smith. 2019a. Lin- guistic knowledge and transferability of contextual representations. arXiv preprint arXiv:1903.08855.", "links": null }, "BIBREF29": { "ref_id": "b29", "title": "Visualizing data using t-SNE", "authors": [ { "first": "Laurens", "middle": [], "last": "Van Der Maaten", "suffix": "" }, { "first": "Geoffrey", "middle": [], "last": "Hinton", "suffix": "" } ], "year": 2008, "venue": "Journal of Machine Learning Research", "volume": "9", "issue": "", "pages": "2579--2605", "other_ids": {}, "num": null, "urls": [], "raw_text": "Laurens van der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-SNE. Journal of Machine Learning Research, 9:2579-2605.", "links": null }, "BIBREF30": { "ref_id": "b30", "title": "Disentangling factors of variation in deep representation using adversarial training", "authors": [ { "first": "Micha\u00ebl", "middle": [], "last": "Mathieu", "suffix": "" }, { "first": "Junbo", "middle": [ "Jake" ], "last": "Zhao", "suffix": "" }, { "first": "Pablo", "middle": [], "last": "Sprechmann", "suffix": "" }, { "first": "Aditya", "middle": [], "last": "Ramesh", "suffix": "" }, { "first": "Yann", "middle": [], "last": "Lecun", "suffix": "" } ], "year": 2016, "venue": "Advances in Neural Information Processing Systems", "volume": "", "issue": "", "pages": "5041--5049", "other_ids": {}, "num": null, "urls": [], "raw_text": "Micha\u00ebl Mathieu, Junbo Jake Zhao, Pablo Sprechmann, Aditya Ramesh, and Yann LeCun. 2016. Disentan- gling factors of variation in deep representation us- ing adversarial training. In Advances in Neural In- formation Processing Systems, pages 5041-5049.", "links": null }, "BIBREF32": { "ref_id": "b32", "title": "Distributed representations of words and phrases and their compositionality", "authors": [ { "first": "Tomas", "middle": [], "last": "Mikolov", "suffix": "" }, { "first": "Ilya", "middle": [], "last": "Sutskever", "suffix": "" }, { "first": "Kai", "middle": [], "last": "Chen", "suffix": "" }, { "first": "Greg", "middle": [ "S" ], "last": "Corrado", "suffix": "" }, { "first": "Jeff", "middle": [], "last": "Dean", "suffix": "" } ], "year": 2013, "venue": "Advances in neural information processing systems", "volume": "", "issue": "", "pages": "3111--3119", "other_ids": {}, "num": null, "urls": [], "raw_text": "Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Cor- rado, and Jeff Dean. 2013a. Distributed representa- tions of words and phrases and their compositional- ity. In Advances in neural information processing systems, pages 3111-3119.", "links": null }, "BIBREF33": { "ref_id": "b33", "title": "Linguistic regularities in continuous space word representations", "authors": [ { "first": "Tomas", "middle": [], "last": "Mikolov", "suffix": "" }, { "first": "Yih", "middle": [], "last": "Wen-Tau", "suffix": "" }, { "first": "Geoffrey", "middle": [], "last": "Zweig", "suffix": "" } ], "year": 2013, "venue": "Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", "volume": "", "issue": "", "pages": "746--751", "other_ids": {}, "num": null, "urls": [], "raw_text": "Tomas Mikolov, Wen-tau Yih, and Geoffrey Zweig. 2013b. Linguistic regularities in continuous space word representations. In Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 746-751.", "links": null }, "BIBREF34": { "ref_id": "b34", "title": "Learning disentangled representations with semi-supervised deep generative models", "authors": [ { "first": "Siddharth", "middle": [], "last": "Narayanaswamy", "suffix": "" }, { "first": "Brooks", "middle": [], "last": "Paige", "suffix": "" }, { "first": "Jan-Willem", "middle": [], "last": "Van De Meent", "suffix": "" }, { "first": "Alban", "middle": [], "last": "Desmaison", "suffix": "" }, { "first": "Noah", "middle": [ "D" ], "last": "Goodman", "suffix": "" }, { "first": "Pushmeet", "middle": [], "last": "Kohli", "suffix": "" }, { "first": "Frank", "middle": [ "D" ], "last": "Wood", "suffix": "" }, { "first": "Philip", "middle": [ "H S" ], "last": "Torr", "suffix": "" } ], "year": 2017, "venue": "Advances in Neural Information Processing Systems", "volume": "", "issue": "", "pages": "5925--5935", "other_ids": {}, "num": null, "urls": [], "raw_text": "Siddharth Narayanaswamy, Brooks Paige, Jan-Willem van de Meent, Alban Desmaison, Noah D. Good- man, Pushmeet Kohli, Frank D. Wood, and Philip H. S. Torr. 2017. Learning disentangled representa- tions with semi-supervised deep generative models. In Advances in Neural Information Processing Sys- tems, pages 5925-5935.", "links": null }, "BIBREF35": { "ref_id": "b35", "title": "Reconstructionbased disentanglement for pose-invariant face recognition", "authors": [ { "first": "Xi", "middle": [], "last": "Peng", "suffix": "" }, { "first": "Xiang", "middle": [], "last": "Yu", "suffix": "" }, { "first": "Kihyuk", "middle": [], "last": "Sohn", "suffix": "" }, { "first": "Dimitris", "middle": [ "N" ], "last": "Metaxas", "suffix": "" }, { "first": "Manmohan", "middle": [], "last": "Chandraker", "suffix": "" } ], "year": 2017, "venue": "IEEE International Conference on Computer Visionn (ICCV)", "volume": "", "issue": "", "pages": "1632--1641", "other_ids": { "DOI": [ "10.1109/ICCV.2017.180" ] }, "num": null, "urls": [], "raw_text": "Xi Peng, Xiang Yu, Kihyuk Sohn, Dimitris N. Metaxas, and Manmohan Chandraker. 2017. Reconstruction- based disentanglement for pose-invariant face recog- nition. In IEEE International Conference on Com- puter Visionn (ICCV), pages 1632-1641.", "links": null }, "BIBREF36": { "ref_id": "b36", "title": "Glove: Global vectors for word representation", "authors": [ { "first": "Jeffrey", "middle": [], "last": "Pennington", "suffix": "" }, { "first": "Richard", "middle": [], "last": "Socher", "suffix": "" }, { "first": "Christopher", "middle": [], "last": "Manning", "suffix": "" } ], "year": 2014, "venue": "Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)", "volume": "", "issue": "", "pages": "1532--1543", "other_ids": {}, "num": null, "urls": [], "raw_text": "Jeffrey Pennington, Richard Socher, and Christopher Manning. 2014. Glove: Global vectors for word rep- resentation. In Proceedings of the 2014 conference on empirical methods in natural language process- ing (EMNLP), pages 1532-1543.", "links": null }, "BIBREF37": { "ref_id": "b37", "title": "Deep contextualized word representations", "authors": [ { "first": "Matthew", "middle": [ "E" ], "last": "Peters", "suffix": "" }, { "first": "Mark", "middle": [], "last": "Neumann", "suffix": "" }, { "first": "Mohit", "middle": [], "last": "Iyyer", "suffix": "" }, { "first": "Matt", "middle": [], "last": "Gardner", "suffix": "" }, { "first": "Christopher", "middle": [], "last": "Clark", "suffix": "" }, { "first": "Kenton", "middle": [], "last": "Lee", "suffix": "" }, { "first": "Luke", "middle": [], "last": "Zettlemoyer", "suffix": "" } ], "year": 2018, "venue": "Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", "volume": "1", "issue": "", "pages": "2227--2237", "other_ids": { "DOI": [ "10.18653/v1/n18-1202" ] }, "num": null, "urls": [], "raw_text": "Matthew E. Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, and Luke Zettlemoyer. 2018. Deep contextualized word rep- resentations. In Proceedings of the 2018 Confer- ence of the North American Chapter of the Associ- ation for Computational Linguistics: Human Lan- guage Technologies, NAACL-HLT 2018, New Or- leans, Louisiana, USA, June 1-6, 2018, Volume 1 (Long Papers), pages 2227-2237. Association for Computational Linguistics.", "links": null }, "BIBREF38": { "ref_id": "b38", "title": "Probing the probing paradigm: Does probing accuracy entail task relevance? CoRR", "authors": [ { "first": "Abhilasha", "middle": [], "last": "Ravichander", "suffix": "" }, { "first": "Yonatan", "middle": [], "last": "Belinkov", "suffix": "" }, { "first": "Eduard", "middle": [ "H" ], "last": "Hovy", "suffix": "" } ], "year": 2020, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Abhilasha Ravichander, Yonatan Belinkov, and Ed- uard H. Hovy. 2020. Probing the probing paradigm: Does probing accuracy entail task relevance? CoRR, abs/2005.00719.", "links": null }, "BIBREF39": { "ref_id": "b39", "title": "Visualizing and measuring the geometry of bert", "authors": [ { "first": "Emily", "middle": [], "last": "Reif", "suffix": "" }, { "first": "Ann", "middle": [], "last": "Yuan", "suffix": "" }, { "first": "Martin", "middle": [], "last": "Wattenberg", "suffix": "" }, { "first": "Fernanda", "middle": [ "B" ], "last": "Viegas", "suffix": "" }, { "first": "Andy", "middle": [], "last": "Coenen", "suffix": "" }, { "first": "Adam", "middle": [], "last": "Pearce", "suffix": "" }, { "first": "Been", "middle": [], "last": "Kim", "suffix": "" } ], "year": 2019, "venue": "Advances in Neural Information Processing Systems", "volume": "", "issue": "", "pages": "8592--8600", "other_ids": {}, "num": null, "urls": [], "raw_text": "Emily Reif, Ann Yuan, Martin Wattenberg, Fernanda B Viegas, Andy Coenen, Adam Pearce, and Been Kim. 2019. Visualizing and measuring the geometry of bert. In Advances in Neural Information Processing Systems, pages 8592-8600.", "links": null }, "BIBREF40": { "ref_id": "b40", "title": "Probing natural language inference models through semantic fragments", "authors": [ { "first": "Kyle", "middle": [], "last": "Richardson", "suffix": "" }, { "first": "Hai", "middle": [], "last": "Hu", "suffix": "" }, { "first": "S", "middle": [], "last": "Lawrence", "suffix": "" }, { "first": "Ashish", "middle": [], "last": "Moss", "suffix": "" }, { "first": "", "middle": [], "last": "Sabharwal", "suffix": "" } ], "year": 2019, "venue": "", "volume": "", "issue": "", "pages": "", "other_ids": { "arXiv": [ "arXiv:1909.07521" ] }, "num": null, "urls": [], "raw_text": "Kyle Richardson, Hai Hu, Lawrence S Moss, and Ashish Sabharwal. 2019. Probing natural lan- guage inference models through semantic fragments. arXiv preprint arXiv:1909.07521.", "links": null }, "BIBREF41": { "ref_id": "b41", "title": "FaceNet: A unified embedding for face recognition and clustering", "authors": [ { "first": "Florian", "middle": [], "last": "Schroff", "suffix": "" }, { "first": "Dmitry", "middle": [], "last": "Kalenichenko", "suffix": "" }, { "first": "James", "middle": [], "last": "Philbin", "suffix": "" } ], "year": 2015, "venue": "IEEE Conference on Computer Vision and Pattern Recognition (CVPR)", "volume": "", "issue": "", "pages": "", "other_ids": { "DOI": [ "10.1109/CVPR.2015.7298682" ] }, "num": null, "urls": [], "raw_text": "Florian Schroff, Dmitry Kalenichenko, and James Philbin. 2015. FaceNet: A unified embedding for face recognition and clustering. In IEEE Confer- ence on Computer Vision and Pattern Recognition (CVPR).", "links": null }, "BIBREF42": { "ref_id": "b42", "title": "Learning structured output representation using deep conditional generative models", "authors": [ { "first": "Kihyuk", "middle": [], "last": "Sohn", "suffix": "" }, { "first": "Honglak", "middle": [], "last": "Lee", "suffix": "" }, { "first": "Xinchen", "middle": [], "last": "Yan", "suffix": "" } ], "year": 2015, "venue": "Advances in neural information processing systems", "volume": "", "issue": "", "pages": "3483--3491", "other_ids": {}, "num": null, "urls": [], "raw_text": "Kihyuk Sohn, Honglak Lee, and Xinchen Yan. 2015. Learning structured output representation using deep conditional generative models. In Advances in neural information processing systems, pages 3483- 3491.", "links": null }, "BIBREF43": { "ref_id": "b43", "title": "BERT rediscovers the classical NLP pipeline", "authors": [ { "first": "Ian", "middle": [], "last": "Tenney", "suffix": "" }, { "first": "Dipanjan", "middle": [], "last": "Das", "suffix": "" }, { "first": "Ellie", "middle": [], "last": "Pavlick", "suffix": "" } ], "year": 2019, "venue": "Proceedings of the Conference of the Association for Computational Linguistics, ACL", "volume": "", "issue": "", "pages": "4593--4601", "other_ids": {}, "num": null, "urls": [], "raw_text": "Ian Tenney, Dipanjan Das, and Ellie Pavlick. 2019a. BERT rediscovers the classical NLP pipeline. In Proceedings of the Conference of the Association for Computational Linguistics, ACL, pages 4593-4601.", "links": null }, "BIBREF44": { "ref_id": "b44", "title": "What do you learn from context? probing for sentence structure in contextualized word representations", "authors": [ { "first": "Ian", "middle": [], "last": "Tenney", "suffix": "" }, { "first": "Patrick", "middle": [], "last": "Xia", "suffix": "" }, { "first": "Berlin", "middle": [], "last": "Chen", "suffix": "" }, { "first": "Alex", "middle": [], "last": "Wang", "suffix": "" }, { "first": "Adam", "middle": [], "last": "Poliak", "suffix": "" }, { "first": "Thomas", "middle": [], "last": "Mccoy", "suffix": "" }, { "first": "Najoung", "middle": [], "last": "Kim", "suffix": "" }, { "first": "Benjamin", "middle": [], "last": "Van Durme", "suffix": "" }, { "first": "Sam", "middle": [], "last": "Bowman", "suffix": "" }, { "first": "Dipanjan", "middle": [], "last": "Das", "suffix": "" }, { "first": "Ellie", "middle": [], "last": "Pavlick", "suffix": "" } ], "year": 2019, "venue": "International Conference on Learning Representations", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Ian Tenney, Patrick Xia, Berlin Chen, Alex Wang, Adam Poliak, R Thomas McCoy, Najoung Kim, Benjamin Van Durme, Sam Bowman, Dipanjan Das, and Ellie Pavlick. 2019b. What do you learn from context? probing for sentence structure in contextu- alized word representations. In International Con- ference on Learning Representations.", "links": null }, "BIBREF45": { "ref_id": "b45", "title": "The information bottleneck method", "authors": [ { "first": "Naftali", "middle": [], "last": "Tishby", "suffix": "" }, { "first": "C", "middle": [], "last": "Fernando", "suffix": "" }, { "first": "William", "middle": [], "last": "Pereira", "suffix": "" }, { "first": "", "middle": [], "last": "Bialek", "suffix": "" } ], "year": 1999, "venue": "Proc. of the Allerton Allerton Conference on Communication, Control and Computing", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Naftali Tishby, Fernando C Pereira, and William Bialek. 1999. The information bottleneck method. In Proc. of the Allerton Allerton Conference on Com- munication, Control and Computing.", "links": null }, "BIBREF46": { "ref_id": "b46", "title": "Modeling garden path effects without explicit hierarchical syntax", "authors": [ { "first": "Marten", "middle": [], "last": "Van Schijndel", "suffix": "" }, { "first": "Tal", "middle": [], "last": "Linzen", "suffix": "" } ], "year": 2018, "venue": "Proceedings of the 40th Annual Conference of the Cognitive Science Society", "volume": "", "issue": "", "pages": "2600--2605", "other_ids": {}, "num": null, "urls": [], "raw_text": "Marten van Schijndel and Tal Linzen. 2018. Modeling garden path effects without explicit hierarchical syn- tax. In Proceedings of the 40th Annual Conference of the Cognitive Science Society, pages 2600-2605. Cognitive Science Society.", "links": null }, "BIBREF47": { "ref_id": "b47", "title": "What do RNN language models learn about filler-gap dependencies?", "authors": [ { "first": "Ethan", "middle": [], "last": "Wilcox", "suffix": "" }, { "first": "Roger", "middle": [], "last": "Levy", "suffix": "" }, { "first": "Takashi", "middle": [], "last": "Morita", "suffix": "" }, { "first": "Richard", "middle": [], "last": "Futrell", "suffix": "" } ], "year": 2018, "venue": "Proceedings of the EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP", "volume": "", "issue": "", "pages": "211--221", "other_ids": {}, "num": null, "urls": [], "raw_text": "Ethan Wilcox, Roger Levy, Takashi Morita, and Richard Futrell. 2018. What do RNN language mod- els learn about filler-gap dependencies? In Proceed- ings of the EMNLP Workshop BlackboxNLP: Ana- lyzing and Interpreting Neural Networks for NLP, pages 211-221. Association for Computational Lin- guistics.", "links": null }, "BIBREF48": { "ref_id": "b48", "title": "End-to-end open-domain question answering with BERTserini", "authors": [ { "first": "Wei", "middle": [], "last": "Yang", "suffix": "" }, { "first": "Yuqing", "middle": [], "last": "Xie", "suffix": "" }, { "first": "Aileen", "middle": [], "last": "Lin", "suffix": "" }, { "first": "Xingyu", "middle": [], "last": "Li", "suffix": "" }, { "first": "Luchen", "middle": [], "last": "Tan", "suffix": "" }, { "first": "Kun", "middle": [], "last": "Xiong", "suffix": "" }, { "first": "Ming", "middle": [], "last": "Li", "suffix": "" }, { "first": "Jimmy", "middle": [], "last": "Lin", "suffix": "" } ], "year": 2019, "venue": "Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics NAACL-HLT", "volume": "", "issue": "", "pages": "72--77", "other_ids": {}, "num": null, "urls": [], "raw_text": "Wei Yang, Yuqing Xie, Aileen Lin, Xingyu Li, Luchen Tan, Kun Xiong, Ming Li, and Jimmy Lin. 2019. End-to-end open-domain question answering with BERTserini. In Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics NAACL-HLT, pages 72- 77.", "links": null } }, "ref_entries": { "FIGREF0": { "num": null, "uris": null, "text": "An illustration of triplet-loss calculation.", "type_str": "figure" }, "FIGREF1": { "num": null, "uris": null, "text": ": t-SNE projection of ELMO states, colored by syntactic function, before (upper) and after (lower) the syntactic transformation.", "type_str": "figure" }, "FIGREF2": { "num": null, "uris": null, "text": "Results of the few-shots parsing setup.", "type_str": "figure" }, "TABREF1": { "num": null, "html": null, "text": "Randomly selected examples of groups of structurally-similar sentences ( \u00a73.1)", "content": "", "type_str": "table" } } } }