File size: 9,940 Bytes
7e60a5e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
import os

import datasets
import pandas as pd
import sacrebleu as scb
from packaging import version
from sacrebleu import CHRF
import string

import evaluate

_CITATION = """\
@ARTICLE{2023arXiv230513252W,
       author = {{Weller}, Orion and {Marone}, Marc and {Weir}, Nathaniel and {Lawrie}, Dawn and {Khashabi}, Daniel and {Van Durme}, Benjamin},
        title = "{``According to ...'' Prompting Language Models Improves Quoting from Pre-Training Data}",
      journal = {arXiv e-prints},
     keywords = {Computer Science - Computation and Language, Computer Science - Artificial Intelligence},
         year = 2023,
        month = may,
          eid = {arXiv:2305.13252},
        pages = {arXiv:2305.13252},
          doi = {10.48550/arXiv.2305.13252},
archivePrefix = {arXiv},
       eprint = {2305.13252},
 primaryClass = {cs.CL},
       adsurl = {https://ui.adsabs.harvard.edu/abs/2023arXiv230513252W},
      adsnote = {Provided by the SAO/NASA Astrophysics Data System}
}
"""

_DESCRIPTION = """\
In order to understand whether models are able
to ground to their pre-training data, we first need
to have a way of measuring this phenomena. We
adopt a narrow definition of grounding (quoting
from source material) while acknowledging that
grounding is a broad term.
To enable fast and efficient measurement of
quoting from pre-training data for many language
model generations across large corpora, we build
off of a D ATA P ORTRAIT (Marone and Van Durme,
2023), which allows for fast membership queries
for each n-gram in the output. This approach en-
ables us to perform a one-time indexing of a large
corpus (e.g. Wikipedia) and at inference time sim-
ply compute a constant time lookup operation (in
milliseconds) for each n-gram in the generation.
We build a D ATA P ORTRAIT on the version of
Wikipedia included in the Pile, 2 as it allows for
us to exactly test the pre-training data included
in many models like GPT-J and is similar to the
training data used in T5. However, we note that for
some models evaluated in this paper (e.g. OpenAI
models) there is no public information about the
Wikipedia version in the models.
We use character based n-grams as opposed to a
token-based n-gram as different models have differ-
ent tokenization schemes; furthermore, character-
based n-gram metrics have widespread usage in
fields such as machine translation with metrics like
chrF and chrF++ (Popović, 2015, 2017). We use
25 character grams for the sketch, approximately 5-
gram words, as we found it empirically gave mean-
ingful results (not too small of an n-gram and not
too large). The D ATA P ORTRAIT checks for exact
matches and is sensitive to orthographic variation
(e.g. case, whitespace). Therefore we view this as
a lower-bound on actual quoting performance.
We define our new metric QUIP-Score as the
character n-gram precision of the generated out-
put compared to the pre-training corpus. More
formally, for generation Y and text corpus C:
P
gram n ∈Y 1 C (gram n )
QUIP(Y ; C) =
,
|gram n ∈ Y |
where 1(.) is an indicator function: 1 if gram n ∈ C
else 0. Thus, a score of 0.5 would indicate that
50% of the generated text n-grams are found in
the pre-training corpus. We macro-average this
quantity over a set of generations to obtain a single
performance number for a given test dataset. 3
"""

_KWARGS_DESCRIPTION = """
Produces QUIP scores for checking grounding from references
Args:
    predictions (list of str): The predicted sentences.
    references (list of list of str): The references. There should be one reference sub-list for each prediction sentence.
Returns:
    'score' (float): The QUIP score,
Examples:
    Example 1--a simple example of calculating chrF:
    predictions = ["The current goodwill balance is $25,173 million as of December 31, 2022."]
    references = [[
                      "Table 7.3: Goodwill (in millions) Consumer Banking and Lending Commercial Banking Corporate and Investment Banking Wealth and Investment Management Corporate Consolidated Company December 31, 2020 $ 16,418 3,018 5,375 1,276 305 26,392 Foreign currency translation β€” β€” β€” β€” β€” β€” Transfers of goodwill β€” (80) β€” (932) 1,012 β€” Divestitures β€” β€” β€” β€” (1,212) (1,212) December 31, 2021 $ 16,418 2,938 5,375 344 105 25,180 Foreign currency translation β€” (7) β€” β€” β€” (7) December 31, 2022 $ 16,418 2,931 5,375 344 105 25,173 Table 7.4 presents the components of other assets."]]
    results = quip.compute(predictions=predictions, references=references, return_match_fraction_by_pred_length=True)
    print(results)
    assert results == 0.5
"""


@evaluate.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION)
class Quip(evaluate.Metric):
    def __init__(self, **kwargs):

        self.set_common = None
        if False:
            common_words_file = "data/NGSL_1.2_stats.csv.zip"
            if os.path.isfile(common_words_file):
                df = pd.read_csv(common_words_file)
                self.set_common = set(df['Lemma'].values.tolist())
        else:
            # https://norvig.com/ngrams/count_1w.txt
            common_words_file = "data/count_1w.txt.zip"
            if os.path.isfile(common_words_file):
                df = pd.read_csv(common_words_file, names=["word", "freq"], header=None, sep='\t')
                df = df.head(1000)
                self.set_common = set(df['word'].values.tolist())
                for k in list(string.ascii_lowercase):
                    keep = {'i', 'I', 'A', 'a'}
                    if k in self.set_common:
                        if k in keep:
                            continue
                        self.set_common.remove(k)

        super().__init__(**kwargs)

    def _info(self):
        if version.parse(scb.__version__) < version.parse("1.4.12"):
            raise ImportWarning(
                "To use `quip`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn't match this condition.\n"
                'You can install it with `pip install "sacrebleu>=1.4.12"`.'
            )
        return evaluate.MetricInfo(
            description=_DESCRIPTION,
            citation=_CITATION,
            homepage="https://github.com/h2oai/h2ogpt",
            inputs_description=_KWARGS_DESCRIPTION,
            features=[
                datasets.Features(
                    {
                        "predictions": datasets.Value("string", id="sequence"),
                        "references": datasets.Sequence(datasets.Value("string", id="sequence"), id="references"),
                    }
                ),
                datasets.Features(
                    {
                        "predictions": datasets.Value("string", id="sequence"),
                        "references": datasets.Value("string", id="sequence"),
                    }
                ),
            ],
            codebase_urls=["https://github.com/h2oai/h2ogpt"],
            reference_urls=[
                "https://github.com/h2oai/h2ogpt",
            ],
        )

    def _compute(
            self,
            predictions=None,
            references=None,
            reduced=True,
            min_len=2,
            max_len=5,
            return_match_count=False,
            return_match_fraction_by_pred_length=False,
            **kwargs,
    ):
        # if only one reference is provided make sure we still use list of lists
        if isinstance(references[0], str):
            references = [[ref] for ref in references]
        references_per_prediction = len(references[0])
        if any(len(refs) != references_per_prediction for refs in references):
            raise ValueError(
                "Quip requires the same number of references for each prediction"
            )
        # transformed_references = [[refs[i] for refs in references] for i in range(references_per_prediction)]

        if reduced:
            punc = """"!"#$%&()*+,-./:;<=>?@[\\]^_{|}~"""

            for predi, pred in enumerate(predictions):
                pred = pred.translate(str.maketrans(punc, ' ' * len(punc))).strip()
                predictions[predi] = ' '.join([x for x in pred.split() if x not in self.set_common])

            for refi, refl in enumerate(references):
                for refj, ref in enumerate(refl):
                    ref = ref.translate(str.maketrans(punc, ' ' * len(punc))).strip()
                    references[refi][refj] = ' '.join([x for x in ref.split() if x not in self.set_common])

        from nltk.util import everygrams
        from utils import flatten_list
        pred_ngrams = set(
            flatten_list([list(everygrams(x.split(), min_len=min_len, max_len=max_len)) for x in predictions]))
        ref_ngrams = set(flatten_list(
            [[list(everygrams(y.split(), min_len=min_len, max_len=max_len)) for y in z] for z in references]))
        residual = pred_ngrams.difference(ref_ngrams)
        if return_match_count:
            return len(pred_ngrams) - len(residual)
        else:
            if not return_match_fraction_by_pred_length:
                # Score = 0.0: No match
                # Score = 1.0: Perfect match
                return 1.0 - len(residual) / len(pred_ngrams)
            else:
                # FIXME: only works with 1 prediction
                nmatches = len(pred_ngrams) - len(residual)
                return min(1.0, nmatches / len(predictions[0].split()))

    def get_reduced_size(self, reduced_query, verbose=True):
        reduced_query_words = reduced_query.split(' ')
        set_common = set(self.df['Lemma'].values.tolist())
        num_common = len([x.lower() in set_common for x in reduced_query_words])
        frac_common = num_common / len(reduced_query) if reduced_query else 0
        # FIXME: report to user bad query that uses too many common words
        if verbose:
            print("frac_common: %s" % frac_common, flush=True)