Add reports from Benchmarking paper "Benchmark Leakage in Large Language Models"

#27
by SinclairWang - opened

What are you reporting:

  • Evaluation dataset(s) found in a pre-training corpus. (e.g. COPA found in ThePile)
  • Evaluation dataset(s) found in a pre-trained model. (e.g. FLAN T5 has been trained on ANLI)

Evaluation dataset(s): Name(s) of the evaluation dataset(s). If available in the HuggingFace Hub please write the path (e.g. uonlp/CulturaX), otherwise provide a link to a paper, GitHub or dataset-card.

gsm8k

hendrycks/competition_math

Contaminated model(s): Name of the model(s) (if any) that have been contaminated with the evaluation dataset. If available in the HuggingFace Hub please list the corresponding paths (e.g. allenai/OLMo-7B).

Contaminated model in gsm8k:

Qwen/Qwen-1_8B, Qwen/Qwen-14B

Contaminated model in hendrycks/competition_math:

BAAI/Aquila2-34B, BAAI/Aquila2-7B, Qwen/Qwen-1_8B, Qwen/Qwen-7B, Qwen/Qwen-14B, THUDM/chatglm3-6b, internlm/internlm2-7b, internlm/internlm2-20b

Contaminated corpora: Name of the corpora used to pretrain models (if any) that have been contaminated with the evaluation dataset. If available in the HuggingFace hub please write the path (e.g. CohereForAI/aya_dataset)

None

Contaminated split(s): If the dataset has Train, Development and/or Test splits please report the contaminated split(s). You can report a percentage of the dataset contaminated; if the entire dataset is compromised, report 100%.

The train and test split. Specific percentage numbers are update in the CSV file.

image.png

Refer to Benchmarking Benchmark Leakage in Large Language Models (https://arxiv.org/abs/2404.18824) for more detailed results.

You may also report instances where there is no contamination. In such cases, follow the previous instructions but report a contamination level of 0%.

Briefly describe your method to detect data contamination

  • Data-based approach
  • Model-based approach

Description of your method, 3-4 sentences. Evidence of data contamination (Read below):

N-gram Accuracy. Specifically, combine the problem and solution into a combined text. Then, evenly select k (e.g., 5) starting points within this sequence. Use the content before each selected point as a prompt to predict the subsequent n-grams (e.g., 5-grams). If the model can correctly predict all the n-grams (measured by a ROUGE-L score greater than 0.75), it indicates that the model has encountered this sample during training. Refer to the paper "Benchmarking Benchmark Leakage in Large Language Models" (https://arxiv.org/pdf/2404.18824) for more details.

image.png

Data-based approaches

Data-based approaches identify evidence of data contamination in a pre-training corpus by directly examining the dataset for instances of the evaluation data. This method involves algorithmically searching through a large pre-training dataset to find occurrences of the evaluation data. You should provide evidence of data contamination in the form: "dataset X appears in line N of corpus Y," "dataset X appears N times in corpus Y," or "N examples from dataset X appear in corpus Y."

Model-based approaches

Model-based approaches, on the other hand, utilize heuristic algorithms to infer the presence of data contamination in a pre-trained model. These methods do not directly analyze the data but instead assess the model's behavior to predict data contamination. Examples include prompting the model to reproduce elements of an evaluation dataset to demonstrate memorization (i.e https://hitz-zentroa.github.io/lm-contamination/blog/) or using perplexity measures to estimate data contamination (). You should provide evidence of data contamination in the form of evaluation results of the algorithm from research papers, screenshots of model outputs that demonstrate memorization of a pre-training dataset, or any other form of evaluation that substantiates the method's effectiveness in detecting data contamination. You can provide a confidence score in your predictions.

Citation

Is there a paper that reports the data contamination or describes the method used to detect data contamination?

URL: https://arxiv.org/pdf/2404.18824
Citation: @article{xu2024benchmarking, title={Benchmarking Benchmark Leakage in Large Language Models}, author={Xu, Ruijie and Wang, Zengzhi and Fan, Run-Ze and Liu, Pengfei}, year={2024}, journal={arXiv preprint arXiv:2404.18824}, url={https://arxiv.org/abs/2404.18824} }

Important! If you wish to be listed as an author in the final report, please complete this information for all the authors of this Pull Request.

Note that all listed contributors are the authors of the reference paper (Benchmarking Benchmark Leakage in Large Language Models).

OSainz changed pull request title from Update contamination_report.csv to Add reports from Benchmarking paper "Benchmark Leakage in Large Language Models"
Workshop on Data Contamination org

Hi @SinclairWang !

Thank you for your contribution! Merging to main :)

Best,
Oscar

OSainz changed pull request status to merged

Sign up or log in to comment