Datasets:

Modalities:
Text
Formats:
parquet
ArXiv:
Libraries:
Datasets
Dask
matthieumeeus97's picture
Update README.md
bcde28e verified
---
dataset_info:
features:
- name: meta
struct:
- name: arxiv_id
dtype: string
- name: language
dtype: string
- name: timestamp
dtype: string
- name: url
dtype: string
- name: yymm
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 857168232
num_examples: 13155
download_size: 382068275
dataset_size: 857168232
---
# ArXiv papers from RedPajama-Data originally published in February 2023
We collect the ArXiv papers released shortly before the training data cutoff date for the [OpenLLaMA models](https://huggingface.co/openlm-research/open_llama_7b).
The OpenLLaMA models (V1) have been trained on [RedPajama data](https://huggingface.co/datasets/togethercomputer/RedPajama-Data-1T).
The last batch of ArXiv papers included in this dataset are papers published in February 2023.
To get the members close to the cutoff data, we collect the 13,155 papers published in "2302" as part of the training dataset.
We process the raw LateX files using this [script](https://github.com/togethercomputer/RedPajama-Data/blob/rp_v1/data_prep/arxiv/run_clean.py).
This dataset has been used as source for 'member' documents to develop (document-level) MIAs against LLMs using data collected shortly before (member) and after (non-member) the training cutoff date for the target model ([the suite of OpenLLaMA models](https://huggingface.co/openlm-research/open_llama_7b)).
For non-members for the RDD setup, we refer to our [Github repo](https://github.com/computationalprivacy/mia_llms_benchmark/tree/main/document_level).
For more details and results see the section of Regression Discontiuity Design (RDD) in the paper ["SoK: Membership Inference Attacks on LLMs are Rushing Nowhere (and How to Fix It)"](https://arxiv.org/pdf/2406.17975).