--- dataset_info: features: - name: meta struct: - name: arxiv_id dtype: string - name: language dtype: string - name: timestamp dtype: string - name: url dtype: string - name: yymm dtype: string - name: text dtype: string splits: - name: train num_bytes: 857168232 num_examples: 13155 download_size: 382068275 dataset_size: 857168232 --- # ArXiv papers from RedPajama-Data originally published in February 2023 We collect the ArXiv papers released shortly before the training data cutoff date for the [OpenLLaMA models](https://huggingface.co/openlm-research/open_llama_7b). The OpenLLaMA models (V1) have been trained on [RedPajama data](https://huggingface.co/datasets/togethercomputer/RedPajama-Data-1T). The last batch of ArXiv papers included in this dataset are papers published in February 2023. To get the members close to the cutoff data, we collect the 13,155 papers published in "2302" as part of the training dataset. We process the raw LateX files using this [script](https://github.com/togethercomputer/RedPajama-Data/blob/rp_v1/data_prep/arxiv/run_clean.py). This dataset has been used as source for 'member' documents to develop (document-level) MIAs against LLMs using data collected shortly before (member) and after (non-member) the training cutoff date for the target model ([the suite of OpenLLaMA models](https://huggingface.co/openlm-research/open_llama_7b)). For non-members for the RDD setup, we refer to our [Github repo](https://github.com/computationalprivacy/mia_llms_benchmark/tree/main/document_level). For more details and results see the section of Regression Discontiuity Design (RDD) in the paper ["SoK: Membership Inference Attacks on LLMs are Rushing Nowhere (and How to Fix It)"](https://arxiv.org/pdf/2406.17975).