--- dataset_info: features: - name: input_ids sequence: int32 splits: - name: train num_bytes: 35168157552.0 num_examples: 1073116 download_size: 18303136476 dataset_size: 35168157552.0 configs: - config_name: default data_files: - split: train path: data/train-* --- # OpenWebTextCorpus tokenized for Gemma This dataset is a pre-tokenized version of the [Skylion007/openwebtext](https://huggingface.co/datasets/Skylion007/openwebtext) dataset using the [gemma](https://huggingface.co/google/gemma-2b) tokenizer. As such, this dataset follows the same licensing as the original openwebtext dataset. This pre-tokenization is done as a performance optimization for using the openwebtext dataset with a Gemma model (gemma-2b, gemma-2b-it, gemma-7b, gemma-7b-it). This dataset was created using [SAELens](https://github.com/jbloomAus/SAELens), with the following settings: - context_size: 8192 - shuffled: true - begin_batch_token: "bos" - begin_sequence_token: null - sequence_separator_token: "bos" - sae_lens_version: "3.3.0"