Edit model card

Models Trained with Human Selection

This is a collection of the language models trained using Pile-CC, each with approximately 1B parameters, trained on different seeds. This project aims to validate the generalization capabilities of the RegMix approach (https://huggingface.co/papers/2407.01492) from small-scale (e.g., 1M parameters) to large-scale (e.g., 1B parameters) models.

Key Features

  • Model Size: 5 separate models trained with different seeds, each with ~1B parameters
  • Training Data: The pile-cc only data mixture on the RegMix-Data dataset

Dataset

The models were trained using the RegMix-Data dataset, which is split into different domains from The Pile dataset.

Training Hyperparameters

Hyperparameter Value
Batch Size 1M tokens
Learning Rate 4e-4
Minimum Learning Rate 1e-5
Learning Rate Schedule Cosine
Warmup Ratio 4%
Total Tokens 25B

How to Load a Model

You can load any model using the corresponding branch with the Hugging Face Transformers library:

from transformers import AutoModel, AutoTokenizer

model = AutoModel.from_pretrained("sail/data-mixture-pile-cc-1b", revision="seed-1")
tokenizer = AutoTokenizer.from_pretrained("sail/data-mixture-pile-cc-1b", revision="seed-1")

Data Mixture

The specific data mixture used for training this 1B model is as follows, which can be also found in our code:

train:
  train_the_pile_pile_cc: 1.0
valid:
  valid_the_pile_pile_cc: 1.0
model_name: tinyllama_1_1b

Model Variants

To access different model variants, simply change the revision parameter in the from_pretrained method to the desired seed (e.g., "seed-2", "seed-3"), and the maxium seed is 5.

Model Performance

We evaluated each model using lm-evaluation-harness. The performance metric for each task is the average of 0-shot to 5-shot accnorm (accuracy normalized, if available) or acc (accuracy) scores.

Seed PIQA LAMBADA MultiRC LogiQA SocialIQA Winogrande RACE OpenBookQA COPA HellaSwag SciQ ARC Easy QQP Average
1 69.23 33.16 50.33 27.57 33.22 52.10 31.80 31.07 65.83 44.15 81.77 51.80 57.04 48.39
2 68.62 33.69 53.15 25.13 32.96 51.24 31.06 30.84 69.80 43.28 83.18 52.00 58.06 48.69
3 69.04 35.68 52.38 26.36 33.45 51.95 30.83 30.16 66.80 42.80 83.32 51.57 57.69 48.62
4 69.35 33.56 50.01 26.24 33.62 50.99 31.81 30.44 65.60 43.00 83.00 52.33 56.14 48.16
5 67.91 35.09 49.93 27.50 33.90 52.85 31.77 30.04 69.40 42.62 80.94 51.25 61.03 48.79

Usage Notes

  • These models are primarily intended for research purposes.
  • Performance may vary depending on the specific task and domain.

Citation

If you use these models in your research, please cite the RegMix paper:

@article{liu2024regmix,
  title={RegMix: Data Mixture as Regression for Language Model Pre-training},
  author={Liu, Qian and Zheng, Xiaosen and Muennighoff, Niklas and Zeng, Guangtao and Dou, Longxu and Pang, Tianyu and Jiang, Jing and Lin, Min},
  journal={arXiv preprint arXiv:2407.01492},
  year={2024}
}

For more information about the RegMix methodology and its applications, please refer to the original paper.

Downloads last month
4
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Datasets used to train sail/data-mixture-pile-cc-1b

Collection including sail/data-mixture-pile-cc-1b