dkumar15's picture
Upload training_code/model/data.py with huggingface_hub
eaf2faa verified
"""
Data pipeline: streams and tokenizes OpenWebText for pretraining.
Packs sequences to max_seq_len for efficiency (no padding waste).
"""
import os
import torch
from torch.utils.data import IterableDataset, DataLoader
from datasets import load_dataset
from transformers import AutoTokenizer
def get_tokenizer(name: str = "mistralai/Mistral-7B-v0.1"):
"""Use Mistral's tokenizer — 32k vocab, BPE, well-trained on diverse data."""
tok = AutoTokenizer.from_pretrained(name, use_fast=True)
if tok.pad_token is None:
tok.pad_token = tok.eos_token
return tok
class PackedPretrainDataset(IterableDataset):
"""
Streams text from HuggingFace dataset, tokenizes on the fly,
and packs into fixed-length sequences for maximum GPU utilization.
"""
def __init__(self, tokenizer, max_seq_len: int, split: str = "train", cache_dir: str = None, seed: int = 42):
self.tokenizer = tokenizer
self.max_seq_len = max_seq_len
self.split = split
self.cache_dir = cache_dir
self.seed = seed
self.eos_id = tokenizer.eos_token_id
def _token_stream(self):
ds = load_dataset(
"HuggingFaceFW/fineweb-edu",
name="sample-10BT",
split=self.split,
streaming=True,
cache_dir=self.cache_dir,
)
ds = ds.shuffle(seed=self.seed, buffer_size=10_000)
for example in ds:
text = example.get("text", "")
if len(text.strip()) < 50:
continue
token_ids = self.tokenizer.encode(text, add_special_tokens=False)
yield from token_ids
yield self.eos_id
def __iter__(self):
buffer = []
for token_id in self._token_stream():
buffer.append(token_id)
if len(buffer) == self.max_seq_len + 1:
input_ids = torch.tensor(buffer[:-1], dtype=torch.long)
labels = torch.tensor(buffer[1:], dtype=torch.long)
yield input_ids, labels
buffer = []
def create_dataloader(tokenizer, config, rank: int = 0, world_size: int = 1, seed_override: int = None):
seed = seed_override if seed_override is not None else config.seed
dataset = PackedPretrainDataset(
tokenizer=tokenizer,
max_seq_len=config.max_seq_len,
split="train",
cache_dir=config.data_cache_dir,
seed=seed + rank,
)
return DataLoader(
dataset,
batch_size=config.batch_size_per_gpu,
num_workers=config.num_workers,
pin_memory=True,
prefetch_factor=4,
)