| | |
| | """Download 300GB from HuggingFace datasets - FineWeb, RedPajama, etc.""" |
| | import os |
| | import json |
| | from datasets import load_dataset |
| | from huggingface_hub import snapshot_download |
| | import time |
| |
|
| | TARGET_GB = 300 |
| | OUTPUT_DIR = "/workspace/scraped_data" |
| | os.makedirs(OUTPUT_DIR, exist_ok=True) |
| |
|
| | |
| | DATASETS = [ |
| | ("HuggingFaceFW/fineweb", "sample-10BT", None), |
| | ("togethercomputer/RedPajama-Data-V2", "sample", "en_head_middle"), |
| | ("allenai/dolma", "v1_6-sample", None), |
| | ("cerebras/SlimPajama-627B", None, None), |
| | ] |
| |
|
| | def get_size_gb(): |
| | total = 0 |
| | for root, dirs, files in os.walk(OUTPUT_DIR): |
| | for f in files: |
| | total += os.path.getsize(os.path.join(root, f)) |
| | return total / 1e9 |
| |
|
| | def download_streaming(name, config, split): |
| | """Stream download to avoid OOM""" |
| | print(f"\n📥 Downloading {name} ({config or 'default'})...") |
| | try: |
| | ds = load_dataset(name, config, split=split or "train", streaming=True, trust_remote_code=True) |
| | |
| | shard_num = 0 |
| | batch = [] |
| | batch_size = 10000 |
| | |
| | for i, example in enumerate(ds): |
| | text = example.get("text") or example.get("content") or str(example) |
| | batch.append({"text": text, "source": name}) |
| | |
| | if len(batch) >= batch_size: |
| | outfile = f"{OUTPUT_DIR}/{name.replace('/', '_')}_{shard_num:05d}.jsonl" |
| | with open(outfile, 'w') as f: |
| | for item in batch: |
| | f.write(json.dumps(item) + "\n") |
| | batch = [] |
| | shard_num += 1 |
| | |
| | size_gb = get_size_gb() |
| | print(f" Progress: {size_gb:.1f} GB / {TARGET_GB} GB ({i:,} examples)") |
| | |
| | if size_gb >= TARGET_GB: |
| | print(f"✅ Target reached!") |
| | return True |
| | |
| | except Exception as e: |
| | print(f" Error: {e}") |
| | return False |
| |
|
| | if __name__ == "__main__": |
| | print(f"🚀 Goddess HF Scraper - Target: {TARGET_GB} GB") |
| | print(f"Output: {OUTPUT_DIR}") |
| | |
| | start = time.time() |
| | |
| | for name, config, split in DATASETS: |
| | if get_size_gb() >= TARGET_GB: |
| | break |
| | done = download_streaming(name, config, split) |
| | if done: |
| | break |
| | |
| | elapsed = time.time() - start |
| | final_size = get_size_gb() |
| | print(f"\n✨ Done! {final_size:.1f} GB in {elapsed/3600:.1f} hours") |
| |
|