| |
| |
| |
| |
| |
| |
|
|
| """Build and deploy an Embedding Atlas end-to-end. |
| |
| Orchestrates the full pipeline: |
| 1. Creates a storage bucket (if needed) |
| 2. Submits a GPU Job to build the atlas (embedding + UMAP) |
| 3. Waits for the Job to complete |
| 4. Deploys a Docker Space that serves the atlas from the bucket |
| |
| ⚠️ EXPERIMENTAL — this workflow is new and may change. |
| |
| Examples: |
| |
| # Minimal — from HF dataset to deployed Space |
| uv run atlas-e2e.py stanfordnlp/imdb \\ |
| --text text --split train \\ |
| --name imdb-atlas --sample 50000 |
| |
| # From prepped parquet (already in a bucket) |
| uv run atlas-e2e.py hf://buckets/user/atlas-data/books.parquet \\ |
| --text title --name open-library-atlas --sample 2000000 |
| |
| # Full control |
| uv run atlas-e2e.py my-org/my-dataset \\ |
| --text text --split train \\ |
| --name my-atlas \\ |
| --sample 1000000 \\ |
| --bucket user/atlas-data \\ |
| --space-id user/my-atlas-viz \\ |
| --flavor a100-large \\ |
| --timeout 2h \\ |
| --batch-size 512 |
| """ |
|
|
| import argparse |
| import subprocess |
| import sys |
| import time |
| from pathlib import Path |
|
|
| from huggingface_hub import HfApi, create_bucket |
|
|
|
|
| def wait_for_job(api: HfApi, job_id: str, poll_interval: int = 30) -> str: |
| """Poll a Job until it completes. Returns the final stage.""" |
| print(f"\nWaiting for Job {job_id}...") |
| while True: |
| job = api.inspect_job(job_id=job_id) |
| stage = job.status.stage |
| if stage in ("COMPLETED", "ERROR"): |
| msg = job.status.message or "" |
| print(f"Job {stage}" + (f": {msg}" if msg else "")) |
| return stage |
| time.sleep(poll_interval) |
|
|
|
|
| def main(): |
| parser = argparse.ArgumentParser( |
| description="Build and deploy an Embedding Atlas end-to-end (experimental)", |
| formatter_class=argparse.RawDescriptionHelpFormatter, |
| epilog=__doc__, |
| ) |
|
|
| |
| parser.add_argument("input", help="HF dataset ID or parquet path") |
| parser.add_argument("--name", required=True, help="Atlas name") |
| parser.add_argument("--text", default="text", help="Text column name") |
|
|
| |
| parser.add_argument("--split", default=None, help="Dataset split") |
| parser.add_argument("--sample", type=int, default=None, help="Number of rows") |
| parser.add_argument("--image", default=None, help="Image column name") |
| parser.add_argument("--model", default=None, help="Embedding model") |
|
|
| |
| parser.add_argument("--bucket", default=None, help="Bucket ID (default: {user}/atlas-data)") |
| parser.add_argument("--space-id", default=None, help="Space ID (default: {user}/{name})") |
| parser.add_argument("--flavor", default="a100-large", help="Job GPU flavor (default: a100-large)") |
| parser.add_argument("--timeout", default="2h", help="Job timeout (default: 2h)") |
| parser.add_argument("--batch-size", type=int, default=256, help="Embedding batch size") |
| parser.add_argument("--space-hardware", default="cpu-basic", help="Space hardware (default: cpu-basic)") |
| parser.add_argument("--private", action="store_true", help="Make Space private") |
|
|
| |
| parser.add_argument("--build-only", action="store_true", help="Only build, don't deploy Space") |
| parser.add_argument("--deploy-only", action="store_true", help="Only deploy from existing bucket data") |
|
|
| args = parser.parse_args() |
|
|
| api = HfApi() |
| user = api.whoami()["name"] |
|
|
| |
| if args.bucket is None: |
| args.bucket = f"{user}/atlas-data" |
| if args.space_id is None: |
| args.space_id = f"{user}/{args.name}" |
|
|
| print("=" * 60) |
| print("Embedding Atlas — End-to-End Pipeline") |
| print("=" * 60) |
| print(f"Input: {args.input}") |
| print(f"Name: {args.name}") |
| print(f"Bucket: {args.bucket}") |
| print(f"Space: {args.space_id}") |
| print(f"Flavor: {args.flavor}") |
| print(f"Sample: {args.sample}") |
| print("=" * 60) |
|
|
| |
| if not args.deploy_only: |
| print(f"\n[1/3] Creating bucket {args.bucket}...") |
| create_bucket(args.bucket, exist_ok=True) |
|
|
| |
| if not args.deploy_only: |
| print(f"\n[2/3] Submitting build Job ({args.flavor})...") |
|
|
| build_script = Path(__file__).parent / "atlas-build-gpu.py" |
| if not build_script.exists(): |
| print(f"ERROR: {build_script} not found") |
| print("atlas-build-gpu.py must be in the same directory as this script") |
| sys.exit(1) |
|
|
| |
| cmd = [ |
| "hf", "jobs", "uv", "run", |
| "--flavor", args.flavor, |
| "-v", f"hf://buckets/{args.bucket}:/data", |
| "-s", "HF_TOKEN", |
| "--timeout", args.timeout, |
| str(build_script), |
| args.input, |
| "--name", args.name, |
| "--text", args.text, |
| "--batch-size", str(args.batch_size), |
| ] |
|
|
| if args.split: |
| cmd.extend(["--split", args.split]) |
| if args.sample: |
| cmd.extend(["--sample", str(args.sample)]) |
| if args.image: |
| cmd.extend(["--image", args.image]) |
| if args.model: |
| cmd.extend(["--model", args.model]) |
|
|
| print(f"Command: {' '.join(cmd)}\n") |
|
|
| |
| result = subprocess.run(cmd, capture_output=True, text=True) |
| output = result.stdout + result.stderr |
|
|
| |
| job_id = None |
| for line in output.split("\n"): |
| if "Job started with ID:" in line: |
| job_id = line.split("Job started with ID:")[-1].strip() |
| break |
|
|
| if job_id is None: |
| print("ERROR: Could not extract Job ID from output:") |
| print(output) |
| sys.exit(1) |
|
|
| print(f"Job submitted: {job_id}") |
| print(f"View: https://huggingface.co/jobs/{user}/{job_id}") |
|
|
| |
| stage = wait_for_job(api, job_id) |
| if stage != "COMPLETED": |
| print(f"\nJob failed. Check logs: https://huggingface.co/jobs/{user}/{job_id}") |
| sys.exit(1) |
|
|
| print("Build complete!") |
|
|
| if args.build_only: |
| print(f"\nBuild finished. Data in bucket: {args.bucket}/{args.name}/") |
| print(f"Deploy later with: uv run atlas-deploy.py --name {args.name} --bucket {args.bucket}") |
| return |
|
|
| |
| print(f"\n[3/3] Deploying Space {args.space_id}...") |
|
|
| deploy_script = Path(__file__).parent / "atlas-deploy.py" |
| if not deploy_script.exists(): |
| print(f"ERROR: {deploy_script} not found") |
| sys.exit(1) |
|
|
| deploy_cmd = [ |
| "uv", "run", |
| str(deploy_script), |
| "--name", args.name, |
| "--bucket", args.bucket, |
| "--space-id", args.space_id, |
| "--hardware", args.space_hardware, |
| "--text-column", args.text, |
| ] |
| if args.private: |
| deploy_cmd.append("--private") |
|
|
| subprocess.run(deploy_cmd, check=True) |
|
|
| print("\n" + "=" * 60) |
| print("Done!") |
| print(f"Space: https://huggingface.co/spaces/{args.space_id}") |
| print(f"Bucket: https://huggingface.co/buckets/{args.bucket}") |
| print("=" * 60) |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|