| from __future__ import annotations |
|
|
| import argparse |
| import datetime as dt |
| import os |
| import re |
| import sys |
| import time |
| import urllib.error |
| import urllib.request |
| from concurrent.futures import ThreadPoolExecutor, as_completed |
| from dataclasses import dataclass |
| from pathlib import Path |
|
|
| BASE_URL_DEFAULT = "https://axavier.org/UnitCommitment.jl/0.4/instances" |
|
|
|
|
| @dataclass(frozen=True) |
| class Task: |
| case: str |
| date: dt.date |
| url: str |
| out_path: Path |
|
|
|
|
| def _iter_dates(start: dt.date, end: dt.date): |
| if end < start: |
| raise ValueError("end_date must be >= start_date") |
| cur = start |
| one = dt.timedelta(days=1) |
| while cur <= end: |
| yield cur |
| cur += one |
|
|
|
|
| def _default_instances_md_path() -> Path: |
| here = Path(__file__).resolve() |
| project_root = here.parents[2] |
| return project_root / "docs" / "src" / "guides" / "instances.md" |
|
|
|
|
| def _parse_matpower_cases_from_instances_md(instances_md_path: Path) -> list[str]: |
| text = instances_md_path.read_text(encoding="utf-8") |
| pattern = re.compile(r"matpower/(case[^/\s`]+)/\d{4}-\d{2}-\d{2}") |
| cases = sorted(set(pattern.findall(text))) |
| if not cases: |
| raise RuntimeError(f"No MATPOWER cases found in {instances_md_path}") |
| return cases |
|
|
|
|
| def _download_one(task: Task, timeout_s: float, retries: int, force: bool) -> tuple[str, str | None]: |
| task.out_path.parent.mkdir(parents=True, exist_ok=True) |
|
|
| if task.out_path.exists() and not force and task.out_path.stat().st_size > 0: |
| return ("skipped", None) |
|
|
| tmp_path = task.out_path.with_suffix(task.out_path.suffix + ".part") |
| req = urllib.request.Request(task.url, headers={"User-Agent": "UnitCommitment downloader"}) |
|
|
| last_err = None |
| for attempt in range(retries + 1): |
| try: |
| with urllib.request.urlopen(req, timeout=timeout_s) as resp: |
| with open(tmp_path, "wb") as f: |
| while True: |
| chunk = resp.read(1024 * 256) |
| if not chunk: |
| break |
| f.write(chunk) |
| os.replace(tmp_path, task.out_path) |
| return ("downloaded", None) |
| except (urllib.error.URLError, urllib.error.HTTPError, TimeoutError, OSError) as e: |
| last_err = e |
| try: |
| if tmp_path.exists(): |
| tmp_path.unlink() |
| except OSError: |
| pass |
| if attempt < retries: |
| time.sleep(min(10.0, 0.5 * (2**attempt))) |
| continue |
| return ("failed", f"{type(last_err).__name__}: {last_err}") |
|
|
| return ("failed", f"{type(last_err).__name__}: {last_err}") |
|
|
|
|
| def _build_tasks( |
| base_url: str, |
| out_dir: Path, |
| cases: list[str], |
| start_date: dt.date, |
| end_date: dt.date, |
| ) -> list[Task]: |
| tasks: list[Task] = [] |
| for case in cases: |
| for d in _iter_dates(start_date, end_date): |
| date_str = d.isoformat() |
| url = f"{base_url}/matpower/{case}/{date_str}.json.gz" |
| out_path = out_dir / case / f"{date_str}.json.gz" |
| tasks.append(Task(case=case, date=d, url=url, out_path=out_path)) |
| return tasks |
|
|
|
|
| def _parse_date(s: str) -> dt.date: |
| return dt.date.fromisoformat(s) |
|
|
|
|
| def main(argv: list[str]) -> int: |
| parser = argparse.ArgumentParser() |
| parser.add_argument("--instances-md", type=Path, default=_default_instances_md_path()) |
| parser.add_argument("--base-url", type=str, default=BASE_URL_DEFAULT) |
| parser.add_argument( |
| "--out-dir", |
| type=Path, |
| default=(Path(__file__).resolve().parents[2] / "instances" / "matpower"), |
| ) |
| parser.add_argument("--start-date", type=_parse_date, default=dt.date(2017, 1, 1)) |
| parser.add_argument("--end-date", type=_parse_date, default=dt.date(2017, 12, 31)) |
| parser.add_argument("--workers", type=int, default=min(32, (os.cpu_count() or 4) * 4)) |
| parser.add_argument("--timeout", type=float, default=60.0) |
| parser.add_argument("--retries", type=int, default=3) |
| parser.add_argument("--force", action="store_true") |
| parser.add_argument("--dry-run", action="store_true") |
| parser.add_argument("--list-cases", action="store_true") |
| args = parser.parse_args(argv) |
|
|
| instances_md_path: Path = args.instances_md |
| if not instances_md_path.exists(): |
| raise FileNotFoundError(str(instances_md_path)) |
|
|
| cases = _parse_matpower_cases_from_instances_md(instances_md_path) |
| if args.list_cases: |
| for c in cases: |
| print(c) |
| return 0 |
|
|
| out_dir: Path = args.out_dir |
| tasks = _build_tasks( |
| base_url=args.base_url.rstrip("/"), |
| out_dir=out_dir, |
| cases=cases, |
| start_date=args.start_date, |
| end_date=args.end_date, |
| ) |
|
|
| print(f"cases={len(cases)} files={len(tasks)} out_dir={out_dir}") |
| if args.dry_run: |
| for t in tasks[:10]: |
| print(f"{t.url} -> {t.out_path}") |
| if len(tasks) > 10: |
| print("...") |
| return 0 |
|
|
| downloaded = 0 |
| skipped = 0 |
| failed = 0 |
| total = len(tasks) |
| t0 = time.time() |
| last_print = 0.0 |
|
|
| with ThreadPoolExecutor(max_workers=max(1, args.workers)) as ex: |
| fut_to_task = { |
| ex.submit(_download_one, t, args.timeout, args.retries, args.force): t for t in tasks |
| } |
| for i, fut in enumerate(as_completed(fut_to_task), start=1): |
| status, err = fut.result() |
| if status == "downloaded": |
| downloaded += 1 |
| elif status == "skipped": |
| skipped += 1 |
| else: |
| failed += 1 |
| task = fut_to_task[fut] |
| sys.stderr.write(f"FAILED {task.url} -> {task.out_path} ({err})\n") |
|
|
| now = time.time() |
| if now - last_print >= 1.0 or i == total: |
| elapsed = max(0.001, now - t0) |
| rate = i / elapsed |
| print( |
| f"{i}/{total} downloaded={downloaded} skipped={skipped} failed={failed} rate={rate:.1f}/s" |
| ) |
| last_print = now |
|
|
| if failed: |
| return 2 |
| return 0 |
|
|
|
|
| if __name__ == "__main__": |
| raise SystemExit(main(sys.argv[1:])) |
|
|