Datasets:
Formats:
parquet
Languages:
English
Size:
10M - 100M
Tags:
biology
chemistry
drug-discovery
clinical-trials
protein-protein-interaction
gene-essentiality
License:
| #!/usr/bin/env python3 | |
| """Unified training harness for PPI baseline models. | |
| Supports SiameseCNN, PIPR, MLPFeatures on PPI M1 binary task. | |
| Usage: | |
| PYTHONPATH=src python scripts_ppi/train_baseline.py \ | |
| --model siamese_cnn \ | |
| --split random \ | |
| --negative negbiodb \ | |
| --dataset balanced \ | |
| --epochs 100 --patience 10 \ | |
| --batch_size 256 --lr 0.001 --seed 42 | |
| Outputs: | |
| results/ppi_baselines/{model}_{dataset}_{split}_{negative}_seed{seed}/ | |
| best.pt — best model checkpoint (val LogAUC) | |
| results.json — test-set metrics | |
| training_log.csv — per-epoch train/val metrics | |
| """ | |
| from __future__ import annotations | |
| import argparse | |
| import csv | |
| import json | |
| import logging | |
| import random | |
| import sys | |
| from pathlib import Path | |
| import numpy as np | |
| import pandas as pd | |
| logging.basicConfig( | |
| level=logging.INFO, | |
| format="%(asctime)s %(levelname)s %(message)s", | |
| datefmt="%H:%M:%S", | |
| stream=sys.stderr, | |
| ) | |
| logger = logging.getLogger(__name__) | |
| ROOT = Path(__file__).parent.parent | |
| # Parquet filenames by (dataset, negative) key | |
| _DATASET_MAP: dict[tuple[str, str], str] = { | |
| ("balanced", "negbiodb"): "ppi_m1_balanced.parquet", | |
| ("realistic", "negbiodb"): "ppi_m1_realistic.parquet", | |
| ("balanced", "uniform_random"): "ppi_m1_uniform_random.parquet", | |
| ("balanced", "degree_matched"): "ppi_m1_degree_matched.parquet", | |
| ("balanced", "ddb"): "ppi_m1_balanced_ddb.parquet", | |
| } | |
| # Split column name by split type | |
| _SPLIT_COL_MAP: dict[str, str] = { | |
| "random": "split_random", | |
| "cold_protein": "split_cold_protein", | |
| "cold_both": "split_cold_both", | |
| "ddb": "split_degree_balanced", | |
| } | |
| def set_seed(seed: int) -> None: | |
| random.seed(seed) | |
| np.random.seed(seed) | |
| try: | |
| import torch | |
| torch.manual_seed(seed) | |
| if torch.cuda.is_available(): | |
| torch.cuda.manual_seed_all(seed) | |
| except ImportError: | |
| pass | |
| def _resolve_dataset_file(dataset: str, split: str, negative: str) -> str | None: | |
| """Resolve parquet filename for a valid experiment configuration.""" | |
| if split == "ddb": | |
| if negative != "negbiodb": | |
| return None | |
| return _DATASET_MAP.get((dataset, "ddb")) | |
| if negative in {"uniform_random", "degree_matched"} and dataset != "balanced": | |
| return None | |
| return _DATASET_MAP.get((dataset, negative)) | |
| def _build_run_name(model: str, dataset: str, split: str, negative: str, seed: int) -> str: | |
| return f"{model}_{dataset}_{split}_{negative}_seed{seed}" | |
| def _json_safe(value): | |
| if isinstance(value, dict): | |
| return {k: _json_safe(v) for k, v in value.items()} | |
| if isinstance(value, (list, tuple)): | |
| return [_json_safe(v) for v in value] | |
| if isinstance(value, (np.floating, float)): | |
| value = float(value) | |
| return value if np.isfinite(value) else None | |
| if isinstance(value, np.integer): | |
| return int(value) | |
| return value | |
| def write_results_json(path: Path, payload: dict) -> None: | |
| with open(path, "w") as f: | |
| json.dump(_json_safe(payload), f, indent=2, allow_nan=False) | |
| # --------------------------------------------------------------------------- | |
| # Dataset | |
| # --------------------------------------------------------------------------- | |
| class PPIDataset: | |
| """PPI dataset backed by a Parquet file.""" | |
| def __init__( | |
| self, | |
| parquet_path: Path, | |
| split_col: str, | |
| fold: str, | |
| model_type: str, | |
| ) -> None: | |
| self.model_type = model_type | |
| df_full = pd.read_parquet(parquet_path) | |
| # Recompute pair degrees from the full merged graph to avoid | |
| # pos/neg asymmetry (DB negatives have degree, positives/controls don't). | |
| if model_type == "mlp_features": | |
| all_ids = pd.concat([df_full["uniprot_id_1"], df_full["uniprot_id_2"]]) | |
| degree_map = all_ids.value_counts() | |
| df_full["protein1_degree"] = df_full["uniprot_id_1"].map(degree_map).astype(float) | |
| df_full["protein2_degree"] = df_full["uniprot_id_2"].map(degree_map).astype(float) | |
| # For cold_both, NaN fold = excluded pairs (dropped by == comparison) | |
| self.df = df_full[df_full[split_col] == fold].reset_index(drop=True) | |
| before = len(self.df) | |
| self.df = self.df.dropna(subset=["sequence_1", "sequence_2"]).reset_index(drop=True) | |
| if len(self.df) < before: | |
| logger.warning("Dropped %d rows with NaN sequences", before - len(self.df)) | |
| logger.info("Fold '%s': %d rows (label 1: %d)", fold, len(self.df), (self.df["Y"] == 1).sum()) | |
| def __len__(self) -> int: | |
| return len(self.df) | |
| def __getitem__(self, idx: int): | |
| row = self.df.iloc[idx] | |
| label = float(row["Y"]) | |
| if self.model_type == "mlp_features": | |
| return ( | |
| row["sequence_1"], row["sequence_2"], | |
| 0.0 if pd.isna(row.get("protein1_degree")) else float(row["protein1_degree"]), | |
| 0.0 if pd.isna(row.get("protein2_degree")) else float(row["protein2_degree"]), | |
| row.get("subcellular_location_1"), | |
| row.get("subcellular_location_2"), | |
| label, | |
| ) | |
| else: | |
| return row["sequence_1"], row["sequence_2"], label | |
| def _collate_sequence_pair(batch, device): | |
| """Collate for SiameseCNN and PIPR (sequence-based models).""" | |
| import torch | |
| from negbiodb_ppi.models.siamese_cnn import seq_to_tensor | |
| seqs1, seqs2, labels = zip(*batch) | |
| return ( | |
| seq_to_tensor(list(seqs1)).to(device), | |
| seq_to_tensor(list(seqs2)).to(device), | |
| torch.tensor(labels, dtype=torch.float32).to(device), | |
| ) | |
| def _collate_features(batch, device): | |
| """Collate for MLPFeatures (feature-based model).""" | |
| import torch | |
| from negbiodb_ppi.models.mlp_features import extract_features | |
| features_list = [] | |
| labels = [] | |
| for seq1, seq2, deg1, deg2, loc1, loc2, label in batch: | |
| features_list.append(extract_features(seq1, seq2, deg1, deg2, loc1, loc2)) | |
| labels.append(label) | |
| return ( | |
| torch.tensor(features_list, dtype=torch.float32).to(device), | |
| None, # placeholder for consistency | |
| torch.tensor(labels, dtype=torch.float32).to(device), | |
| ) | |
| def make_dataloader(dataset: PPIDataset, batch_size: int, shuffle: bool, device): | |
| from torch.utils.data import DataLoader | |
| if dataset.model_type == "mlp_features": | |
| collate_fn = lambda b: _collate_features(b, device) # noqa: E731 | |
| else: | |
| collate_fn = lambda b: _collate_sequence_pair(b, device) # noqa: E731 | |
| return DataLoader( | |
| dataset, | |
| batch_size=batch_size, | |
| shuffle=shuffle, | |
| num_workers=0, | |
| collate_fn=collate_fn, | |
| ) | |
| # --------------------------------------------------------------------------- | |
| # Training | |
| # --------------------------------------------------------------------------- | |
| def _run_epoch(model, loader, criterion, optimizer, device, train: bool, model_type: str): | |
| import torch | |
| model.train(train) | |
| total_loss = 0.0 | |
| all_labels: list[float] = [] | |
| all_preds: list[float] = [] | |
| n_batches = 0 | |
| with torch.set_grad_enabled(train): | |
| for batch in loader: | |
| if model_type == "mlp_features": | |
| features, _, labels = batch | |
| logits = model(features) | |
| else: | |
| seq1_tokens, seq2_tokens, labels = batch | |
| logits = model(seq1_tokens, seq2_tokens) | |
| loss = criterion(logits, labels) | |
| if train: | |
| optimizer.zero_grad() | |
| loss.backward() | |
| optimizer.step() | |
| total_loss += loss.item() | |
| all_labels.extend(labels.cpu().tolist()) | |
| all_preds.extend(torch.sigmoid(logits).detach().cpu().tolist()) | |
| n_batches += 1 | |
| avg_loss = total_loss / max(n_batches, 1) | |
| return avg_loss, np.array(all_labels), np.array(all_preds) | |
| def _compute_val_metric(y_true, y_score): | |
| from negbiodb.metrics import log_auc | |
| if len(np.unique(y_true)) < 2: | |
| return float("nan") | |
| return log_auc(y_true, y_score) | |
| def train_model( | |
| model, | |
| train_loader, | |
| val_loader, | |
| epochs: int, | |
| patience: int, | |
| lr: float, | |
| output_dir: Path, | |
| device, | |
| model_type: str, | |
| ): | |
| import torch | |
| import torch.nn as nn | |
| criterion = nn.BCEWithLogitsLoss() | |
| optimizer = torch.optim.Adam(model.parameters(), lr=lr) | |
| best_val_metric = float("-inf") | |
| patience_counter = 0 | |
| training_log = [] | |
| for epoch in range(1, epochs + 1): | |
| train_loss, _, _ = _run_epoch( | |
| model, train_loader, criterion, optimizer, device, train=True, model_type=model_type | |
| ) | |
| val_loss, val_y, val_pred = _run_epoch( | |
| model, val_loader, criterion, optimizer, device, train=False, model_type=model_type | |
| ) | |
| val_metric = _compute_val_metric(val_y, val_pred) | |
| row = {"epoch": epoch, "train_loss": train_loss, "val_loss": val_loss, "val_log_auc": val_metric} | |
| training_log.append(row) | |
| logger.info( | |
| "Epoch %3d | train_loss=%.4f | val_loss=%.4f | val_LogAUC=%.4f", | |
| epoch, train_loss, val_loss, val_metric, | |
| ) | |
| if not np.isnan(val_metric) and val_metric > best_val_metric: | |
| best_val_metric = val_metric | |
| patience_counter = 0 | |
| torch.save(model.state_dict(), output_dir / "best.pt") | |
| logger.info(" -> Saved best checkpoint (val_LogAUC=%.4f)", best_val_metric) | |
| else: | |
| patience_counter += 1 | |
| if patience_counter >= patience: | |
| logger.info("Early stopping at epoch %d (patience=%d)", epoch, patience) | |
| break | |
| log_path = output_dir / "training_log.csv" | |
| with open(log_path, "w", newline="") as f: | |
| writer = csv.DictWriter(f, fieldnames=training_log[0].keys()) | |
| writer.writeheader() | |
| writer.writerows(training_log) | |
| logger.info("Training log saved -> %s", log_path) | |
| return best_val_metric | |
| def evaluate(model, test_loader, checkpoint_path: Path, device, model_type: str) -> dict[str, float]: | |
| import torch | |
| from negbiodb.metrics import compute_all_metrics | |
| state = torch.load(checkpoint_path, map_location=device, weights_only=True) | |
| model.load_state_dict(state) | |
| model.eval() | |
| criterion = torch.nn.BCEWithLogitsLoss() | |
| _, y_true, y_score = _run_epoch( | |
| model, test_loader, criterion, None, device, train=False, model_type=model_type | |
| ) | |
| if len(np.unique(y_true)) < 2: | |
| logger.warning("Test set has only one class — metrics will be NaN.") | |
| return {k: float("nan") for k in ["auroc", "auprc", "mcc", "log_auc", "bedroc", "ef_1pct", "ef_5pct"]} | |
| return compute_all_metrics(y_true, y_score) | |
| # --------------------------------------------------------------------------- | |
| # Model factory | |
| # --------------------------------------------------------------------------- | |
| def build_model(model_type: str): | |
| if model_type == "siamese_cnn": | |
| from negbiodb_ppi.models.siamese_cnn import SiameseCNN | |
| return SiameseCNN() | |
| elif model_type == "pipr": | |
| from negbiodb_ppi.models.pipr import PIPR | |
| return PIPR() | |
| elif model_type == "mlp_features": | |
| from negbiodb_ppi.models.mlp_features import MLPFeatures | |
| return MLPFeatures() | |
| else: | |
| raise ValueError(f"Unknown model: {model_type!r}") | |
| # --------------------------------------------------------------------------- | |
| # Main | |
| # --------------------------------------------------------------------------- | |
| def main(argv: list[str] | None = None) -> int: | |
| parser = argparse.ArgumentParser(description="Train a PPI baseline model.") | |
| parser.add_argument("--model", required=True, | |
| choices=["siamese_cnn", "pipr", "mlp_features"]) | |
| parser.add_argument("--split", required=True, choices=list(_SPLIT_COL_MAP)) | |
| parser.add_argument("--negative", required=True, | |
| choices=["negbiodb", "uniform_random", "degree_matched"]) | |
| parser.add_argument("--dataset", default="balanced", | |
| choices=["balanced", "realistic"]) | |
| parser.add_argument("--epochs", type=int, default=100) | |
| parser.add_argument("--patience", type=int, default=10) | |
| parser.add_argument("--batch_size", type=int, default=256) | |
| parser.add_argument("--lr", type=float, default=1e-3) | |
| parser.add_argument("--seed", type=int, default=42) | |
| parser.add_argument("--data_dir", type=Path, | |
| default=ROOT / "exports" / "ppi") | |
| parser.add_argument("--output_dir", type=Path, | |
| default=ROOT / "results" / "ppi_baselines") | |
| args = parser.parse_args(argv) | |
| set_seed(args.seed) | |
| if args.split == "ddb" and args.dataset != "balanced": | |
| logger.error("DDB split only supported for dataset=balanced.") | |
| return 1 | |
| filename = _resolve_dataset_file(args.dataset, args.split, args.negative) | |
| if filename is None: | |
| logger.error( | |
| "Invalid combination: dataset=%s, split=%s, negative=%s", | |
| args.dataset, args.split, args.negative, | |
| ) | |
| return 1 | |
| parquet_path = args.data_dir / filename | |
| if not parquet_path.exists(): | |
| logger.error("Dataset not found: %s", parquet_path) | |
| logger.error("Run `scripts_ppi/prepare_exp_data.py` first.") | |
| return 1 | |
| split_col = _SPLIT_COL_MAP[args.split] | |
| run_name = _build_run_name( | |
| args.model, args.dataset, args.split, args.negative, args.seed | |
| ) | |
| out_dir = args.output_dir / run_name | |
| out_dir.mkdir(parents=True, exist_ok=True) | |
| logger.info("Run: %s -> %s", run_name, out_dir) | |
| # Device | |
| try: | |
| import torch | |
| if torch.cuda.is_available(): | |
| device = torch.device("cuda") | |
| elif hasattr(torch.backends, "mps") and torch.backends.mps.is_available(): | |
| device = torch.device("mps") | |
| else: | |
| device = torch.device("cpu") | |
| except ImportError: | |
| logger.error("torch not found.") | |
| return 1 | |
| logger.info("Device: %s", device) | |
| # Build datasets | |
| train_ds = PPIDataset(parquet_path, split_col, "train", args.model) | |
| val_ds = PPIDataset(parquet_path, split_col, "val", args.model) | |
| test_ds = PPIDataset(parquet_path, split_col, "test", args.model) | |
| if len(train_ds) == 0 or len(val_ds) == 0 or len(test_ds) == 0: | |
| logger.error("Empty split. Check split_col=%s in %s.", split_col, parquet_path.name) | |
| return 1 | |
| train_loader = make_dataloader(train_ds, args.batch_size, shuffle=True, device=device) | |
| val_loader = make_dataloader(val_ds, args.batch_size, shuffle=False, device=device) | |
| test_loader = make_dataloader(test_ds, args.batch_size, shuffle=False, device=device) | |
| model = build_model(args.model).to(device) | |
| n_params = sum(p.numel() for p in model.parameters() if p.requires_grad) | |
| logger.info("Model: %s | params: %d", args.model, n_params) | |
| best_val = train_model( | |
| model, train_loader, val_loader, | |
| epochs=args.epochs, patience=args.patience, lr=args.lr, | |
| output_dir=out_dir, device=device, model_type=args.model, | |
| ) | |
| checkpoint = out_dir / "best.pt" | |
| if not checkpoint.exists(): | |
| logger.warning("No checkpoint saved — all epochs produced NaN val LogAUC.") | |
| null_metrics = {k: None for k in [ | |
| "log_auc", "auprc", "bedroc", "ef_1pct", "ef_5pct", "mcc", "auroc", | |
| ]} | |
| results = { | |
| "run_name": run_name, | |
| "model": args.model, | |
| "split": args.split, | |
| "negative": args.negative, | |
| "dataset": args.dataset, | |
| "seed": args.seed, | |
| "best_val_log_auc": None, | |
| "test_metrics": null_metrics, | |
| "n_train": len(train_ds), | |
| "n_val": len(val_ds), | |
| "n_test": len(test_ds), | |
| } | |
| results_path = out_dir / "results.json" | |
| write_results_json(results_path, results) | |
| logger.info("Null results saved → %s", results_path) | |
| return 0 | |
| test_metrics = evaluate(model, test_loader, checkpoint, device, args.model) | |
| results = { | |
| "run_name": run_name, | |
| "model": args.model, | |
| "split": args.split, | |
| "negative": args.negative, | |
| "dataset": args.dataset, | |
| "seed": args.seed, | |
| "best_val_log_auc": best_val, | |
| "test_metrics": test_metrics, | |
| "n_train": len(train_ds), | |
| "n_val": len(val_ds), | |
| "n_test": len(test_ds), | |
| } | |
| results_path = out_dir / "results.json" | |
| write_results_json(results_path, results) | |
| logger.info("Results saved -> %s", results_path) | |
| for metric, value in test_metrics.items(): | |
| logger.info(" %s: %.4f", metric, value if value is not None else float("nan")) | |
| return 0 | |
| if __name__ == "__main__": | |
| sys.exit(main()) | |