|
|
| import argparse |
| import sys |
| import os |
| import torch |
| import torch.nn as nn |
| from torch.utils.data import DataLoader |
| import numpy as np |
| import pandas as pd |
| from tqdm import tqdm |
| from pathlib import Path |
| from transformers import AutoTokenizer |
|
|
| |
| sys.path.append(os.path.join(os.path.dirname(__file__))) |
|
|
| from dataset import ProteinTaxonomyDataset |
| from model import TaxonomyAwareESM |
| from asymmetric_loss import load_ia_weights |
|
|
| def evaluate_gpu(model, dataloader, ic_weights, device, thresholds=None, pred_output_path=None, metrics_output_path=None): |
| """ |
| Calculates Weighted F-max and S-min using GPU streaming to avoid OOM. |
| (Copied from train.py) |
| """ |
| model.eval() |
| |
| if thresholds is None: |
| thresholds = torch.linspace(0, 1, 101, device=device) |
| |
| |
| sum_prec = torch.zeros(len(thresholds), device=device) |
| sum_rec = torch.zeros(len(thresholds), device=device) |
| sum_ru = torch.zeros(len(thresholds), device=device) |
| sum_mi = torch.zeros(len(thresholds), device=device) |
| |
| total_samples = 0 |
| |
| |
| f_pred = None |
| if pred_output_path: |
| os.makedirs(os.path.dirname(pred_output_path), exist_ok=True) |
| f_pred = open(pred_output_path, 'w') |
| idx_to_go = {v: k for k, v in dataloader.dataset.go_to_idx.items()} |
| |
| with torch.no_grad(): |
| for batch in tqdm(dataloader, desc="GPU Eval"): |
| input_ids = batch['input_ids'].to(device) |
| attention_mask = batch['attention_mask'].to(device) |
| tax_vector = batch['tax_vector'].to(device) |
| labels = batch['labels'].to(device) |
| entry_ids = batch['entry_id'] |
| |
| |
| if isinstance(entry_ids, str): |
| entry_ids = [entry_ids] |
| if not isinstance(entry_ids, (list, tuple)): |
| if isinstance(entry_ids, torch.Tensor): |
| entry_ids = entry_ids.tolist() |
| else: |
| entry_ids = list(entry_ids) |
| |
| |
| logits = model(input_ids, attention_mask, tax_vector) |
| probs = torch.sigmoid(logits) |
| |
| |
| if f_pred: |
| probs_cpu = probs.cpu().numpy() |
| for i, entry_id in enumerate(entry_ids): |
| indices = np.where(probs_cpu[i] > 0.01)[0] |
| for idx in indices: |
| term = idx_to_go[idx] |
| score = probs_cpu[i][idx] |
| f_pred.write(f"{entry_id}\t{term}\t{score:.4f}\n") |
| |
| |
| |
| true_ic = (labels * ic_weights).sum(dim=1) |
| true_ic = torch.maximum(true_ic, torch.tensor(1e-9, device=device)) |
| |
| |
| |
| probs_unsqueezed = probs.unsqueeze(1) |
| thresholds_unsqueezed = thresholds.view(1, -1, 1) |
| |
| pred_binary = (probs_unsqueezed >= thresholds_unsqueezed).float() |
| |
| labels_unsqueezed = labels.unsqueeze(1) |
| ic_weights_unsqueezed = ic_weights.view(1, 1, -1) |
| |
| |
| intersection_ic = (pred_binary * labels_unsqueezed * ic_weights_unsqueezed).sum(dim=2) |
| |
| |
| pred_ic = (pred_binary * ic_weights_unsqueezed).sum(dim=2) |
| |
| |
| precision = intersection_ic / (pred_ic + 1e-9) |
| |
| |
| recall = intersection_ic / (true_ic.view(-1, 1) + 1e-9) |
| |
| |
| ru = true_ic.view(-1, 1) - intersection_ic |
| ru = torch.clamp(ru, min=0.0) |
| |
| |
| mi = pred_ic - intersection_ic |
| mi = torch.clamp(mi, min=0.0) |
| |
| |
| sum_prec += precision.sum(dim=0) |
| sum_rec += recall.sum(dim=0) |
| sum_ru += ru.sum(dim=0) |
| sum_mi += mi.sum(dim=0) |
| |
| total_samples += input_ids.size(0) |
| |
| del logits, probs, pred_binary, intersection_ic, pred_ic, ru, mi |
| |
| if f_pred: |
| f_pred.close() |
| print(f"Saved predictions to {pred_output_path}") |
| |
| |
| avg_prec = sum_prec / total_samples |
| avg_rec = sum_rec / total_samples |
| avg_ru = sum_ru / total_samples |
| avg_mi = sum_mi / total_samples |
| |
| |
| f1_scores = 2 * avg_prec * avg_rec / (avg_prec + avg_rec + 1e-9) |
| best_fmax = f1_scores.max().item() |
| best_t_idx = f1_scores.argmax().item() |
| best_threshold_f = thresholds[best_t_idx].item() |
| |
| |
| s_scores = torch.sqrt(avg_ru**2 + avg_mi**2) |
| min_s = s_scores.min().item() |
| best_s_idx = s_scores.argmin().item() |
| best_threshold_s = thresholds[best_s_idx].item() |
| |
| metrics = { |
| 'fmax_w': best_fmax, |
| 'threshold_fmax': best_threshold_f, |
| 'smin': min_s, |
| 'threshold_smin': best_threshold_s, |
| } |
| |
| |
| if metrics_output_path: |
| metrics_data = { |
| 'threshold': thresholds.cpu().numpy(), |
| 'precision': avg_prec.cpu().numpy(), |
| 'recall': avg_rec.cpu().numpy(), |
| 'f1': f1_scores.cpu().numpy(), |
| 'ru': avg_ru.cpu().numpy(), |
| 'mi': avg_mi.cpu().numpy(), |
| 's': s_scores.cpu().numpy() |
| } |
| pd.DataFrame(metrics_data).to_csv(metrics_output_path, sep='\t', index=False) |
| print(f"Saved detailed metrics to {metrics_output_path}") |
|
|
| return metrics |
|
|
| def main(): |
| parser = argparse.ArgumentParser() |
| parser.add_argument("--data_path", type=str, required=True, help="Path to mounted dataset") |
| parser.add_argument("--output_dir", type=str, default="outputs", help="Directory for checkpoints and predictions") |
| parser.add_argument("--checkpoint", type=str, default="latest_model.pth", help="Checkpoint filename to load (in output_dir)") |
| parser.add_argument("--epoch", type=int, default=3, help="Epoch to associate with predictions") |
| parser.add_argument("--esm_model_name", type=str, default="facebook/esm2_t33_650M_UR50D", help="ESM model name") |
| parser.add_argument("--force_novel", action="store_true", help="Force re-evaluation of novel dataset") |
| |
| args = parser.parse_args() |
| |
| device = torch.device('cuda' if torch.cuda.is_available() else 'cpu') |
| print(f"Using device: {device}") |
| |
| output_dir = Path(args.output_dir) |
| data_path = Path(args.data_path) |
| |
| |
| val_novel_fasta = data_path / "validation_superset" / "validation_novel" / "validation_novel.fasta" |
| val_novel_term = data_path / "validation_superset" / "validation_novel" / "validation_novel_terms.tsv" |
| |
| val_homolog_fasta = data_path / "validation_superset" / "validation_homolog" / "validation_homolog.fasta" |
| val_homolog_term = data_path / "validation_superset" / "validation_homolog" / "validation_homolog_terms.tsv" |
| |
| species_vec = data_path / "taxon_embedding" / "species_vectors.tsv" |
| |
| |
| go_vocab_path = "src/go_terms.json" |
| if not os.path.exists(go_vocab_path): |
| go_vocab_path = "go_terms.json" |
| |
| ia_path = data_path / "IA.tsv" |
| go_matrix_path = data_path / "go_info" / "go_ancestor_matrix.npz" |
| go_mapping_path = data_path / "go_info" / "go_term_mappings.pkl" |
| |
| |
| print("Loading tokenizer...") |
| tokenizer = AutoTokenizer.from_pretrained(args.esm_model_name) |
| |
| |
| |
| print("Initializing mapping dataset...") |
| dummy_dataset = ProteinTaxonomyDataset( |
| val_novel_fasta, val_novel_term, species_vec, go_vocab_path, max_len=1024, esm_tokenizer=tokenizer, |
| go_matrix_path=str(go_matrix_path), go_mapping_path=str(go_mapping_path) |
| ) |
| |
| print("Initializing Model...") |
| model = TaxonomyAwareESM( |
| num_classes=dummy_dataset.num_classes, |
| pretrained_model_name=args.esm_model_name, |
| use_lora=True, |
| lora_rank=512, |
| vocab_sizes=dummy_dataset.vocab_sizes |
| ).to(device) |
| |
| |
| ckpt_path = output_dir / args.checkpoint |
| if not ckpt_path.exists(): |
| print(f"Error: Checkpoint not found at {ckpt_path}") |
| return |
| |
| print(f"Loading checkpoint from {ckpt_path}...") |
| checkpoint = torch.load(ckpt_path, map_location=device) |
| model.load_state_dict(checkpoint['model_state_dict']) |
| print(f"Model loaded (Epoch {checkpoint['epoch']})") |
| |
| |
| print("Loading IC weights...") |
| ic_weights = load_ia_weights( |
| str(ia_path) if ia_path.exists() else "IA.tsv", |
| dummy_dataset.go_to_idx, |
| dummy_dataset.num_classes |
| ).to(device) |
| |
| |
| |
| |
| novel_preds_path = output_dir / f"gpu_preds_novel_epoch_{args.epoch}.tsv" |
| novel_metrics_path = output_dir / f"metrics_novel_epoch_{args.epoch}.tsv" |
| |
| if novel_metrics_path.exists() and not args.force_novel: |
| print(f"Novel metrics already exist at {novel_metrics_path}. Skipping.") |
| with open(novel_metrics_path, 'r') as f: |
| print(f.read()) |
| else: |
| print("Evaluating Novel Dataset...") |
| val_novel_loader = DataLoader(dummy_dataset, batch_size=32, shuffle=False, num_workers=4, pin_memory=True) |
| metrics_novel = evaluate_gpu( |
| model, val_novel_loader, ic_weights, device, |
| pred_output_path=novel_preds_path, |
| metrics_output_path=novel_metrics_path |
| ) |
| print("Novel Metrics:", metrics_novel) |
| |
| |
| homolog_preds_path = output_dir / f"gpu_preds_homolog_epoch_{args.epoch}.tsv" |
| homolog_metrics_path = output_dir / f"metrics_homolog_epoch_{args.epoch}.tsv" |
| |
| print("Evaluating Homolog Dataset...") |
| val_homolog_dataset = ProteinTaxonomyDataset( |
| val_homolog_fasta, val_homolog_term, species_vec, go_vocab_path, max_len=1024, esm_tokenizer=tokenizer, |
| go_matrix_path=str(go_matrix_path), go_mapping_path=str(go_mapping_path) |
| ) |
| val_homolog_loader = DataLoader(val_homolog_dataset, batch_size=32, shuffle=False, num_workers=4, pin_memory=True) |
| |
| metrics_homolog = evaluate_gpu( |
| model, val_homolog_loader, ic_weights, device, |
| pred_output_path=homolog_preds_path, |
| metrics_output_path=homolog_metrics_path |
| ) |
| print("Homolog Metrics:", metrics_homolog) |
| |
| print("Evaluation All Done.") |
|
|
| if __name__ == "__main__": |
| main() |
|
|