depscreen / ml /scripts /kfold_evaluate.py
halsabbah's picture
style: apply ruff format to pass CI format check
95974bc
"""
K-Fold Cross-Validation harness for the DSM-5 symptom classifier.
Replaces single train/val/test split with stratified K-fold CV at the post level.
All metrics reported as mean ± std across folds.
Usage:
python kfold_evaluate.py [options]
Options:
--model-name: Base model (default: distilbert-base-uncased)
--k: Number of folds (default: 5)
--epochs: Training epochs per fold (default: 5)
--batch-size: Batch size (default: 16)
--lr: Learning rate (default: 2e-5)
--max-length: Max token length (default: 128)
--output-dir: Where to save CV results
"""
import argparse
import json
import logging
import os
import sys
from pathlib import Path
import numpy as np
import pandas as pd
import torch
import torch.nn as nn
from iterstrat.ml_stratifiers import MultilabelStratifiedKFold
from sklearn.metrics import accuracy_score, precision_recall_fscore_support
from sklearn.preprocessing import MultiLabelBinarizer
from torch.optim import AdamW
from torch.utils.data import DataLoader
from tqdm import tqdm
from transformers import AutoTokenizer, get_linear_schedule_with_warmup
# Reuse existing model/dataset/collate from training script
sys.path.insert(0, str(Path(__file__).parent))
# Reuse preprocessing functions
from preprocess_redsm5 import (
SYMPTOM_LABELS,
SYMPTOM_READABLE,
clean_sentence,
create_negative_samples,
create_positive_samples,
load_data,
)
from train_redsm5_model import SymptomClassifier, SymptomDataset, collate_fn
logging.basicConfig(level=logging.INFO)
logger = logging.getLogger(__name__)
# Silence tokenizer parallelism warnings
os.environ["TOKENIZERS_PARALLELISM"] = "false"
def build_post_label_matrix(df: pd.DataFrame) -> tuple[pd.DataFrame, np.ndarray]:
"""Build a multi-label binary matrix at the post level for stratification.
Each post gets a binary vector indicating which symptoms appear in it.
This lets iterstrat stratify by symptom distribution across posts.
Returns:
post_df: DataFrame with one row per unique post_id
label_matrix: (n_posts, n_classes) binary matrix
"""
label_names = sorted(SYMPTOM_LABELS.keys(), key=lambda x: SYMPTOM_LABELS[x])
# Group sentences by post, collect unique symptoms per post
post_symptoms = df.groupby("post_id")["label"].apply(set).reset_index()
post_symptoms.columns = ["post_id", "symptoms"]
# Build binary matrix
mlb = MultiLabelBinarizer(classes=label_names)
label_matrix = mlb.fit_transform(post_symptoms["symptoms"])
return post_symptoms, label_matrix
def compute_class_weights(train_df: pd.DataFrame) -> dict[int, float]:
"""Compute inverse-frequency class weights for a training fold."""
counts = train_df["label_id"].value_counts().sort_index()
total = len(train_df)
n_classes = len(counts)
weights = {}
for label_id, count in counts.items():
weights[int(label_id)] = total / (n_classes * count)
return weights
def train_one_fold(
train_df: pd.DataFrame,
val_df: pd.DataFrame,
model_name: str,
epochs: int,
batch_size: int,
lr: float,
max_length: int,
device: torch.device,
fold_idx: int,
soft_labels: list[list[float]] | None = None,
distill_alpha: float = 0.5,
distill_temperature: float = 3.0,
loss_type: str = "ce",
label_smoothing: float = 0.0,
focal_gamma: float = 2.0,
effective_num_beta: float = 0.999,
hard_label_classes: list[int] | None = None,
use_llrd: bool = False,
use_fgm: bool = False,
fgm_epsilon: float = 0.5,
use_rdrop: bool = False,
rdrop_alpha: float = 0.1,
**kwargs,
) -> dict:
"""Train and evaluate a single fold. Returns metrics dict.
Args:
soft_labels: If provided, uses knowledge distillation loss.
List of probability vectors aligned to train_df rows.
distill_alpha: Weight for hard labels (1-alpha for soft). Default 0.5.
distill_temperature: Softening temperature. Default 3.0.
"""
label_names = sorted(SYMPTOM_LABELS.keys(), key=lambda x: SYMPTOM_LABELS[x])
num_classes = len(label_names)
# Tokenizer
tokenizer = AutoTokenizer.from_pretrained(model_name)
# Datasets — use distillation-aware dataset if soft labels available
if soft_labels is not None:
from distillation_utils import DistillationDataset, DistillationLoss, collate_fn_distill
train_dataset = DistillationDataset(
train_df["clean_text"].tolist(),
train_df["label_id"].tolist(),
soft_labels,
tokenizer,
max_length,
)
train_collate = collate_fn_distill
else:
train_dataset = SymptomDataset(
train_df["clean_text"].tolist(),
train_df["label_id"].tolist(),
tokenizer,
max_length,
)
train_collate = collate_fn
val_dataset = SymptomDataset(
val_df["clean_text"].tolist(),
val_df["label_id"].tolist(),
tokenizer,
max_length,
)
num_workers = 0 if device.type == "mps" else 2
train_loader = DataLoader(
train_dataset,
batch_size=batch_size,
shuffle=True,
collate_fn=train_collate,
num_workers=num_workers,
)
val_loader = DataLoader(
val_dataset,
batch_size=batch_size,
collate_fn=collate_fn,
num_workers=num_workers,
)
# Model (fresh for each fold)
# Pooling strategy: "cls", "mean", or "cls_mean" (concatenated)
pooling = kwargs.get("pooling", "mean")
model = SymptomClassifier(num_classes=num_classes, model_name=model_name, pooling=pooling)
model.to(device)
# Class weights for this fold's training data
if loss_type == "effective_num" or loss_type == "focal":
from distillation_utils import compute_effective_number_weights
class_counts = train_df["label_id"].value_counts().to_dict()
weight_tensor = compute_effective_number_weights(class_counts, num_classes, effective_num_beta)
weight_tensor = weight_tensor.to(device)
logger.info(f" Using effective-number weights (β={effective_num_beta})")
else:
class_weights = compute_class_weights(train_df)
weight_tensor = torch.zeros(num_classes)
for label_id_str, weight in class_weights.items():
weight_tensor[int(label_id_str)] = weight
weight_tensor = weight_tensor.to(device)
# Loss function selection
if loss_type == "focal":
if soft_labels is not None:
from distillation_utils import FocalDistillationLoss
criterion = FocalDistillationLoss(
alpha=distill_alpha,
temperature=distill_temperature,
gamma=focal_gamma,
class_weights=weight_tensor,
label_smoothing=label_smoothing,
)
logger.info(
f" Using focal+distillation loss (γ={focal_gamma}, α={distill_alpha}, T={distill_temperature}, ls={label_smoothing})"
)
else:
from distillation_utils import FocalLoss
criterion = FocalLoss(gamma=focal_gamma, class_weights=weight_tensor, label_smoothing=label_smoothing)
logger.info(f" Using focal loss (γ={focal_gamma}, ls={label_smoothing})")
elif soft_labels is not None:
from distillation_utils import DistillationLoss
per_class_alpha = None
if hard_label_classes:
per_class_alpha = {cls_id: 1.0 for cls_id in hard_label_classes}
logger.info(f" Per-class hard labels for class IDs: {hard_label_classes}")
criterion = DistillationLoss(
alpha=distill_alpha,
temperature=distill_temperature,
class_weights=weight_tensor,
per_class_alpha=per_class_alpha,
)
if label_smoothing > 0:
criterion.ce_loss = nn.CrossEntropyLoss(weight=weight_tensor, label_smoothing=label_smoothing)
logger.info(
f" Using {loss_type}+distillation loss (α={distill_alpha}, T={distill_temperature}, ls={label_smoothing})"
)
else:
criterion = nn.CrossEntropyLoss(weight=weight_tensor, label_smoothing=label_smoothing)
logger.info(f" Using CE loss (ls={label_smoothing})")
# Optimizer — LLRD or standard AdamW
if use_llrd:
from distillation_utils import build_llrd_param_groups
param_groups = build_llrd_param_groups(model, lr=lr, decay_factor=0.8, weight_decay=0.01)
optimizer = AdamW(param_groups)
logger.info(" Using LLRD (decay=0.8, wd=0.01)")
else:
optimizer = AdamW(model.parameters(), lr=lr)
total_steps = len(train_loader) * epochs
scheduler = get_linear_schedule_with_warmup(
optimizer,
num_warmup_steps=total_steps // 10,
num_training_steps=total_steps,
)
# FGM adversarial training
fgm = None
if use_fgm:
from distillation_utils import FGM
fgm = FGM(model, epsilon=fgm_epsilon)
logger.info(f" Using FGM adversarial training (ε={fgm_epsilon})")
if use_rdrop:
from distillation_utils import compute_rdrop_loss
logger.info(f" Using R-Drop regularization (α={rdrop_alpha})")
# SWA: accumulate weights from good epochs
use_swa = kwargs.get("swa", False)
swa_start_epoch = max(2, epochs // 2) # Start averaging from halfway
swa_states = []
if use_swa:
logger.info(f" Using SWA (averaging from epoch {swa_start_epoch + 1})")
# Threshold tuning
use_threshold_tuning = kwargs.get("threshold_tuning", False)
# Training loop
best_val_f1 = 0
best_state = None
for epoch in range(epochs):
# Train
model.train()
total_loss = 0
for batch in tqdm(train_loader, desc=f"Fold {fold_idx + 1} Epoch {epoch + 1}", leave=False):
input_ids = batch["input_ids"].to(device)
attention_mask = batch["attention_mask"].to(device)
labels = batch["label"].to(device)
optimizer.zero_grad()
logits = model(input_ids, attention_mask)
# Primary loss
if soft_labels is not None and "soft_label" in batch:
teacher_soft = batch["soft_label"].to(device)
loss = criterion(logits, labels, teacher_soft)
else:
if isinstance(criterion, nn.CrossEntropyLoss):
loss = criterion(logits, labels)
else:
loss = criterion(logits, labels, None)
# R-Drop: second forward pass with different dropout mask
if use_rdrop:
logits2 = model(input_ids, attention_mask)
loss = loss + compute_rdrop_loss(logits, logits2, alpha=rdrop_alpha)
loss.backward()
# FGM: adversarial perturbation step
if fgm is not None:
fgm.attack()
logits_adv = model(input_ids, attention_mask)
if soft_labels is not None and "soft_label" in batch:
loss_adv = criterion(logits_adv, labels, teacher_soft)
else:
if isinstance(criterion, nn.CrossEntropyLoss):
loss_adv = criterion(logits_adv, labels)
else:
loss_adv = criterion(logits_adv, labels, None)
loss_adv.backward()
fgm.restore()
torch.nn.utils.clip_grad_norm_(model.parameters(), 1.0)
optimizer.step()
scheduler.step()
total_loss += loss.item()
# Validate
model.eval()
all_preds = []
all_labels = []
with torch.no_grad():
for batch in val_loader:
input_ids = batch["input_ids"].to(device)
attention_mask = batch["attention_mask"].to(device)
labels = batch["label"].to(device)
logits = model(input_ids, attention_mask)
preds = torch.argmax(logits, dim=1)
all_preds.extend(preds.cpu().numpy())
all_labels.extend(labels.cpu().numpy())
_, _, micro_f1, _ = precision_recall_fscore_support(all_labels, all_preds, average="micro")
_, _, macro_f1, _ = precision_recall_fscore_support(all_labels, all_preds, average="macro")
logger.info(
f" Fold {fold_idx + 1} Epoch {epoch + 1}: "
f"loss={total_loss / len(train_loader):.4f} "
f"micro_f1={micro_f1:.4f} macro_f1={macro_f1:.4f}"
)
if micro_f1 > best_val_f1:
best_val_f1 = micro_f1
best_state = {k: v.cpu().clone() for k, v in model.state_dict().items()}
# SWA: collect weights from later epochs
if use_swa and epoch >= swa_start_epoch:
swa_states.append({k: v.cpu().clone() for k, v in model.state_dict().items()})
# SWA: average collected weights
if use_swa and len(swa_states) >= 2:
avg_state = {}
for key in swa_states[0]:
avg_state[key] = torch.stack([s[key].float() for s in swa_states]).mean(0)
if swa_states[0][key].dtype != torch.float32:
avg_state[key] = avg_state[key].to(swa_states[0][key].dtype)
# Evaluate SWA model vs best checkpoint — keep whichever is better
model.load_state_dict(avg_state)
model.to(device)
model.eval()
swa_preds, swa_labels = [], []
with torch.no_grad():
for batch in val_loader:
logits = model(batch["input_ids"].to(device), batch["attention_mask"].to(device))
swa_preds.extend(torch.argmax(logits, dim=1).cpu().numpy())
swa_labels.extend(batch["label"].numpy())
_, _, swa_micro, _ = precision_recall_fscore_support(swa_labels, swa_preds, average="micro")
_, _, swa_macro, _ = precision_recall_fscore_support(swa_labels, swa_preds, average="macro")
logger.info(f" SWA: micro={swa_micro:.4f} macro={swa_macro:.4f} (best checkpoint: micro={best_val_f1:.4f})")
if swa_micro >= best_val_f1:
best_state = avg_state
logger.info(" SWA model selected (better than best checkpoint)")
else:
logger.info(" Best checkpoint kept (better than SWA)")
# Load best model and compute final metrics on this fold's val set
model.load_state_dict(best_state)
model.to(device)
model.eval()
all_probs = []
all_labels = []
with torch.no_grad():
for batch in val_loader:
input_ids = batch["input_ids"].to(device)
attention_mask = batch["attention_mask"].to(device)
labels = batch["label"].to(device)
logits = model(input_ids, attention_mask)
probs = torch.softmax(logits, dim=1)
all_probs.extend(probs.cpu().numpy())
all_labels.extend(labels.cpu().numpy())
all_probs = np.array(all_probs)
all_labels = np.array(all_labels)
# Threshold tuning: optimize per-class thresholds to maximize per-class F1
if use_threshold_tuning:
best_thresholds = np.full(num_classes, 0.0)
for cls_id in range(num_classes):
best_cls_f1 = -1
best_t = 0.0
cls_true = (all_labels == cls_id).astype(int)
if cls_true.sum() == 0:
continue
for t in np.arange(0.05, 0.95, 0.05):
cls_pred = (all_probs[:, cls_id] >= t).astype(int)
if cls_pred.sum() == 0:
continue
p, r, f, _ = precision_recall_fscore_support(cls_true, cls_pred, average="binary", zero_division=0)
if f > best_cls_f1:
best_cls_f1 = f
best_t = t
best_thresholds[cls_id] = best_t
# Apply tuned thresholds: for each sample, pick the class with highest
# (probability - threshold) as the prediction
adjusted_scores = all_probs - best_thresholds[np.newaxis, :]
all_preds = np.argmax(adjusted_scores, axis=1)
logger.info(f" Threshold tuning: {dict(zip(label_names, [f'{t:.2f}' for t in best_thresholds]))}")
else:
all_preds = np.argmax(all_probs, axis=1)
accuracy = accuracy_score(all_labels, all_preds)
micro_p, micro_r, micro_f1, _ = precision_recall_fscore_support(all_labels, all_preds, average="micro")
macro_p, macro_r, macro_f1, _ = precision_recall_fscore_support(all_labels, all_preds, average="macro")
per_class_p, per_class_r, per_class_f1, per_class_support = precision_recall_fscore_support(
all_labels,
all_preds,
average=None,
labels=list(range(num_classes)),
zero_division=0,
)
per_class = {}
for i, name in enumerate(label_names):
per_class[name] = {
"precision": float(per_class_p[i]),
"recall": float(per_class_r[i]),
"f1": float(per_class_f1[i]),
"support": int(per_class_support[i]),
}
fold_metrics = {
"fold": fold_idx + 1,
"accuracy": accuracy,
"micro_f1": micro_f1,
"macro_f1": macro_f1,
"micro_precision": micro_p,
"micro_recall": micro_r,
"macro_precision": macro_p,
"macro_recall": macro_r,
"per_class": per_class,
"train_samples": len(train_df),
"val_samples": len(val_df),
}
# Clean up GPU memory
del model, best_state
if device.type == "mps":
torch.mps.empty_cache()
elif device.type == "cuda":
torch.cuda.empty_cache()
return fold_metrics
def aggregate_cv_results(fold_results: list[dict]) -> dict:
"""Aggregate per-fold metrics into mean ± std."""
scalar_keys = [
"accuracy",
"micro_f1",
"macro_f1",
"micro_precision",
"micro_recall",
"macro_precision",
"macro_recall",
]
aggregated = {}
for key in scalar_keys:
values = [f[key] for f in fold_results]
aggregated[key] = {
"mean": float(np.mean(values)),
"std": float(np.std(values)),
"values": values,
}
# Per-class aggregation
all_class_names = list(fold_results[0]["per_class"].keys())
per_class_agg = {}
for cls in all_class_names:
cls_metrics = {}
for metric in ["precision", "recall", "f1", "support"]:
values = [f["per_class"][cls][metric] for f in fold_results]
cls_metrics[metric] = {
"mean": float(np.mean(values)),
"std": float(np.std(values)),
}
per_class_agg[cls] = cls_metrics
aggregated["per_class"] = per_class_agg
return aggregated
def print_cv_report(aggregated: dict, fold_results: list[dict]):
"""Print a formatted CV report."""
print("\n" + "=" * 70)
print("K-FOLD CROSS-VALIDATION RESULTS")
print("=" * 70)
print(f"\n{'Metric':<25} {'Mean':>10} {'± Std':>10} Per-fold values")
print("-" * 70)
for key in ["accuracy", "micro_f1", "macro_f1", "micro_precision", "micro_recall"]:
m = aggregated[key]
vals = " ".join(f"{v:.3f}" for v in m["values"])
print(f"{key:<25} {m['mean']:>10.4f} {m['std']:>10.4f} [{vals}]")
print(f"\n{'Symptom':<25} {'F1 Mean':>10} {'± Std':>10} {'Prec Mean':>10} {'Rec Mean':>10} {'Avg Support':>12}")
print("-" * 80)
per_class = aggregated["per_class"]
# Sort by F1 descending
sorted_classes = sorted(per_class.keys(), key=lambda c: per_class[c]["f1"]["mean"], reverse=True)
for cls in sorted_classes:
m = per_class[cls]
print(
f"{cls:<25} {m['f1']['mean']:>10.4f} {m['f1']['std']:>10.4f} "
f"{m['precision']['mean']:>10.4f} {m['recall']['mean']:>10.4f} "
f"{m['support']['mean']:>12.1f}"
)
print("\nSamples per fold:")
for f in fold_results:
print(f" Fold {f['fold']}: train={f['train_samples']}, val={f['val_samples']}")
def main():
parser = argparse.ArgumentParser(description="K-Fold CV for DSM-5 symptom classifier")
parser.add_argument("--model-name", type=str, default="distilbert-base-uncased")
parser.add_argument("--k", type=int, default=5, help="Number of folds")
parser.add_argument("--epochs", type=int, default=5)
parser.add_argument("--batch-size", type=int, default=16)
parser.add_argument("--lr", type=float, default=2e-5)
parser.add_argument("--max-length", type=int, default=128)
parser.add_argument("--max-negatives", type=int, default=400)
parser.add_argument("--output-dir", type=str, default=None)
parser.add_argument("--redsm5-dir", type=str, default=None)
parser.add_argument(
"--data-dir",
type=str,
default=None,
help="Path to pre-split cleaned data dir (with train.csv, val.csv, metadata.json)",
)
parser.add_argument(
"--distill", type=str, default=None, help="Path to train_distilled.csv for knowledge distillation"
)
parser.add_argument(
"--distill-alpha", type=float, default=0.5, help="Weight for hard labels in distillation (default: 0.5)"
)
parser.add_argument(
"--distill-temperature", type=float, default=3.0, help="Temperature for softening distributions (default: 3.0)"
)
parser.add_argument(
"--loss-type",
type=str,
default="ce",
choices=["ce", "effective_num", "focal"],
help="Loss type: ce (inverse-freq), effective_num (Cui et al.), focal (Lin et al.)",
)
parser.add_argument(
"--label-smoothing", type=float, default=0.0, help="Label smoothing epsilon (default: 0.0, recommended: 0.1)"
)
parser.add_argument("--focal-gamma", type=float, default=2.0, help="Focal loss gamma (default: 2.0)")
parser.add_argument(
"--effective-num-beta", type=float, default=0.999, help="Effective number beta (default: 0.999)"
)
parser.add_argument(
"--augmented",
type=str,
default=None,
help="Path to augmented_samples.csv (added to training only, never validation)",
)
parser.add_argument(
"--hard-label-classes",
type=str,
default=None,
help="Comma-separated class IDs to use hard labels only (e.g. '4,7,9' for PSYCHOMOTOR,COGNITIVE,SPECIAL)",
)
parser.add_argument(
"--pooling",
type=str,
default="mean",
choices=["cls", "mean", "cls_mean"],
help="Pooling strategy: cls, mean (default), or cls_mean (concatenated)",
)
parser.add_argument("--llrd", action="store_true", help="Enable layer-wise learning rate decay")
parser.add_argument("--fgm", action="store_true", help="Enable FGM adversarial training")
parser.add_argument("--fgm-epsilon", type=float, default=0.5, help="FGM perturbation epsilon")
parser.add_argument("--rdrop", action="store_true", help="Enable R-Drop regularization")
parser.add_argument("--rdrop-alpha", type=float, default=0.1, help="R-Drop KL weight")
parser.add_argument("--swa", action="store_true", help="Enable Stochastic Weight Averaging")
parser.add_argument("--threshold-tuning", action="store_true", help="Enable per-class threshold tuning")
args = parser.parse_args()
# Paths
project_root = Path(__file__).parent.parent.parent.parent
redsm5_dir = Path(args.redsm5_dir) if args.redsm5_dir else project_root / "redsm5"
base_dir = Path(__file__).parent.parent
output_dir = Path(args.output_dir) if args.output_dir else base_dir / "evaluation" / "cv_results"
output_dir.mkdir(parents=True, exist_ok=True)
# Device
if torch.backends.mps.is_available():
device = torch.device("mps")
elif torch.cuda.is_available():
device = torch.device("cuda")
else:
device = torch.device("cpu")
logger.info(f"Using device: {device}")
# ── Build full dataset ──
# If --data-dir is provided with pre-split CSVs, load directly instead of re-preprocessing
data_dir_path = Path(args.data_dir) if args.data_dir else None
if data_dir_path and (data_dir_path / "train.csv").exists() and (data_dir_path / "metadata.json").exists():
logger.info(f"Loading pre-split data from {data_dir_path}...")
train_presplit = pd.read_csv(data_dir_path / "train.csv")
val_presplit = pd.read_csv(data_dir_path / "val.csv")
combined = pd.concat([train_presplit, val_presplit], ignore_index=True)
combined = combined.sample(frac=1, random_state=42).reset_index(drop=True)
logger.info(f"Total dataset: {len(combined)} samples from pre-split CSVs")
else:
logger.info("Loading and preprocessing ReDSM5 dataset...")
posts, annotations = load_data(redsm5_dir)
positives = create_positive_samples(annotations)
negatives = create_negative_samples(posts, annotations, max_negatives=args.max_negatives)
combined = pd.concat([positives, negatives], ignore_index=True)
combined = combined.sample(frac=1, random_state=42).reset_index(drop=True)
logger.info(f"Total dataset: {len(combined)} samples, {combined['post_id'].nunique()} posts")
# ── Build post-level multi-label matrix for stratification ──
logger.info("Building post-level label matrix for stratification...")
post_df, label_matrix = build_post_label_matrix(combined)
logger.info(f"Posts: {len(post_df)}, Label matrix shape: {label_matrix.shape}")
# Verify label matrix has signal
label_sums = label_matrix.sum(axis=0)
label_names_sorted = sorted(SYMPTOM_LABELS.keys(), key=lambda x: SYMPTOM_LABELS[x])
for i, name in enumerate(label_names_sorted):
logger.info(f" {name}: {int(label_sums[i])} posts")
# ── Load augmented data if provided ──
augmented_df = None
if args.augmented:
augmented_df = pd.read_csv(args.augmented)
# Ensure required columns exist
required_cols = {"clean_text", "label", "label_id", "sentence_id"}
if not required_cols.issubset(set(augmented_df.columns)):
logger.error(f"Augmented CSV missing columns: {required_cols - set(augmented_df.columns)}")
return
logger.info(f"Loaded {len(augmented_df)} augmented samples (training-only)")
# ── Load soft labels if distillation enabled ──
distill_lookup = None
if args.distill:
from distillation_utils import load_soft_labels_for_df
logger.info(f"Distillation enabled: loading soft labels from {args.distill}")
# We load the full distilled CSV once, then subset per fold
distilled_df = pd.read_csv(args.distill)
label_names_sorted = sorted(SYMPTOM_LABELS.keys(), key=lambda x: SYMPTOM_LABELS[x])
soft_columns = [f"soft_{name}" for name in label_names_sorted]
# Build lookup: sentence_id → soft label vector
distill_lookup = {}
for _, row in distilled_df.iterrows():
if row.get("soft_label_valid", True):
sid = row["sentence_id"]
distill_lookup[sid] = [float(row[c]) for c in soft_columns]
logger.info(f" Loaded {len(distill_lookup)} soft label vectors")
# ── K-Fold stratified splitting at post level ──
mskf = MultilabelStratifiedKFold(n_splits=args.k, shuffle=True, random_state=42)
fold_results = []
for fold_idx, (train_post_idx, val_post_idx) in enumerate(mskf.split(post_df["post_id"], label_matrix)):
logger.info(f"\n{'=' * 60}")
logger.info(f"FOLD {fold_idx + 1}/{args.k}")
logger.info(f"{'=' * 60}")
# Map post indices back to sentence-level data
train_post_ids = set(post_df.iloc[train_post_idx]["post_id"])
val_post_ids = set(post_df.iloc[val_post_idx]["post_id"])
# Verify no leakage
assert len(train_post_ids & val_post_ids) == 0, "Post leakage between train and val!"
train_df = combined[combined["post_id"].isin(train_post_ids)].reset_index(drop=True)
val_df = combined[combined["post_id"].isin(val_post_ids)].reset_index(drop=True)
# Add augmented data to training only (never validation)
if augmented_df is not None:
aug_with_cols = augmented_df[
["post_id", "sentence_id", "sentence_text", "clean_text", "label", "label_id"]
].copy()
train_df = pd.concat([train_df, aug_with_cols], ignore_index=True)
train_df = train_df.sample(frac=1, random_state=42 + fold_idx).reset_index(drop=True)
logger.info(f" Train: {len(train_df)} sentences ({len(train_post_ids)} posts + augmented)")
logger.info(f" Val: {len(val_df)} sentences ({len(val_post_ids)} posts, no augmented)")
# Log per-class distribution in val fold
val_dist = val_df["label"].value_counts()
min_class = val_dist.min()
min_class_name = val_dist.idxmin()
logger.info(f" Smallest val class: {min_class_name} = {min_class} samples")
# Build fold-specific soft labels if distillation enabled
fold_soft_labels = None
if distill_lookup is not None:
label_names_sorted = sorted(SYMPTOM_LABELS.keys(), key=lambda x: SYMPTOM_LABELS[x])
fold_soft_labels = []
missing = 0
for _, row in train_df.iterrows():
sid = row["sentence_id"]
if sid in distill_lookup:
fold_soft_labels.append(distill_lookup[sid])
else:
# One-hot fallback
one_hot = [0.0] * len(label_names_sorted)
one_hot[int(row["label_id"])] = 1.0
fold_soft_labels.append(one_hot)
missing += 1
if missing > 0:
logger.warning(f" {missing}/{len(train_df)} samples missing soft labels — one-hot fallback")
fold_metrics = train_one_fold(
train_df=train_df,
val_df=val_df,
model_name=args.model_name,
epochs=args.epochs,
batch_size=args.batch_size,
lr=args.lr,
max_length=args.max_length,
device=device,
fold_idx=fold_idx,
soft_labels=fold_soft_labels,
distill_alpha=args.distill_alpha,
distill_temperature=args.distill_temperature,
loss_type=args.loss_type,
label_smoothing=args.label_smoothing,
focal_gamma=args.focal_gamma,
effective_num_beta=args.effective_num_beta,
hard_label_classes=[int(x) for x in args.hard_label_classes.split(",")]
if args.hard_label_classes
else None,
pooling=args.pooling,
swa=args.swa,
threshold_tuning=args.threshold_tuning,
use_llrd=args.llrd,
use_fgm=args.fgm,
fgm_epsilon=args.fgm_epsilon,
use_rdrop=args.rdrop,
rdrop_alpha=args.rdrop_alpha,
)
fold_results.append(fold_metrics)
logger.info(
f" Fold {fold_idx + 1} DONE: "
f"micro_f1={fold_metrics['micro_f1']:.4f}, "
f"macro_f1={fold_metrics['macro_f1']:.4f}"
)
# ── Aggregate results ──
aggregated = aggregate_cv_results(fold_results)
print_cv_report(aggregated, fold_results)
# ── Save ──
cv_output = {
"config": {
"model_name": args.model_name,
"k_folds": args.k,
"epochs": args.epochs,
"batch_size": args.batch_size,
"learning_rate": args.lr,
"max_length": args.max_length,
"max_negatives": args.max_negatives,
"loss_type": args.loss_type,
"label_smoothing": args.label_smoothing,
"focal_gamma": args.focal_gamma if args.loss_type == "focal" else None,
"effective_num_beta": args.effective_num_beta if args.loss_type in ("effective_num", "focal") else None,
"distillation": args.distill is not None,
"distill_alpha": args.distill_alpha if args.distill else None,
"distill_temperature": args.distill_temperature if args.distill else None,
"total_samples": len(combined),
"total_posts": combined["post_id"].nunique(),
},
"aggregated": aggregated,
"per_fold": fold_results,
}
output_file = output_dir / f"cv_results_{args.model_name.replace('/', '_')}_{args.k}fold.json"
with open(output_file, "w") as f:
json.dump(cv_output, f, indent=2, default=str)
logger.info(f"\nCV results saved to: {output_file}")
# Summary line for quick reference
micro = aggregated["micro_f1"]
macro = aggregated["macro_f1"]
print(
f"\nFINAL: Micro-F1 = {micro['mean']:.4f} ± {micro['std']:.4f}, "
f"Macro-F1 = {macro['mean']:.4f} ± {macro['std']:.4f}"
)
if __name__ == "__main__":
main()