| | |
| | |
| | """ |
| | Quick coverage evaluation using pre-generated attention analysis outputs. |
| | |
| | This script walks through a detailed attention directory (e.g. |
| | checkpoints_asllrp/detailed_eval-init.trans_*.*/sample_XXX), loads the |
| | frame_alignment.json produced for each sample, aligns it with the |
| | corresponding reference gloss sequence, and approximates frame coverage |
| | against ASLLRP ground-truth annotations. |
| | |
| | Usage: |
| | python quick_coverage_eval.py \ |
| | --detail-dir checkpoints_asllrp/detailed_eval-init.trans_20251228_041351 \ |
| | --ref-file preprocessed-asllrp/dev.bpe.gloss |
| | """ |
| |
|
| | import argparse |
| | import json |
| | import re |
| | from pathlib import Path |
| | from collections import defaultdict |
| | from difflib import SequenceMatcher |
| |
|
| |
|
| | def clean_gloss_text(text: str) -> str: |
| | """Remove BPE annotations and trim whitespace.""" |
| | return text.replace('@@ ', '').replace('@@', '').strip() |
| |
|
| |
|
| | def load_reference_sentences(ref_path: Path): |
| | """Load reference sentences (sorted by numeric index).""" |
| | refs = [] |
| | with ref_path.open('r', encoding='utf-8') as f: |
| | for line in f: |
| | line = line.strip() |
| | if not line: |
| | continue |
| | parts = line.split() |
| | try: |
| | idx = int(parts[0]) |
| | except ValueError: |
| | continue |
| | sent = clean_gloss_text(' '.join(parts[1:])) |
| | refs.append((idx, sent)) |
| | refs.sort(key=lambda x: x[0]) |
| | ordered = [sent for _, sent in refs] |
| | return ordered |
| |
|
| |
|
| | def load_mapping(mapping_path: Path): |
| | """Map gloss sentence -> list of ASLLRP utterance IDs.""" |
| | mapping = defaultdict(list) |
| | with mapping_path.open('r', encoding='utf-8') as f: |
| | for line in f: |
| | line = line.strip() |
| | if not line or ':' not in line: |
| | continue |
| | utt_id, gloss = line.split(':', 1) |
| | mapping[clean_gloss_text(gloss.strip())].append(utt_id.strip()) |
| | return mapping |
| |
|
| |
|
| | def pop_video_id(mapping, sentence): |
| | ids = mapping.get(sentence) |
| | if not ids: |
| | return None |
| | return ids.pop(0) |
| |
|
| |
|
| | def normalized_iou(pred, gt): |
| | """Compute IoU between two normalized [start, end] intervals.""" |
| | start = max(pred[0], gt[0]) |
| | end = min(pred[1], gt[1]) |
| | inter = max(0.0, end - start) |
| | union = max(pred[1], gt[1]) - min(pred[0], gt[0]) |
| | if union <= 0: |
| | return 0.0 |
| | return inter / union |
| |
|
| |
|
| | def normalize_interval(start, end, total): |
| | if total <= 0: |
| | return 0.0, 0.0 |
| | return start / total, end / total |
| |
|
| |
|
| | def normalize_token(token): |
| | """Normalize gloss tokens to be more tolerant (case/punctuation insensitive).""" |
| | if token is None: |
| | return "" |
| | token = token.lower().replace("@@", "") |
| | token = re.sub(r'[^a-z0-9]+', '', token) |
| | return token |
| |
|
| |
|
| | def token_similarity(a, b): |
| | if not a or not b: |
| | return 1.0 if a == b else 0.0 |
| | if a == b: |
| | return 1.0 |
| | return SequenceMatcher(None, a, b).ratio() |
| |
|
| |
|
| | def compute_coverage(detail_dir: Path, |
| | ref_sentences, |
| | mapping_path: Path, |
| | gt_json_path: Path, |
| | output_path: Path, |
| | expansion_factors, |
| | overflow_penalty): |
| | mapping = load_mapping(mapping_path) |
| | with gt_json_path.open('r', encoding='utf-8') as f: |
| | gt_data = json.load(f) |
| |
|
| | samples = sorted([d for d in detail_dir.iterdir() if d.is_dir()]) |
| | assert len(samples) <= len(ref_sentences), \ |
| | "Reference sentences shorter than number of samples" |
| |
|
| | overall = { |
| | "matched_tokens": 0, |
| | "complete_coverage_hits": 0, |
| | "iou_sum": 0.0, |
| | "samples_with_matches": 0, |
| | "skipped_samples": 0, |
| | } |
| | per_sample = [] |
| |
|
| | for idx, sample_dir in enumerate(samples): |
| | frame_file = sample_dir / "frame_alignment.json" |
| | if not frame_file.exists(): |
| | overall["skipped_samples"] += 1 |
| | continue |
| |
|
| | with frame_file.open('r', encoding='utf-8') as f: |
| | frame_data = json.load(f) |
| |
|
| | sentence = ref_sentences[idx] |
| | video_id = pop_video_id(mapping, sentence) |
| | if not video_id or video_id not in gt_data: |
| | overall["skipped_samples"] += 1 |
| | continue |
| |
|
| | gt_glosses = gt_data[video_id]["glosses"] |
| | if not gt_glosses: |
| | overall["skipped_samples"] += 1 |
| | continue |
| |
|
| | gt_total = max(g['end_24fps'] for g in gt_glosses if 'end_24fps' in g) |
| | if gt_total <= 0: |
| | overall["skipped_samples"] += 1 |
| | continue |
| |
|
| | gt_entries = [{ |
| | 'gloss': gt['gloss'], |
| | 'norm': normalize_token(gt['gloss']), |
| | 'start': gt['start_24fps'], |
| | 'end': gt['end_24fps'] |
| | } for gt in gt_glosses if 'start_24fps' in gt and 'end_24fps' in gt] |
| | gt_used = [False] * len(gt_entries) |
| | last_match_idx = 0 |
| |
|
| | matches = [] |
| |
|
| | total_frames_pred = max(frame_data.get("total_video_frames", 0), 1) |
| | for pred in frame_data.get("frame_ranges", []): |
| | word = pred['word'] |
| | word_norm = normalize_token(word) |
| |
|
| | match_idx = None |
| | |
| | for idx in range(last_match_idx, len(gt_entries)): |
| | if gt_used[idx]: |
| | continue |
| | if token_similarity(word_norm, gt_entries[idx]['norm']) >= 0.7: |
| | match_idx = idx |
| | break |
| |
|
| | if match_idx is None: |
| | |
| | for idx in range(len(gt_entries)): |
| | if gt_used[idx]: |
| | continue |
| | if token_similarity(word_norm, gt_entries[idx]['norm']) >= 0.7: |
| | match_idx = idx |
| | break |
| |
|
| | if match_idx is None: |
| | continue |
| |
|
| | gt_used[match_idx] = True |
| | last_match_idx = max(last_match_idx, match_idx) |
| | gt_entry = gt_entries[match_idx] |
| |
|
| | pred_norm = normalize_interval( |
| | pred['start_frame'], pred['end_frame'], total_frames_pred) |
| | gt_norm = normalize_interval(gt_entry['start'], gt_entry['end'], gt_total) |
| |
|
| | iou = normalized_iou(pred_norm, gt_norm) |
| | complete = pred_norm[0] <= gt_norm[0] and pred_norm[1] >= gt_norm[1] |
| |
|
| | matches.append({ |
| | "word": word, |
| | "pred_norm": pred_norm, |
| | "gt_norm": gt_norm, |
| | "iou": iou, |
| | "complete": complete, |
| | "pred_frames": (pred['start_frame'], pred['end_frame'], total_frames_pred), |
| | "gt_frames": (gt_entry['start'], gt_entry['end'], gt_total) |
| | }) |
| |
|
| | if not matches: |
| | overall["skipped_samples"] += 1 |
| | continue |
| |
|
| | sample_stats = { |
| | "sample": sample_dir.name, |
| | "video_id": video_id, |
| | "matched": len(matches), |
| | "complete_coverage": sum(1 for m in matches if m["complete"]), |
| | "avg_iou": sum(m["iou"] for m in matches) / len(matches), |
| | } |
| | per_sample.append(sample_stats) |
| |
|
| | overall["matched_tokens"] += sample_stats["matched"] |
| | overall["complete_coverage_hits"] += sample_stats["complete_coverage"] |
| | overall["iou_sum"] += sample_stats["avg_iou"] |
| | overall["samples_with_matches"] += 1 |
| |
|
| | |
| | for factor in expansion_factors: |
| | factor_stats = overall.setdefault("factor_stats", {}).setdefault( |
| | factor, {"coverage_sum": 0.0, "count": 0, "perfect_hits": 0, "penalized": 0} |
| | ) |
| |
|
| | for match in matches: |
| | pred_start, pred_end, pred_total = match["pred_frames"] |
| | gt_start, gt_end, gt_total = match["gt_frames"] |
| |
|
| | if pred_total <= 0 or gt_total <= 0: |
| | continue |
| |
|
| | pred_start_abs = pred_start / pred_total * gt_total |
| | pred_end_abs = pred_end / pred_total * gt_total |
| | if pred_end_abs <= pred_start_abs: |
| | pred_end_abs = pred_start_abs + 1e-6 |
| |
|
| | center = (pred_start_abs + pred_end_abs) / 2.0 |
| | half_len = (pred_end_abs - pred_start_abs) / 2.0 * factor |
| | start_exp = max(0.0, center - half_len) |
| | end_exp = min(gt_total, center + half_len) |
| |
|
| | overlap = max(0.0, min(end_exp, gt_end) - max(start_exp, gt_start)) |
| | gt_len = max(gt_end - gt_start, 1e-6) |
| | coverage = overlap / gt_len |
| |
|
| | penalized = False |
| | if start_exp < gt_start or end_exp > gt_end: |
| | coverage = max(0.0, coverage - overflow_penalty) |
| | penalized = True |
| |
|
| | factor_stats["coverage_sum"] += coverage |
| | factor_stats["count"] += 1 |
| | if coverage >= 1.0: |
| | factor_stats["perfect_hits"] += 1 |
| | if penalized: |
| | factor_stats["penalized"] += 1 |
| |
|
| | factor_summary = {} |
| | factor_stats = overall.get("factor_stats", {}) |
| | for factor, stats in factor_stats.items(): |
| | if stats["count"] == 0: |
| | continue |
| | factor_summary[str(factor)] = { |
| | "avg_coverage": stats["coverage_sum"] / stats["count"], |
| | "perfect_rate": stats["perfect_hits"] / stats["count"], |
| | "penalized_rate": stats["penalized"] / stats["count"], |
| | "count": stats["count"], |
| | } |
| |
|
| | overall_summary = { |
| | "samples_evaluated": len(samples), |
| | "samples_with_matches": overall["samples_with_matches"], |
| | "samples_skipped": overall["skipped_samples"], |
| | "avg_complete_coverage": ( |
| | overall["complete_coverage_hits"] / overall["matched_tokens"] |
| | if overall["matched_tokens"] > 0 else 0.0 |
| | ), |
| | "avg_iou": ( |
| | overall["iou_sum"] / overall["samples_with_matches"] |
| | if overall["samples_with_matches"] > 0 else 0.0 |
| | ), |
| | "word_level": factor_summary |
| | } |
| |
|
| | output = { |
| | "detail_dir": str(detail_dir), |
| | "overall": overall_summary, |
| | "sample_stats": per_sample, |
| | } |
| | with output_path.open('w', encoding='utf-8') as f: |
| | json.dump(output, f, indent=2) |
| |
|
| | return output |
| |
|
| |
|
| | def main(): |
| | parser = argparse.ArgumentParser(description="Quick coverage evaluator") |
| | parser.add_argument("--detail-dir", type=Path, required=True, |
| | help="Path to detailed attention directory") |
| | parser.add_argument("--ref-file", type=Path, |
| | default=Path("preprocessed-asllrp/dev.bpe.gloss"), |
| | help="Reference gloss file with indices") |
| | parser.add_argument("--mapping-file", type=Path, |
| | default=Path("../ASLLRP_utterances_mapping.txt"), |
| | help="Utterance mapping file (video_id: gloss ...)") |
| | parser.add_argument("--gt-json", type=Path, |
| | default=Path("../ASLLRP_utterances_with_frames.json"), |
| | help="JSON with ground-truth frame annotations") |
| | parser.add_argument("--output", type=Path, |
| | default=Path("coverage_summary.json"), |
| | help="Output summary JSON path") |
| | parser.add_argument("--expansion-factors", type=str, default="1.0,1.5,2.0", |
| | help="Comma-separated list of expansion multipliers to test") |
| | parser.add_argument("--overflow-penalty", type=float, default=0.5, |
| | help="Penalty to subtract if expanded window exceeds GT range") |
| | args = parser.parse_args() |
| |
|
| | expansion_factors = [float(x) for x in args.expansion_factors.split(',') if x.strip()] |
| |
|
| | ref_sentences = load_reference_sentences(args.ref_file) |
| | summary = compute_coverage( |
| | detail_dir=args.detail_dir, |
| | ref_sentences=ref_sentences, |
| | mapping_path=args.mapping_file, |
| | gt_json_path=args.gt_json, |
| | output_path=args.output, |
| | expansion_factors=expansion_factors, |
| | overflow_penalty=args.overflow_penalty |
| | ) |
| |
|
| | print(json.dumps(summary["overall"], indent=2)) |
| | print(f"\nPer-sample stats saved to {args.output}") |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|