diff --git a/analysis/current/analysis-report-hybrid.json b/analysis/current/analysis-report-hybrid.json index 054e57a2434f1e6dbe683593b407e2436026b93d..245cc71d535fa164c0a085a1ef36871a30b53320 100644 --- a/analysis/current/analysis-report-hybrid.json +++ b/analysis/current/analysis-report-hybrid.json @@ -1,8 +1,8 @@ { "schema_version": "1.0", "repo": "huggingface/transformers", - "snapshot_id": "20260505T000047Z", - "generated_at": "2026-05-05T00:04:56Z", + "snapshot_id": "20260505T060044Z", + "generated_at": "2026-05-05T06:09:27Z", "evidence_quality": "full", "llm_enrichment": true, "meta_bugs": [ @@ -1346,6 +1346,156 @@ } ] }, + { + "cluster_id": "cluster-45561-4", + "summary": "Cluster of 1 issues and 3 PRs centered on issue #45561.", + "status": "open", + "confidence": 0.8, + "canonical_issue_number": 45561, + "canonical_pr_number": 45645, + "issue_numbers": [ + 45561 + ], + "pr_numbers": [ + 45639, + 45645, + 45778 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target", + "soft_similarity" + ], + "pr_comparisons": [ + { + "left_pr_number": 45639, + "right_pr_number": 45645, + "code_similarity": 0.532, + "size_similarity": 0.619, + "file_overlap": 0.5, + "area_overlap": 0.543, + "patch_similarity": 0.451, + "shared_filenames": [ + "src/transformers/testing_utils.py", + "tests/utils/test_testing_utils.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/testing_utils.py", + "left_ranges": [ + [ + 3529, + 3536 + ], + [ + 3753, + 3787 + ] + ], + "right_ranges": [ + [ + 3525, + 3558 + ], + [ + 3782, + 3788 + ] + ] + }, + { + "filename": "tests/utils/test_testing_utils.py", + "left_ranges": [ + [ + 1, + 86 + ] + ], + "right_ranges": [ + [ + 1, + 114 + ] + ] + } + ] + }, + { + "left_pr_number": 45639, + "right_pr_number": 45778, + "code_similarity": 0.341, + "size_similarity": 0.141, + "file_overlap": 0.5, + "area_overlap": 0.386, + "patch_similarity": 0.181, + "shared_filenames": [ + "src/transformers/testing_utils.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/testing_utils.py", + "left_ranges": [ + [ + 3529, + 3536 + ], + [ + 3753, + 3787 + ] + ], + "right_ranges": [ + [ + 3529, + 3537 + ], + [ + 3762, + 3770 + ] + ] + } + ] + }, + { + "left_pr_number": 45645, + "right_pr_number": 45778, + "code_similarity": 0.175, + "size_similarity": 0.087, + "file_overlap": 0.25, + "area_overlap": 0.18, + "patch_similarity": 0.128, + "shared_filenames": [ + "src/transformers/testing_utils.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/testing_utils.py", + "left_ranges": [ + [ + 3525, + 3558 + ], + [ + 3782, + 3788 + ] + ], + "right_ranges": [ + [ + 3529, + 3537 + ], + [ + 3762, + 3770 + ] + ] + } + ] + } + ] + }, { "cluster_id": "cluster-43240-3", "summary": "Cluster of 1 issues and 2 PRs centered on issue #43240.", @@ -1679,81 +1829,6 @@ } ] }, - { - "cluster_id": "cluster-45561-3", - "summary": "Cluster of 1 issues and 2 PRs centered on issue #45561.", - "status": "open", - "confidence": 0.8, - "canonical_issue_number": 45561, - "canonical_pr_number": 45645, - "issue_numbers": [ - 45561 - ], - "pr_numbers": [ - 45639, - 45645 - ], - "evidence_types": [ - "closing_reference", - "shared_issue_target", - "soft_similarity" - ], - "pr_comparisons": [ - { - "left_pr_number": 45639, - "right_pr_number": 45645, - "code_similarity": 0.532, - "size_similarity": 0.619, - "file_overlap": 0.5, - "area_overlap": 0.543, - "patch_similarity": 0.451, - "shared_filenames": [ - "src/transformers/testing_utils.py", - "tests/utils/test_testing_utils.py" - ], - "shared_file_areas": [ - { - "filename": "src/transformers/testing_utils.py", - "left_ranges": [ - [ - 3529, - 3536 - ], - [ - 3753, - 3787 - ] - ], - "right_ranges": [ - [ - 3525, - 3558 - ], - [ - 3782, - 3788 - ] - ] - }, - { - "filename": "tests/utils/test_testing_utils.py", - "left_ranges": [ - [ - 1, - 86 - ] - ], - "right_ranges": [ - [ - 1, - 114 - ] - ] - } - ] - } - ] - }, { "cluster_id": "cluster-43698-3", "summary": "Cluster of 1 issues and 2 PRs centered on issue #43698.", @@ -2172,13 +2247,14 @@ "reason": "PRs in cluster-45081-3 are treated as duplicates because they converge on issue #45081 with closing_reference, shared_issue_target, soft_similarity evidence." }, { - "cluster_id": "cluster-45561-3", + "cluster_id": "cluster-45561-4", "canonical_pr_number": 45645, "duplicate_pr_numbers": [ - 45639 + 45639, + 45778 ], "target_issue_number": 45561, - "reason": "PRs in cluster-45561-3 are treated as duplicates because they converge on issue #45561 with closing_reference, shared_issue_target, soft_similarity evidence." + "reason": "PRs in cluster-45561-4 are treated as duplicates because they converge on issue #45561 with closing_reference, shared_issue_target, soft_similarity evidence." }, { "cluster_id": "cluster-45647-3", @@ -2191,10 +2267,10 @@ } ], "best_issue": { - "cluster_id": "cluster-45647-3", - "issue_number": 45647, - "reason": "Issue #45647 is the strongest global issue candidate because it is open, belongs to a cluster with 3 artifacts, and carries 12 discussion comments plus 8 inbound references.", - "score": 162.0 + "cluster_id": "cluster-45561-4", + "issue_number": 45561, + "reason": "Issue #45561 is the strongest global issue candidate because it is open, belongs to a cluster with 4 artifacts, and carries 2 discussion comments plus 8 inbound references.", + "score": 152.0 }, "best_pr": { "cluster_id": "cluster-43979-11", diff --git a/analysis/current/analysis-report-hybrid.llm-reviews.json b/analysis/current/analysis-report-hybrid.llm-reviews.json index 9a0ef737745819e48fc493aca0d022755725d70d..0dd584408f8186d1cb732c474be4fe725b9ffaeb 100644 --- a/analysis/current/analysis-report-hybrid.llm-reviews.json +++ b/analysis/current/analysis-report-hybrid.llm-reviews.json @@ -1,8 +1,8 @@ { "schema_version": "1.0", "repo": "huggingface/transformers", - "snapshot_id": "20260505T000047Z", - "generated_at": "2026-05-05T00:04:56Z", + "snapshot_id": "20260505T060044Z", + "generated_at": "2026-05-05T06:09:27Z", "model": "gpt-5.4-mini?service_tier=flex", "reviews": [ { @@ -13,8 +13,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-13244-3", "nodes": [ @@ -26,12 +26,12 @@ "issue:13244|issue:16618", "issue:15980|issue:16618" ], - "prepared_review_unit_hash": "30feca13ee61f1371a80548f6ff5c721f885c74e66adddd7b85f7a1178a6bb01", + "prepared_review_unit_hash": "d56c5d5a3f11345f515d7f21dbf2c83ed36190b41540e76cc58fa67ffd9340ba", "component_budget": { "node_count": 3, "item_count": 3, "soft_pair_count": 2, - "serialized_chars": 2477, + "serialized_chars": 2478, "estimated_input_tokens": 620, "estimated_eval_tokens": 1496 }, @@ -39,7 +39,7 @@ "node_count": 3, "item_count": 3, "soft_pair_count": 2, - "serialized_chars": 2477, + "serialized_chars": 2478, "estimated_input_tokens": 620, "estimated_eval_tokens": 1496 }, @@ -48,24 +48,24 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Only one issue is plausibly related to model-loading/auth errors; the Tapas tokenization report is unrelated. The two model-loading issues still look like different failure modes, so I would not merge them as duplicates.", - "confidence": 0.62, - "canonical_issue_reason": "issue:16618 is the broadest user-facing model-loading failure report; issue:13244 is unrelated, and issue:15980 is a narrower auth/error-message case.", + "summary": "These issues do not appear to be duplicates: one is about TAPAS tokenization parity, one about an unauthenticated private-model download error, and one about failing to load bert-base-uncased. The cluster is heterogeneous rather than a single underlying bug.", + "confidence": 0.97, + "canonical_issue_reason": null, "canonical_pr_reason": null, - "best_issue_reason": "issue:16618 is the best representative for the cluster because it is the most generic and likely to attract related model-load reports.", + "best_issue_reason": null, "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:13244", "right": "issue:16618", "accept": false, - "reason": "Tapas tokenization vs. loading 'bert-base-uncased' are different subsystems and different bugs." + "reason": "Different problems: TAPAS tokenization mismatch vs. model loading failure for bert-base-uncased." }, { "left": "issue:15980", "right": "issue:16618", "accept": false, - "reason": "One is a private-model login/error-message problem; the other is a generic inability to load a public model, so they do not appear to be the same underlying bug." + "reason": "Different code paths: private-model auth/download error vs. a general model-load failure." } ] }, @@ -84,8 +84,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -206,511 +206,415 @@ "issue:44841|issue:45084", "issue:41093|issue:41762" ], - "prepared_review_unit_hash": "96df2ee5abb1460ae9963ed3ad6a623223b9c52434dd7feaef98a645afab8255", + "prepared_review_unit_hash": "afb0273794e45991c23982a6b35e3342e60e0aa4b344c67203d5ae6ad741fae4", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, "soft_pair_count": 67, - "serialized_chars": 47123, - "estimated_input_tokens": 11781, - "estimated_eval_tokens": 23818 + "serialized_chars": 47120, + "estimated_input_tokens": 11780, + "estimated_eval_tokens": 23816 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "Mostly similarity noise. The only clearly supported duplicate family is the DeepSpeed ZeRO-3 size-0/index-0 loading bug, centered on 43638 with close variants 43596 and 41762. The mask-shape-mismatch links are not accepted here and should not be treated as a duplicate family.", - "confidence": 0.86, - "canonical_issue_reason": "43638 is the best anchor because it states the generic ZeRO-3 size-0/index-0 loading failure and matches the accepted close variants most directly.", + "summary": "This set is mostly a grab bag of unrelated model/runtime/docs issues. The only clear duplicate-looking pair is the ZeRO-3/BERT size-0 IndexError reports.", + "confidence": 0.79, + "canonical_issue_reason": "issue:43596 is the clearest representative of the one plausible duplicate thread here (ZeRO-3 + BERT + size-0 IndexError), and it is the more general/earlier report.", "canonical_pr_reason": null, - "best_issue_reason": "43638 is the most representative issue for the ZeRO-3 loading bug family and is broad enough to subsume the accepted nearby reports.", + "best_issue_reason": "issue:43596 best captures the only strong duplicate cluster in this batch and is the cleanest canonical issue candidate.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:44322", "right": "issue:45464", "accept": false, - "reason": "Same model family, but different API paths and symptoms; too broad to be one bug." + "reason": "Different Qwen3.5 failures: config attribute error vs streaming chat/completions failure." }, { "left": "issue:45237", "right": "issue:45685", "accept": false, - "reason": "Different backends and failures; not the same concrete bug." + "reason": "Different hardware/backend problems on different models; not the same bug." }, { "left": "issue:43716", "right": "issue:45237", "accept": false, - "reason": "Different model/runtime failures; no shared code-path." + "reason": "Mistral-3 dtype mismatch with image preprocessing is unrelated to GPT-OSS on AMD GPUs." }, { "left": "issue:43828", "right": "issue:45237", "accept": false, - "reason": "Different model/runtime failures; not the same bug." + "reason": "Phi-tiny-MoE autocast dtype mismatch is a different code path from GPT-OSS AMD GPU failures." }, { "left": "issue:45237", "right": "issue:45507", "accept": false, - "reason": "Different model/backend problems; not mergeable as one fix." + "reason": "AMD GPU runtime issue vs GraniteMoEHybrid invalid method; no shared underlying defect." }, { "left": "issue:44292", "right": "issue:45237", "accept": false, - "reason": "Different model and hardware context; not the same bug." + "reason": "Qwen-3-8B-NVFP4 error and GPT-OSS AMD GPU error are unrelated model/runtime issues." }, { "left": "issue:30277", "right": "issue:35545", "accept": false, - "reason": "Unrelated distributed-training and export/runtime failures." + "reason": "DeepSpeed collective mismatch during training vs ModernBERT ONNX export are unrelated." }, { "left": "issue:35545", "right": "issue:42915", "accept": false, - "reason": "Different model/export issue; no shared concrete bug." + "reason": "ONNX export failure and Qwen3Moe FP8 failure do not look like the same bug." }, { "left": "issue:35545", "right": "issue:41720", "accept": false, - "reason": "Different runtime failure and model path; not a duplicate." + "reason": "ModernBERT export problem and Qwen3 auto-device-map cuda assert are different failures." }, { "left": "issue:44509", "right": "issue:44512", "accept": false, - "reason": "Both are docs cleanup, but they target different removed references." + "reason": "Both are v5 docs cleanup items, but they reference different removed docs content and are not the same change." }, { "left": "issue:35545", "right": "issue:36010", "accept": false, - "reason": "Unrelated export bug vs import error; no shared fix." + "reason": "ModernBERT export and GenerationMixin import breakage are unrelated." }, { "left": "issue:43638", "right": "issue:44805", "accept": false, - "reason": "Both are IndexError-ish, but the error contexts and root causes differ." + "reason": "Both are IndexError reports, but the titles point to different concrete failures and models." }, { "left": "issue:43596", "right": "issue:44805", "accept": false, - "reason": "Different failure modes: ZeRO-3 load error versus mask-shape mismatch." + "reason": "Same broad ZeRO-3/indexing area, but the reported error and symptom differ too much to merge." }, { "left": "issue:44805", "right": "issue:45161", "accept": false, - "reason": "Different symptoms and model paths; too broad." + "reason": "Mask shape mismatch is unrelated to GPT-OSS tensor-parallel failure." }, { "left": "issue:44805", "right": "issue:44918", "accept": false, - "reason": "Different training-time errors; not the same underlying bug." + "reason": "Different model/trainer failures; no evidence of a shared code path." }, { "left": "issue:44661", "right": "issue:44805", "accept": false, - "reason": "Tokenizer mapping bug vs tensor mask mismatch; unrelated." + "reason": "Tokenizer-mapping setup bug is unrelated to the mask/indexing error." }, { "left": "issue:34634", "right": "issue:35545", "accept": false, - "reason": "Different component and failure type; not duplicates." + "reason": "BarkProcessor voice_preset issue is unrelated to ModernBERT ONNX export." }, { "left": "issue:30277", "right": "issue:43638", "accept": false, - "reason": "Both mention ZeRO-3, but the concrete errors and code paths differ." + "reason": "Both involve DeepSpeed, but one is collective mismatch and the other is Bert zero-3 indexing; different bugs." }, { "left": "issue:43596", "right": "issue:43638", "accept": true, - "reason": "Same ZeRO-3 index-0/size-0 loading failure; clearly the same underlying bug." + "reason": "Near-identical ZeRO-3 + BERT + size-0 IndexError reports; same underlying bug and plausible single fix." }, { "left": "issue:43975", "right": "issue:44457", "accept": false, - "reason": "Different models and different detokenization/LoRA behaviors." + "reason": "DeepSeek detokenization bug and LoRA save/reload inconsistency are different problems." }, { "left": "issue:43039", "right": "issue:45161", "accept": false, - "reason": "Different feature area and failure mode." + "reason": "Liger Kernel cross-entropy routing and GPT-OSS tensor-parallel failure are unrelated." }, { "left": "issue:44292", "right": "issue:44457", "accept": false, - "reason": "Different model/runtime issue; no shared code-path." + "reason": "Qwen NVFP4 runtime issue and LoRA merge/save mismatch are not the same defect." }, { "left": "issue:34928", "right": "issue:44805", "accept": false, - "reason": "Both involve tensor shapes, but the specific bugs are different." + "reason": "Activation checkpointing/FSDP recompute mismatch is unrelated to the mask shape IndexError." }, { "left": "issue:24643", "right": "issue:30277", "accept": false, - "reason": "Different distributed training errors with no concrete overlap." + "reason": "DeepSpeed weight-dimension error and collective mismatch are different training failures." }, { "left": "issue:43638", "right": "issue:45161", "accept": false, - "reason": "ZeRO-3 load error versus GPT-OSS tensor-parallel issue; unrelated." - }, - { - "left": "issue:34928", - "right": "issue:36331", - "accept": false, - "reason": "Training-shape mismatch vs trainer API signature mismatch; unrelated." - }, - { - "left": "issue:43596", - "right": "issue:45161", - "accept": false, - "reason": "Different model-loading bug and tensor-parallel issue." + "reason": "ZeRO-3/BERT index error is unrelated to GPT-OSS tensor-parallel not working." }, { "left": "issue:43716", "right": "issue:45507", "accept": false, - "reason": "Different model families and different runtime failures." + "reason": "Different model families and different failure modes." }, { "left": "issue:39290", "right": "issue:45081", "accept": false, - "reason": "Different model/config failures with no concrete overlap." + "reason": "Gemma3 sliding_window_pattern and Mistral tokenizer regex patch are unrelated." }, { "left": "issue:43828", "right": "issue:45161", "accept": false, - "reason": "Different model/runtime behavior and different root cause." + "reason": "Autocast dtype mismatch is not the same as GPT-OSS TP failure." }, { "left": "issue:43425", "right": "issue:44292", "accept": false, - "reason": "Compatibility issue vs model runtime failure; not a duplicate." + "reason": "Torch version compatibility and Qwen NVFP4 runtime error are different issues." }, { "left": "issue:44918", "right": "issue:45161", "accept": false, - "reason": "Different training feature failures; no shared concrete bug." + "reason": "Qwen3.5 embedding unpacking under TRL is unrelated to GPT-OSS TP behavior." }, { "left": "issue:34928", "right": "issue:41720", "accept": false, - "reason": "Activation-checkpointing/FSDP issue versus Qwen3 device-mapping crash." + "reason": "FSDP checkpointing mismatch and Qwen3 auto-device-map assert are unrelated." }, { "left": "issue:43716", "right": "issue:45685", "accept": false, - "reason": "Different model-specific failures with no shared bug." + "reason": "Mistral-3 dtype mismatch and MPS histogram kernel error are different backend bugs." }, { "left": "issue:43541", "right": "issue:43716", "accept": false, - "reason": "Different models and different runtime failures." + "reason": "Mixtral grouped_mm tracing error is unrelated to Mistral-3 image preprocessing dtype mismatch." }, { "left": "issue:43828", "right": "issue:44292", "accept": false, - "reason": "Both are dtype/runtime issues, but not the same underlying problem." - }, - { - "left": "issue:43039", - "right": "issue:43828", - "accept": false, - "reason": "Different feature areas and failure causes." - }, - { - "left": "issue:30277", - "right": "issue:36331", - "accept": false, - "reason": "Distributed collective mismatch vs trainer API incompatibility; unrelated." + "reason": "Both are model runtime errors, but the model and failure mechanism differ." }, { "left": "issue:43039", - "right": "issue:44805", - "accept": false, - "reason": "Liger-kernel/cross-entropy issue is unrelated to mask-shape mismatch." - }, - { - "left": "issue:34928", - "right": "issue:36010", - "accept": false, - "reason": "Training checkpointing bug vs import error; no shared bug." - }, - { - "left": "issue:43901", - "right": "issue:44512", + "right": "issue:44918", "accept": false, - "reason": "Both are docs issues, but they concern different removed references." + "reason": "Liger Kernel cross-entropy routing and Qwen3.5 embedding unpacking are unrelated." }, { "left": "issue:41762", - "right": "issue:43596", - "accept": true, - "reason": "Same ZeRO-3 size-0/index-0 loading bug; different model examples only." - }, - { - "left": "issue:39290", - "right": "issue:41720", - "accept": false, - "reason": "Different model/config failures with no concrete overlap." - }, - { - "left": "issue:43039", - "right": "issue:44560", - "accept": false, - "reason": "Different feature areas and runtime errors." - }, - { - "left": "issue:43866", - "right": "issue:44863", - "accept": false, - "reason": "Both are load failures, but the root causes and models differ." - }, - { - "left": "issue:43854", - "right": "issue:43866", - "accept": false, - "reason": "Different model/test problem vs corrupted checkpoint report." - }, - { - "left": "issue:39290", - "right": "issue:43531", - "accept": false, - "reason": "Similar wording, but different model/config contexts." - }, - { - "left": "issue:41093", "right": "issue:43638", "accept": false, - "reason": "Both are IndexErrors, but they refer to different failure modes." - }, - { - "left": "issue:43425", - "right": "issue:45070", - "accept": false, - "reason": "Compatibility/version issue vs pydantic config bug; unrelated." - }, - { - "left": "issue:45362", - "right": "issue:45464", - "accept": false, - "reason": "Same model family, but different API scenarios and symptoms." + "reason": "Both involve ZeRO-3, but Gemma3 loading and Bert loading failures are not the same concrete bug." }, { - "left": "issue:43792", - "right": "issue:45070", + "left": "issue:43827", + "right": "issue:44512", "accept": false, - "reason": "Different subsystem: model load/runtime vs config serialization." + "reason": "Both are docs issues, but they concern different removed v5 references and should stay separate." }, { - "left": "issue:43404", - "right": "issue:43866", + "left": "issue:43716", + "right": "issue:44560", "accept": false, - "reason": "Different model-loading bugs and different underlying causes." + "reason": "Mistral-3 dtype mismatch and Qwen3-vl StopIteration are unrelated." }, { - "left": "issue:45464", + "left": "issue:43828", "right": "issue:45685", "accept": false, - "reason": "Unrelated runtime failures." + "reason": "Autocast dtype mismatch and MPS histogram kernel error are different backend issues." }, { - "left": "issue:43541", + "left": "issue:44560", "right": "issue:45685", "accept": false, - "reason": "Different model classes and different runtime failures." + "reason": "Video embedding StopIteration and MPS Int histogram support are unrelated." }, { - "left": "issue:43541", + "left": "issue:43039", "right": "issue:43828", "accept": false, - "reason": "Different model/runtime bugs; not the same fix." + "reason": "Different model/backend failures with no shared code-path evidence." }, { - "left": "issue:36010", - "right": "issue:39290", + "left": "issue:30277", + "right": "issue:36331", "accept": false, - "reason": "Import error vs config attribute error; unrelated." + "reason": "DeepSpeed collective mismatch and CustomTrainer signature breakage are unrelated." }, { - "left": "issue:44841", - "right": "issue:45084", + "left": "issue:43039", + "right": "issue:44805", "accept": false, - "reason": "Different processor/template bugs." + "reason": "Liger Kernel cross-entropy issue is unrelated to the mask/indexing IndexError." }, { - "left": "issue:41093", - "right": "issue:41762", + "left": "issue:34928", + "right": "issue:36010", "accept": false, - "reason": "Different error signatures and likely different root causes." - }, - { - "left": "issue:41762", - "right": "issue:43638", - "accept": true, - "reason": "Same ZeRO-3 index-0/size-0 load failure; same underlying bug." + "reason": "Activation checkpointing/FSDP mismatch and GenerationMixin import error are unrelated." }, { - "left": "issue:43827", + "left": "issue:43901", "right": "issue:44512", "accept": false, - "reason": "Both are v5 docs issues, but they remove different obsolete references." - }, - { - "left": "issue:43716", - "right": "issue:44560", - "accept": false, - "reason": "Different model-specific failures with no shared bug." + "reason": "Different documentation topics and different removed behaviors." }, { - "left": "issue:43828", - "right": "issue:45685", + "left": "issue:41762", + "right": "issue:43596", "accept": false, - "reason": "Different runtime errors and different model/backends." + "reason": "Same DeepSpeed/size-0 theme, but Gemma3 zero-3 loading and Bert zero-3 loading are not the same bug." }, { - "left": "issue:44560", - "right": "issue:45685", + "left": "issue:39290", + "right": "issue:41720", "accept": false, - "reason": "Different model/backend failures; not duplicates." + "reason": "Gemma3 sliding_window_pattern and Qwen3 auto device mapping are different defects." }, { "left": "issue:43039", "right": "issue:44560", "accept": false, - "reason": "Different feature areas and error paths." + "reason": "Liger Kernel loss routing and Qwen3-vl video StopIteration are unrelated." }, { "left": "issue:43866", "right": "issue:44863", "accept": false, - "reason": "Checkpoint corruption vs unsupported implementation; not the same underlying bug." + "reason": "Checkpoint corruption and NemotronH checkpoint loading are not the same underlying issue." }, { "left": "issue:43854", "right": "issue:43866", "accept": false, - "reason": "Different model/test problems; no shared concrete bug." + "reason": "Unit-test load failure and checkpoint corruption are different problems." }, { "left": "issue:39290", "right": "issue:43531", "accept": false, - "reason": "Different model config problems, only loosely related by wording." + "reason": "Both mention Qwen3-related sliding-window behavior, but the titles point to different concrete failures." }, { "left": "issue:41093", "right": "issue:43638", "accept": false, - "reason": "Different IndexErrors with different shapes and contexts." + "reason": "Mask/tensor shape mismatch and Bert zero-3 size-0 index error are not clearly the same bug." }, { "left": "issue:43425", "right": "issue:45070", "accept": false, - "reason": "Version incompatibility vs pydantic field regression; unrelated." + "reason": "Torch version incompatibility and pydantic PretrainedConfig field breakage are unrelated." }, { "left": "issue:45362", "right": "issue:45464", "accept": false, - "reason": "Related model family, but different endpoints and failures." + "reason": "Qwen3.5 chat crash and Qwen3.5 streaming inference failure are different symptoms with no evidence of one root cause." }, { "left": "issue:43792", "right": "issue:45070", "accept": false, - "reason": "Different subsystems and causes; not a duplicate." + "reason": "Whisper load failure and pydantic config breakage are unrelated." }, { "left": "issue:43404", "right": "issue:43866", "accept": false, - "reason": "Different model-loading problems with different root causes." + "reason": "Mistral3 untied lm_head weight and Ovis2 checkpoint corruption are different defects." }, { "left": "issue:45464", "right": "issue:45685", "accept": false, - "reason": "Unrelated runtime failures." + "reason": "Qwen3.5 streaming inference failure is unrelated to the MPS histogram kernel issue." }, { "left": "issue:43541", "right": "issue:45685", "accept": false, - "reason": "Different backend/model issues; no shared fix." + "reason": "Mixtral dynamo-tracing grouped_mm error and MPS Int histogram error are unrelated." }, { "left": "issue:43541", "right": "issue:43828", "accept": false, - "reason": "Different model/runtime bugs." + "reason": "Grouped_mm tracing failure and autocast dtype mismatch are different bugs." }, { "left": "issue:36010", "right": "issue:39290", "accept": false, - "reason": "ImportError and config attribute error are unrelated." + "reason": "GenerationMixin import breakage and Gemma3 sliding_window_pattern issue are unrelated." }, { "left": "issue:44841", "right": "issue:45084", "accept": false, - "reason": "Different processor/template bugs." + "reason": "Processor load failure and template compilation error are different code paths." }, { "left": "issue:41093", "right": "issue:41762", "accept": false, - "reason": "Different error signatures and likely different root causes." + "reason": "Both are indexing/loading errors, but the specific symptoms and models differ too much." } ] }, "evaluator_result": { - "accept": false, - "feedback": "The ZeRO-3 size-0/index-0 loading family is grounded in the packet, but the summary overstates evidence by claiming a true duplicate family for the mask-shape-mismatch issues as well. In the packet, the mask-shape pairs are all marked false, so they should not be described as standing out as a duplicate family. The soft-edge verdicts should stay conservative and only support the ZeRO-3 family." + "accept": true, + "feedback": "The assessment is grounded in the packet and stays conservative. The one accepted duplicate pair, issue:43596 and issue:43638, is well supported by the near-identical ZeRO-3/BERT size-0 IndexError titles and timestamps, and the rejection of the rest avoids overclaiming from broad thematic similarity alone. The summary is appropriately cautious and does not overstate evidence." }, "evaluator_used": true, - "retried": true, + "retried": false, "accepted_nontrivial_soft_edge": true, "error_kind": null, "error_message": null @@ -723,8 +627,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -790,8 +694,8 @@ "issue:41093|issue:43596", "issue:43054|issue:43994", "issue:43716|issue:44918", - "issue:45632|issue:45698", "issue:44918|issue:45685", + "issue:45632|issue:45698", "issue:43404|issue:45005", "issue:43645|issue:45399", "issue:40990|issue:44960", @@ -811,20 +715,20 @@ "issue:43828|issue:44928", "issue:43866|issue:44403" ], - "prepared_review_unit_hash": "fdea643cfcca02e53c4cbdd9dda220a62824c0db7de7972e5c4e6b593b7d4540", + "prepared_review_unit_hash": "b1b1d3bd3df7e53a75860fbf5af49e93061f9dce33bf3cfd6aa2bc4aa83daf90", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, "soft_pair_count": 33, - "serialized_chars": 36620, + "serialized_chars": 36618, "estimated_input_tokens": 9155, "estimated_eval_tokens": 18566 }, @@ -833,216 +737,192 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is mostly a set of unrelated Transformers/vLLM/model-loading issues. The only clear duplicate group is the v5 non-persistent buffer regression, which shows up in multiple similarly worded reports.", - "confidence": 0.9, - "canonical_issue_reason": "issue:43950 is the most specific and central report for the non-persistent buffer regression: it clearly describes silent corruption in from_pretrained() and matches the other buffer-related reports.", + "summary": "Mostly a mixed-topic cluster rather than one duplicate set. Two clear duplicate pockets stand out: SigLIP2 bad outputs, and the v5 non-persistent-buffer corruption regression. Most other soft pairs are only broadly related by subsystem or symptom and should stay separate.", + "confidence": 0.88, + "canonical_issue_reason": null, "canonical_pr_reason": null, - "best_issue_reason": "issue:43950 is the best representative issue because it states the bug precisely, is broad enough to cover the related buffer-corruption reports, and is actionable for a fix.", + "best_issue_reason": "issue:43950 is the strongest standalone representative: it describes the concrete from_pretrained/non-persistent-buffer corruption regression clearly and is corroborated by near-duplicate wording in issue:43644 and issue:44534.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43296", "right": "issue:43366", "accept": false, - "reason": "Different problems: PaddleOCR-VL loading failure vs GGUF gpt-oss support." + "reason": "Different models and different failures: PaddleOCR-VL/vLLM load error vs. gpt-oss GGUF support." }, { "left": "issue:43572", "right": "issue:45070", "accept": false, - "reason": "Both involve config/API regressions, but they are different fields and different failure modes." + "reason": "One is a missing StableLmConfig field after v5 update; the other is a broader PretrainedConfig/pydantic-field regression." }, { "left": "issue:44387", "right": "issue:45005", "accept": false, - "reason": "One is an int4 memory/OOM regression; the other is tied-weights handling in translation models." + "reason": "Int4 quantization OOM and tied-weights bugs are different code paths and fixes." }, { "left": "issue:43541", "right": "issue:45161", "accept": false, - "reason": "Different MoE issues: grouped_mm tracing error vs tensor-parallel support not working." + "reason": "Mixtral torch-dynamo grouped_mm tracing failure is unrelated to GPT-OSS tensor-parallel behavior." }, { "left": "issue:43828", "right": "issue:45464", "accept": false, - "reason": "Different symptoms and code paths: autocast dtype mismatch vs streaming chat/completions failure." + "reason": "Autocast dtype mismatch in Phi-tiny-MoE is not the same as Qwen3.5 streaming chat/completions failure." }, { "left": "issue:44291", "right": "issue:45070", "accept": false, - "reason": "Unexpected init_empty_weights argument is unrelated to the pydantic PretrainedConfig field breakage." + "reason": "Unexpected _is_hf_initialized argument during init_empty_weights is a different bug than the pydantic PretrainedConfig field issue." }, { "left": "issue:43366", "right": "issue:45464", "accept": false, - "reason": "GGUF architecture support and streaming inference are unrelated." + "reason": "Both mention modern models, but GPT-OSS GGUF support and Qwen3.5 streaming inference are separate problems." }, { "left": "issue:43531", "right": "issue:45070", "accept": false, - "reason": "Qwen3-MoE sliding_window bug is not the same as the PretrainedConfig/pydantic regression." + "reason": "Qwen3-MoE sliding_window handling is not the same as a PretrainedConfig field regression." }, { "left": "issue:39290", "right": "issue:43742", "accept": false, - "reason": "Different models and distinct missing-attribute/key-error failures." + "reason": "Gemma3TextConfig missing attribute in vLLM and MobileLLM key-error loading are distinct model/load bugs." }, { "left": "issue:43541", "right": "issue:44589", "accept": false, - "reason": "Grouped_mm tracing failure is unrelated to missing Float8 storage support." + "reason": "Torch-dynamo grouped_mm tracing failure is unrelated to the missing Float8 storage type error." }, { "left": "issue:41093", "right": "issue:43596", "accept": false, - "reason": "Different indexing errors in different execution contexts." + "reason": "Mask length mismatch during indexing is a different failure from the zero-size tensor index error with ZeRO-3/BERT." }, { "left": "issue:43054", "right": "issue:43994", - "accept": false, - "reason": "Both are about SigLIP2, but one is a relative quality complaint and the other is an AutoModel/pipeline correctness issue; not clearly the same bug." + "accept": true, + "reason": "Both describe the same SigLIP2 regression: degraded/nonsensical text embeddings or outputs from the same model family, likely the same processor/model mismatch." }, { "left": "issue:43716", "right": "issue:44918", "accept": false, - "reason": "Image preprocessor/model dtype mismatch is unrelated to TRL SFT input-embedding unpacking." + "reason": "Mistral-3 image-preprocessor dtype mismatch and Qwen3.5 SFT embedding unpacking are unrelated bugs." }, { - "left": "issue:45632", - "right": "issue:45698", + "left": "issue:44918", + "right": "issue:45685", "accept": false, - "reason": "Both involve custom module loading, but one is a local-path cache collision and the other is save_pretrained/from_pretrained picking the wrong module." + "reason": "TRL SFT embedding unpacking and the MPS histogram_mps Int error are different failures in different stacks." }, { - "left": "issue:44918", - "right": "issue:45685", + "left": "issue:45632", + "right": "issue:45698", "accept": false, - "reason": "Qwen3.5 SFT embedding handling is unrelated to the MPS histogram kernel error." + "reason": "Both involve custom loading, but one is a trust_remote_code cache-path collision and the other is wrong module selection after save_pretrained; not the same concrete bug." }, { "left": "issue:43404", "right": "issue:45005", "accept": false, - "reason": "Related to tied weights in name only; these are different models and different bugs." + "reason": "Both mention tied weights, but one is a Mistral3 lm_head tying bug and the other is a translation-models v5 regression." }, { "left": "issue:43645", "right": "issue:45399", "accept": false, - "reason": "Jupyter custom-model initialization and flash-attn fallback gating are unrelated." + "reason": "Notebook custom-model initialization and flash-attn fallback gating are unrelated." }, { "left": "issue:40990", "right": "issue:44960", "accept": false, - "reason": "Different models and different evaluation/perplexity claims." + "reason": "High perplexity on gpt-oss-20b and the GLM5 issue do not share a concrete fix path." }, { "left": "issue:44530", "right": "issue:45005", "accept": false, - "reason": "PagedAttentionCache group-type error is not the same as tied-weights regressions." + "reason": "Qwen3.5 PagedAttentionCache linear_attention crash is not the same as tied-weights behavior in translation models." }, { "left": "issue:42915", "right": "issue:45399", "accept": false, - "reason": "FineGrainedFP8Config failure is unrelated to flash-attn fallback checks." + "reason": "FineGrainedFP8Config failure and flash-attn fallback checks are separate issues." }, { "left": "issue:44360", "right": "issue:44512", "accept": false, - "reason": "DSA indexer behavior and documentation typo are unrelated." + "reason": "DSA indexer ReLU discussion and a docs typo about the removed run command are unrelated." }, { "left": "issue:44863", "right": "issue:45005", "accept": false, - "reason": "NemotronH checkpoint loading is unrelated to tied-weights handling." + "reason": "NemotronH checkpoint-loading failure is different from the tied-weights translation-model regression." }, { "left": "issue:43873", "right": "issue:45005", "accept": false, - "reason": "Quantization offloading behavior is unrelated to tied-weight regressions." + "reason": "Quantization/offloading behavior and tied-weights handling are not the same bug." }, { "left": "issue:39186", "right": "issue:44918", "accept": false, - "reason": "FSDP 2-D weight error and Qwen3.5 embedding unpacking are different failure modes." + "reason": "FSDP weight dimensionality error and Qwen3.5 embedding unpacking are unrelated." }, { "left": "issue:44492", "right": "issue:44512", "accept": false, - "reason": "A cache-strategy typo is not the same as stale docs about the removed run command." + "reason": "A cache-strategy typo is not the same as outdated docs about the removed transformers run command." }, { "left": "issue:44918", "right": "issue:45507", "accept": false, - "reason": "Different model families and unrelated invalid-method/runtime errors." + "reason": "TRL SFT embedding unpacking and GraniteMoEHybrid invalid method calls are different code paths." }, { "left": "issue:43975", "right": "issue:44292", "accept": false, - "reason": "Incorrect detokenization and NVFP4 runtime failure are unrelated." + "reason": "Incorrect detokenization for deepseek-coder is unrelated to the Qwen-3-8B-NVFP4 runtime error." }, { "left": "issue:43950", "right": "issue:44534", "accept": true, - "reason": "These describe the same v5 non-persistent buffer corruption regression with nearly identical wording and behavior." + "reason": "Same v5 regression: non-persistent buffers are being overwritten/corrupted on load, just phrased differently." }, { "left": "issue:43644", "right": "issue:43950", "accept": true, - "reason": "Both report the same non-persistent buffer corruption regression in Transformers 5.x." - }, - { - "left": "issue:44292", - "right": "issue:45507", - "accept": false, - "reason": "NVFP4 runtime failure and GraniteMoEHybrid invalid-method calls are unrelated." - }, - { - "left": "issue:39692", - "right": "issue:43054", - "accept": false, - "reason": "Documentation example errors are not the same as SigLIP2 embedding-quality regression." - }, - { - "left": "issue:43828", - "right": "issue:44928", - "accept": false, - "reason": "Different Qwen/Phi training-time issues with different root causes." - }, - { - "left": "issue:43866", - "right": "issue:44403", - "accept": false, - "reason": "Checkpoint corruption and noisy load logging are not the same bug." + "reason": "Same underlying non-persistent-buffer corruption bug in Transformers 5.x." } ] }, "evaluator_result": { "accept": true, - "feedback": "The cluster summary is grounded in the packet, and the duplicate call on the v5 non-persistent buffer regression (43950/43644/44534) is supported. The other soft-edge rejections read conservative overall, with no clear overstatement of evidence." + "feedback": "The summary is grounded in the packet and stays conservative overall. The accepted soft edges are well supported by the issue titles, especially the non-persistent-buffer regression trio and the SigLIP2 output pair. The rejected pairs are mostly appropriately kept separate based on distinct models, failure modes, or subsystems, with no clear overstatement of evidence." }, "evaluator_used": true, "retried": false, @@ -1058,8 +938,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -1149,250 +1029,256 @@ "issue:43854|issue:44863", "issue:42915|issue:43872" ], - "prepared_review_unit_hash": "24078b38d4b9832995866c3b200f02029243d3137da6d096b4a4309effab2f29", + "prepared_review_unit_hash": "5e37c8dfa8b5574eace33cd366420468aeab63c6c0c49ff005fcd56b808a5210", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, "soft_pair_count": 36, - "serialized_chars": 37861, - "estimated_input_tokens": 9466, - "estimated_eval_tokens": 19188 + "serialized_chars": 37866, + "estimated_input_tokens": 9467, + "estimated_eval_tokens": 19190 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is heterogeneous and mostly splits into a few separate duplicate subgroups rather than one true cluster-wide duplicate set. Clear duplicate pairs appear for the docs pipeline-removal reports, the Qwen3 ZeRO-3 loading failures, and the _is_hf_initialized/init_empty_weights TypeError; most other suggested edges are not duplicates.", - "confidence": 0.31, - "canonical_issue_reason": "No single issue is a clean canonical for the whole cluster; it splits into multiple unrelated subclusters (docs, SigLIP2 quality/regression, ZeRO-3 loading, and init_empty_weights/bitsandbytes TypeErrors).", + "summary": "This is a mixed cluster of mostly unrelated model-loading, tokenizer, training, and docs issues. Only two soft edges look like true duplicates: the v5 docs pipeline cleanup pair and the `_is_hf_initialized` loading regression pair.", + "confidence": 0.77, + "canonical_issue_reason": "issue:43872 is the clearest exact-duplicate anchor here: issue:44291 describes the same `_is_hf_initialized` kwarg failure in the init_empty_weights / quantized-loading path.", "canonical_pr_reason": null, - "best_issue_reason": "If one issue must anchor the cluster, 45020 is the broadest umbrella candidate among the loading-regression reports, but it still does not cover the whole set.", + "best_issue_reason": "issue:45020 is the broadest cluster-level umbrella for recent remote_code/model-loading regressions, even though most items are not exact duplicates of it.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:36331", "right": "issue:41762", "accept": false, - "reason": "Unrelated: trainer compute_loss signature mismatch vs Gemma3/DeepSpeed ZeRO-3 loading failure." + "reason": "Different failure modes: custom trainer signature mismatch vs Gemma3 ZeRO-3 index error." }, { "left": "issue:41553", "right": "issue:43883", "accept": false, - "reason": "Different problems: Voxtral AutoTokenizer error messaging vs Molmo missing all_tied_weights_keys." + "reason": "Unrelated models and code paths: Voxtral AutoTokenizer error message vs Molmo tied-weights attribute error." }, { "left": "issue:43872", "right": "issue:44589", "accept": false, - "reason": "Both are loading-related, but the failure modes differ: bitsandbytes _is_hf_initialized kwarg vs missing Float8 storage object." + "reason": "Both are loading-related TypeErrors, but they point to different incompatibilities (_is_hf_initialized vs Float8 storage lookup)." }, { "left": "issue:38175", "right": "issue:43054", - "accept": true, - "reason": "Both report SigLIP2 behaving incorrectly at inference quality/output-distribution level; likely the same underlying SigLIP2 implementation issue." + "accept": false, + "reason": "Both mention SigLIP2, but one is zero probabilities and the other is degraded text embeddings; not clearly the same bug." }, { "left": "issue:41553", "right": "issue:43881", "accept": false, - "reason": "Unrelated: Voxtral tokenizer error message vs GLM-4v-9b loading failure." + "reason": "Different models and unrelated loading failures." }, { "left": "issue:43054", "right": "issue:45701", "accept": false, - "reason": "SigLIP2 embedding quality issue vs a broader tokenization-version regression; not the same code-path." + "reason": "SigLIP2 embedding quality issue vs generic tokenization change regression; not the same underlying bug." }, { "left": "issue:42947", "right": "issue:44387", "accept": false, - "reason": "Different regressions: LoRA gradient checkpointing ineffective vs int4 quantization memory growth/OOM." + "reason": "Both involve memory, but one is LoRA gradient checkpointing and the other is int4 reserved-memory growth; different regressions." }, { "left": "issue:44403", "right": "issue:44451", "accept": false, - "reason": "Unnecessary loading noise vs inability to load ScandiBERT; symptoms and targets do not match." + "reason": "Generic loading noise vs a concrete ScandiBERT loading failure; not a duplicate." }, { "left": "issue:43643", "right": "issue:45020", "accept": false, - "reason": "Related theme, but not the same bug: missing fields from AutoConfig with trust_remote_code vs a broader remote_code breakage report." + "reason": "401remote_code issues are broad, but this is specifically missing fields from AutoConfig; not enough evidence of the same bug." }, { "left": "issue:43881", "right": "issue:43883", "accept": false, - "reason": "Different model load failures: GLM-4v-9b load problem vs Molmo missing all_tied_weights_keys." + "reason": "Different model families and different failures." }, { "left": "issue:43975", "right": "issue:45520", "accept": false, - "reason": "Detokenization regression vs import_utils flash_attn KeyError; unrelated code paths." + "reason": "Tokenizer detokenization regression vs flash_attn import key error on Python 3.13." }, { "left": "issue:43054", "right": "issue:43493", "accept": false, - "reason": "Same model family, but this pair is not explicit enough to treat as a duplicate without stronger evidence of the same root cause." + "reason": "Related SigLIP2 area, but this is a performance/embedding-quality report and the other is a broader implementation discrepancy claim." }, { "left": "issue:43975", "right": "issue:45237", "accept": false, - "reason": "Tokenizer/detokenization regression vs AMD GPU runtime incompatibility; unrelated." + "reason": "Tokenizer regression vs AMD GPU runtime support issue." }, { "left": "issue:44403", "right": "issue:44488", "accept": false, - "reason": "Different load issues for different models; no clear shared underlying bug." + "reason": "Noise/warnings on load vs a specific model load failure; different problems." }, { "left": "issue:42947", "right": "issue:43856", "accept": false, - "reason": "LoRA gradient-checkpointing issue vs Qwen3 MoE memory usage; different concrete failures." + "reason": "Both are training/memory reports, but they describe different symptoms and likely different code paths." }, { "left": "issue:42491", "right": "issue:44368", "accept": false, - "reason": "LoRA/Qwen3 MoE conversion issue vs a warning about tie_word_embeddings; not the same problem." + "reason": "Qwen3 MoE version-compatibility/load issue vs a tie_word_embeddings warning in Qwen3.5 LoRA fine-tuning." }, { "left": "issue:41762", "right": "issue:42915", "accept": false, - "reason": "Gemma3 ZeRO-3 load failure vs Qwen3 MoE FP8 failure; same broad area, different bugs." + "reason": "Gemma3 ZeRO-3 index error vs Qwen3 MoE FP8 failure; unrelated." + }, + { + "left": "issue:43975", + "right": "issue:44451", + "accept": false, + "reason": "Different model/tokenizer regressions with no clear shared root cause." }, { "left": "issue:44661", "right": "issue:45020", "accept": false, - "reason": "add-new-model-like/tokenizer mapping bug vs remote_code regressions; different code paths." + "reason": "add-new-model-like mapping bug is unrelated to the broader remote_code loading regressions." }, { "left": "issue:43824", "right": "issue:43872", "accept": false, - "reason": "ImportError for Qwen2_5_VL symbol vs bitsandbytes init kwarg TypeError; unrelated." + "reason": "Qwen2.5-VL import error vs bitsandbytes _is_hf_initialized TypeError; unrelated." }, { "left": "issue:44991", "right": "issue:45399", "accept": false, - "reason": "Tokenizer loading failure for a specific model vs flash-attn fallback checks; unrelated." + "reason": "Tokenizer loading regression vs flash-attn fallback gating; different bugs." }, { "left": "issue:45020", "right": "issue:45356", "accept": false, - "reason": "Broad remote_code breakage vs Kimi-K2.5 tokenizer codec regression; not the same bug." + "reason": "Broad remote_code breakage vs a specific Kimi-K2.5 tokenizer regression; not the same concrete defect." }, { "left": "issue:43782", "right": "issue:44960", "accept": false, - "reason": "Qwen3VL weight_only loading error vs GLM5; unrelated models and failures." + "reason": "Different model families and distinct loading errors." }, { "left": "issue:43827", "right": "issue:44509", "accept": true, - "reason": "Near-verbatim docs issue: both say v5 docs still reference removed pipeline tasks like text2text-generation/summarization/translation." + "reason": "Both report the same stale docs problem: pipeline task names removed in v5 are still referenced." }, { "left": "issue:45081", "right": "issue:45399", "accept": false, - "reason": "Tokenizer regex crash vs flash-attn fallback blocked by checks; different issues." + "reason": "Tokenizer regex patch crash vs flash-attn fallback gating; unrelated." }, { "left": "issue:43792", "right": "issue:44220", "accept": false, - "reason": "Whisper runtime failure vs _torch_extract_fbank_features bug; related speech area but not clearly the same concrete defect." + "reason": "Whisper runtime failure and a feature-extraction-function issue may be related, but the evidence is not strong enough to call them the same bug." }, { "left": "issue:43404", "right": "issue:43950", "accept": false, - "reason": "Mistral3 lm_head tying bug vs non-persistent buffer corruption in from_pretrained; unrelated." + "reason": "Weight tying bug in Mistral3 vs silent non-persistent buffer corruption; different regressions." }, { "left": "issue:43742", "right": "issue:45399", "accept": false, - "reason": "MobileLLM load KeyError vs flash-attn fallback gating; not the same underlying bug." + "reason": "MobileLLM loading key error vs flash-attn fallback behavior; unrelated." }, { "left": "issue:42915", "right": "issue:45020", "accept": false, - "reason": "Qwen3 MoE FP8 failure vs broad remote_code regressions; too different to merge." + "reason": "Qwen3 MoE FP8 failure is much narrower than the broad remote_code regression bucket." }, { "left": "issue:43257", "right": "issue:43866", "accept": false, - "reason": "Qwen3 MoE weight conversion under accelerate/deepspeed vs Ovis2 checkpoint corruption; unrelated." + "reason": "Qwen3 MoE weight-conversion/load issue vs Ovis2 checkpoint corruption; unrelated." }, { "left": "issue:43742", "right": "issue:45020", "accept": false, - "reason": "Specific MobileLLM KeyError vs general remote_code breakage; not the same bug." + "reason": "Specific MobileLLM key error does not match the broad remote_code regression." }, { "left": "issue:41553", "right": "issue:44991", "accept": false, - "reason": "Different model/tokenizer loading failures with no clear shared code-path." + "reason": "Voxtral tokenizer error message vs tokenizer-loading failure for a different model; unrelated." }, { "left": "issue:43940", "right": "issue:45313", - "accept": true, - "reason": "Same concrete class of bug: ZeRO-3 weight loading fails for Qwen3-family models, just on different variants." + "accept": false, + "reason": "Both are ZeRO-3 weight-loading issues, but they affect different Qwen variants and describe different missing-weight patterns." }, { "left": "issue:43872", "right": "issue:44291", "accept": true, - "reason": "Same `_is_hf_initialized`/init_empty_weights incompatibility causing TypeError during model loading." + "reason": "Same underlying `_is_hf_initialized` TypeError during init_empty_weights / quantized model loading." }, { "left": "issue:43854", "right": "issue:44863", "accept": false, - "reason": "GLM-4.7-Flash unit-test load issue vs NemotronH checkpoint loading failure; unrelated." + "reason": "GLM-4.7-Flash test loading problem vs NemotronH checkpoint-loading failure; unrelated." }, { "left": "issue:42915", "right": "issue:43872", "accept": false, - "reason": "Qwen3 MoE FP8 training/loading issue vs bitsandbytes `_is_hf_initialized` TypeError; different failures." + "reason": "Different failures: FP8 Qwen3 MoE loading vs `_is_hf_initialized` quantized-loading TypeError." } ] }, "evaluator_result": { "accept": true, - "feedback": "The analyst summary is grounded in the packet and stays conservative. It correctly treats the cluster as heterogeneous, avoids over-claiming a single canonical issue, and only accepts soft edges that look clearly mergeable by symptom/root cause (docs pipeline removals, SigLIP2 inference-quality regressions, Qwen3 ZeRO-3 loading failures, and the _is_hf_initialized/init_empty_weights TypeError). The rejected edges are generally supported by distinct titles and failure modes." + "feedback": "The summary is grounded in the packet, and the two accepted soft edges are well-supported by the issue titles. The remaining soft-edge rejections stay conservative and do not overclaim duplicate status." }, "evaluator_used": true, "retried": false, @@ -1408,8 +1294,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -1464,8 +1350,8 @@ "soft_pairs": [ "issue:44928|issue:45588", "issue:41553|issue:44488", - "issue:36331|issue:41093", "issue:43257|issue:45305", + "issue:36331|issue:41093", "issue:43872|issue:45538", "issue:44530|issue:44863", "issue:41093|issue:44805", @@ -1475,8 +1361,8 @@ "issue:43366|issue:43531", "issue:43550|issue:44910", "issue:43645|issue:43646", - "issue:43950|issue:45325", "issue:43366|issue:44589", + "issue:43950|issue:45325", "issue:44560|issue:44918", "issue:42617|issue:43366", "issue:42915|issue:43957", @@ -1493,220 +1379,220 @@ "issue:38175|issue:39692", "issue:45313|issue:45357" ], - "prepared_review_unit_hash": "1d86b55b0d5beed4871125f8573403060984a20786a702acd59dba03a7e4f391", + "prepared_review_unit_hash": "8bf8f40d4c157d33c86d4fe87cd01e8ca5272531b6a58533cdc480a1bd89acd5", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, "soft_pair_count": 30, - "serialized_chars": 36301, - "estimated_input_tokens": 9076, - "estimated_eval_tokens": 18408 + "serialized_chars": 36297, + "estimated_input_tokens": 9075, + "estimated_eval_tokens": 18406 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "Mostly a heterogeneous set of Transformers bug reports, with only a few true duplicate-looking subclusters. The strongest matches are the custom model initialization regression, the identical mask-shape IndexError, and a SigLIP2 broken-inference report; the rest are too different in bug class or code path.", - "confidence": 0.62, - "canonical_issue_reason": "No single issue cleanly covers the whole cluster, but issue:43646 is the best canonical anchor among the true duplicates: it states the broad Transformers 5.0 custom-model-initialization regression, while issue:43645 is just a narrower Jupyter repro.", + "summary": "Mostly a heterogeneous cluster of unrelated issues, with only a few clear duplicate pockets. The strongest overlaps are the mask/tensor shape IndexError reports, the Transformers 5.0 custom model initialization regression, and the SigLIP2 output regression; the rest look distinct.", + "confidence": 0.82, + "canonical_issue_reason": "No single canonical issue for the whole cluster; it splits into multiple unrelated bug families. The closest duplicate anchor is the mask-shape IndexError pair.", "canonical_pr_reason": null, - "best_issue_reason": "issue:43646 is the most general, reusable bug report in the strongest duplicate subcluster, so it is the best issue to track even though the overall cluster is not one single bug.", + "best_issue_reason": "No global best issue for the entire cluster for the same reason; if forced to anchor one duplicate pocket, the earliest mask-shape IndexError report is the cleanest representative.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:44928", "right": "issue:45588", "accept": false, - "reason": "Different failures: Qwen3.5 RLHF/3D position_ids vs a flash-attention integration AttributeError on sink-less models." + "reason": "Different Qwen3.5 bugs: SDPA/3D position_ids NaN vs flash_attention s_aux=None crash." }, { "left": "issue:41553", "right": "issue:44488", "accept": false, - "reason": "Both are loading/error-reporting issues, but one is AutoTokenizer/Voxtral messaging and the other is loading a different model; no shared code path." + "reason": "Different models and failures; Voxtral tokenizer error message vs sleng-bert load failure." }, { - "left": "issue:36331", - "right": "issue:41093", + "left": "issue:43257", + "right": "issue:45305", "accept": false, - "reason": "Trainer API keyword mismatch versus a tensor/mask shape IndexError; unrelated bugs." + "reason": "Both DeepSpeed-related, but one is weight conversion and the other gradient averaging; not the same bug." }, { - "left": "issue:43257", - "right": "issue:45305", + "left": "issue:36331", + "right": "issue:41093", "accept": false, - "reason": "Both involve DeepSpeed/Qwen3, but one is weight conversion on load and the other is gradient averaging with GAS; different code paths." + "reason": "Different code paths: custom trainer signature mismatch vs mask/tensor shape IndexError." }, { "left": "issue:43872", "right": "issue:45538", "accept": false, - "reason": "bitsandbytes constructor incompatibility versus a tokenizer max_length constant; unrelated." + "reason": "bitsandbytes init incompatibility vs CLIPTokenizer max_length bug." }, { "left": "issue:44530", "right": "issue:44863", "accept": false, - "reason": "Both mention model internals, but one is a PagedAttentionCache crash on Qwen3.5 and the other is a NemotronH checkpoint loading failure." + "reason": "Qwen3.5 PagedAttention cache crash vs NemotronH checkpoint loading." }, { "left": "issue:41093", "right": "issue:44805", "accept": true, - "reason": "Same IndexError family with the same mask-vs-tensor shape mismatch; 44805 looks like a later/generalized repro of the same bug." + "reason": "Same generic mask/tensor shape mismatch IndexError; 44805 looks like a later duplicate report." }, { "left": "issue:44479", "right": "issue:44560", "accept": false, - "reason": "Different regressions: a broad video-input regression across several Qwen VL variants versus a StopIteration on Qwen3-vl-embedding." + "reason": "Both are video/Qwen issues, but the failures differ: broad video-input regression vs StopIteration in Qwen3-vl-embedding." }, { "left": "issue:24643", "right": "issue:30064", "accept": false, - "reason": "DeepSpeed training weight-dimension error versus image processor handling of void segmentation maps; unrelated." + "reason": "Training/DeepSpeed weight shape error vs image processor void-segmentation issue." }, { "left": "issue:43366", "right": "issue:43828", "accept": false, - "reason": "GGUF/gpt-oss support request versus a torch.autocast dtype mismatch on Phi-tiny-MoE; not the same bug." + "reason": "GGUF/gpt-oss support vs Phi-tiny-MoE autocast dtype mismatch." }, { "left": "issue:43366", "right": "issue:43531", "accept": false, - "reason": "GGUF support request versus a Qwen3-MoE sliding_window issue; different changes." + "reason": "Unrelated Qwen3-MoE sliding_window bug vs GGUF gpt-oss support." }, { "left": "issue:43550", "right": "issue:44910", "accept": false, - "reason": "Both touch attention, but Bamba SDPA/torch.compile and Qwen3.5 flash-attention position_ids are separate failures." + "reason": "Different models and kernels: Bamba torch.compile/SDPA vs Qwen3.5 flash-attention position_ids crash." }, { "left": "issue:43645", "right": "issue:43646", "accept": true, - "reason": "These describe the same Transformers 5.0 custom-model-initialization regression; the notebook mention is just a specific repro." + "reason": "Same Transformers 5.0 custom model initialization regression; one is just a notebook-specific reproduction." }, { - "left": "issue:43950", - "right": "issue:45325", + "left": "issue:43366", + "right": "issue:44589", "accept": false, - "reason": "Non-persistent buffer loading corruption versus Qwen2.5-VL rope-position scaling; different subsystems." + "reason": "GGUF architecture support vs missing Float8 storage type." }, { - "left": "issue:43366", - "right": "issue:44589", + "left": "issue:43950", + "right": "issue:45325", "accept": false, - "reason": "GGUF architecture support versus a missing Float8 storage type error; unrelated." + "reason": "Buffer loading regression vs Qwen2.5-VL rope/temporal position scaling." }, { "left": "issue:44560", "right": "issue:44918", "accept": false, - "reason": "Both are Qwen VL/input-embedding related, but one is a video StopIteration and the other is TRL SFT unpacking failure; not the same concrete bug." + "reason": "Video StopIteration in Qwen3-vl-embedding vs Qwen3.5 TRL unpacking failure." }, { "left": "issue:42617", "right": "issue:43366", "accept": false, - "reason": "3d_parallel.py runtime problem versus GGUF gpt-oss support; unrelated." + "reason": "3d_parallel.py runtime failure unrelated to GGUF support." }, { "left": "issue:42915", "right": "issue:43957", "accept": false, - "reason": "FineGrainedFP8Config failure on Qwen3Moe versus meta-device loading regressions; different code paths." + "reason": "Qwen3Moe FineGrainedFP8Config failure vs meta-device model loading." }, { "left": "issue:43824", "right": "issue:43957", "accept": false, - "reason": "Import error for a Qwen2.5-VL class versus meta-device model loading breakage; unrelated." + "reason": "Import/export of Qwen2_5_VL class vs meta-device loading regression." }, { "left": "issue:36010", "right": "issue:42915", "accept": false, - "reason": "GenerationMixin import failure versus Qwen3Moe FP8 config failure; unrelated." + "reason": "GenerationMixin import error vs Qwen3Moe FP8 config." }, { "left": "issue:38175", "right": "issue:43994", "accept": true, - "reason": "Same SigLIP2 checkpoint with broken inference outputs; zero probabilities and nonsensical pipeline/AutoModel results look like the same underlying regression." + "reason": "Same SigLIP2 model/output regression; both report broken predictions for google/siglip2-base-patch16-224." }, { "left": "issue:43646", "right": "issue:43950", "accept": false, - "reason": "Custom model initialization regression versus non-persistent buffer corruption on from_pretrained(); different issues." + "reason": "Custom model init regression vs non-persistent buffer corruption." }, { "left": "issue:44155", "right": "issue:45584", "accept": false, - "reason": "AudioFlamingo3 batched-inference token leak versus Whisper empty-transcription generation failure; unrelated." + "reason": "AudioFlamingo3 track leak vs Whisper empty transcription bug." }, { "left": "issue:44877", "right": "issue:45030", "accept": false, - "reason": "Strict config blocking granite_speech versus tiny-random glm4v config validation; different model families and failures." + "reason": "Different config-validation failures on different models." }, { "left": "issue:44560", "right": "issue:44805", "accept": false, - "reason": "Qwen3-vl video StopIteration versus a generic mask-shape IndexError; not the same bug." + "reason": "Different error classes: video StopIteration vs generic mask shape mismatch." }, { "left": "issue:43638", "right": "issue:44661", "accept": false, - "reason": "Zero-size tensor IndexError under DeepSpeed ZeRO-3 versus tokenizer-mapping-name handling in add-new-model-like; unrelated." + "reason": "Zero3/Bert index error vs tokenizer-mapping restriction in add-new-model-like." }, { "left": "issue:43596", "right": "issue:44661", "accept": false, - "reason": "Both are IndexErrors, but one is ZeRO-3/BertModel loading and the other is a tokenizer mapping edge case; different root causes." + "reason": "Zero3/Bert index error vs tokenizer-mapping restriction in add-new-model-like." }, { "left": "issue:43317", "right": "issue:44756", "accept": false, - "reason": "Dequantized model offload loading failure versus mmap OOM on Strix Halo; different memory-loading problems." + "reason": "Dequantized model offload failure vs Strix Halo mmap OOM." }, { "left": "issue:38175", "right": "issue:39692", "accept": false, - "reason": "Both mention SigLIP2, but one is bad model outputs and the other is a documentation example with model/processor mismatch plus quantization failure; not the same concrete bug." + "reason": "Same model family but different issues: doc-example validation/quantization vs output probabilities." }, { "left": "issue:45313", "right": "issue:45357", "accept": false, - "reason": "ZeRO-3 loading weights for Qwen3.5 language_model versus save_pretrained storing incorrect visual encoder keys; separate regressions." + "reason": "Load-time ZeRO-3 weight issue vs save_pretrained visual encoder key regression." } ] }, "evaluator_result": { "accept": true, - "feedback": "The summary is grounded in the packet and stays conservative about a mostly heterogeneous cluster. The accepted duplicate subclusters are plausible from the issue titles alone: 43645/43646, 41093/44805, and 38175/43994. The non-duplicate verdicts also avoid over-claiming by generally citing distinct failure modes or code paths. No PR-pair mergeability concerns apply here." + "feedback": "The summary is grounded in the packet: the cluster is mostly heterogeneous, and the three accepted duplicate pockets are tightly supported by the titles. The rejected soft edges are conservative and do not overclaim evidence." }, "evaluator_used": true, "retried": false, @@ -1722,8 +1608,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -1780,10 +1666,10 @@ "issue:44960|issue:44977", "issue:44451|issue:44488", "issue:42491|issue:43720", - "issue:43854|issue:43931", "issue:43708|issue:44075", - "issue:41720|issue:42915", + "issue:43854|issue:43931", "issue:42491|issue:45276", + "issue:41720|issue:42915", "issue:43856|issue:43873", "issue:43257|issue:43854", "issue:39692|issue:43994", @@ -1794,254 +1680,261 @@ "issue:43644|issue:43645", "issue:44910|issue:45588", "issue:45362|issue:45406", - "issue:44488|issue:44991", "issue:42491|issue:45292", + "issue:44488|issue:44991", "issue:43404|issue:43454", "issue:36010|issue:39404", "issue:45276|issue:45335", "issue:44164|issue:44291", "issue:44402|issue:44488", "issue:39692|issue:44960", - "issue:42947|issue:45704", "issue:43425|issue:43957", + "issue:42947|issue:45704", "issue:43335|issue:45663", "issue:45663|issue:45759", "issue:43873|issue:44387", "issue:44387|issue:44756", - "issue:43854|issue:45216" + "issue:43854|issue:45216", + "issue:44291|issue:44756" ], - "prepared_review_unit_hash": "458006a1b71cabd454ad0c61ab0d278d12bbaa6fc6e012e772b71deaf57f7e9a", + "prepared_review_unit_hash": "8637dd97acadfb7b439285db8292b5acf7620bdea49ad9e3fa2adf2155f78a6e", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, - "soft_pair_count": 33, - "serialized_chars": 37020, - "estimated_input_tokens": 9255, - "estimated_eval_tokens": 18766 + "soft_pair_count": 34, + "serialized_chars": 37316, + "estimated_input_tokens": 9329, + "estimated_eval_tokens": 18914 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "One clear duplicate pair stands out: the non-persistent-buffer regression in Transformers 5.0.0. The rest are related by subsystem or model family, but they do not look like the same concrete bug.", - "confidence": 0.91, - "canonical_issue_reason": "issue:43644 is the earlier, broader report of the non-persistent buffer regression in Transformers 5.0.0; it is the better canonical thread for the duplicate pair with issue:44534.", + "summary": "This set is mostly a grab bag of distinct Transformers bugs; only two soft edges look like true duplicates: the Transformers 5 non-persistent-buffer junk issue and the resize_token_embeddings propagation bug. The rest are too model- or code-path-specific to merge.", + "confidence": 0.81, + "canonical_issue_reason": "issue:43644 is the strongest representative because it has an almost exact duplicate in 44534 and the bug statement is very specific and reusable as a canonical report.", "canonical_pr_reason": null, - "best_issue_reason": "issue:43644 is the strongest issue candidate overall because it captures a concrete, version-specific regression with a clear duplicate and broad impact.", + "best_issue_reason": "issue:43644 is the best overall issue candidate: clear repro class, strong duplicate evidence, and a concrete Transformers 5 regression with broad impact.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43644", "right": "issue:44534", "accept": true, - "reason": "Same bug: both report Transformers v5 filling non-persistent buffers with junk; 44534 is just a shorter rewording of the same regression." + "reason": "Same bug: Transformers v5 filling non-persistent buffers with junk; titles describe the same regression nearly verbatim." }, { "left": "issue:44960", "right": "issue:44977", "accept": false, - "reason": "Both mention flash-attention/generation failures, but they affect different models and appear to be different code-path bugs." + "reason": "Both involve generation/flash-attention, but they target different models and failure modes; not the same concrete bug." }, { "left": "issue:44451", "right": "issue:44488", "accept": false, - "reason": "Both are model-loading failures for different BERT checkpoints, but there is no clear evidence they share the same underlying defect." + "reason": "Both are model-loading complaints, but the titles point to different models and no shared root cause is evident." }, { "left": "issue:42491", "right": "issue:43720", "accept": false, - "reason": "Completely different problems: Qwen3 MoE LoRA compatibility vs BitNet packed-weight loading in accelerate." - }, - { - "left": "issue:43854", - "right": "issue:43931", - "accept": false, - "reason": "Different model-load failures with different symptoms; no sign of a shared root cause." + "reason": "Unrelated: Qwen3 MoE LoRA compatibility vs BitNet packed-weight loading in CI." }, { "left": "issue:43708", "right": "issue:44075", "accept": false, - "reason": "Trainer checkpoint math and SGD argument handling are unrelated bugs." + "reason": "Different problems: checkpoint resume step counting vs SGD optimizer arguments not being used." }, { - "left": "issue:41720", - "right": "issue:42915", + "left": "issue:43854", + "right": "issue:43931", "accept": false, - "reason": "Different Qwen3 issues: device-mapping CUDA assert vs FineGrainedFP8Config failure." + "reason": "Different model-loading failures for different models and different symptoms." }, { "left": "issue:42491", "right": "issue:45276", "accept": false, - "reason": "Unrelated issues; one is a Qwen3 MoE LoRA compatibility problem, the other is resize_token_embeddings behavior in Gemma4." + "reason": "Different subsystems: Qwen3 MoE LoRA loading vs token embedding resize propagation." + }, + { + "left": "issue:41720", + "right": "issue:42915", + "accept": false, + "reason": "Both are Qwen3-related, but one is a device-mapping CUDA assert and the other is a FineGrainedFP8Config failure." }, { "left": "issue:43856", "right": "issue:43873", "accept": false, - "reason": "Both involve Qwen3 MoE/quantization, but one is training memory usage and the other is offloading behavior; not the same bug." + "reason": "Both mention memory/quantization themes, but the concrete bugs are different." }, { "left": "issue:43257", "right": "issue:43854", "accept": false, - "reason": "Different failures: Qwen3 MoE weight conversion with accelerate/deepspeed vs GLM-4.7 loading in tests." + "reason": "Different model families and unrelated loading paths." }, { "left": "issue:39692", "right": "issue:43994", "accept": false, - "reason": "Both concern SigLIP2, but one is a broken documentation example and the other is bad outputs in AutoModel/pipeline; not the same concrete issue." + "reason": "Both are SigLIP2-related, but one is a docs example with model/processor mismatch and the other is a nonsensical-output complaint." }, { "left": "issue:42915", "right": "issue:44756", "accept": false, - "reason": "Qwen3Moe FP8 failure and Strix Halo mmap OOM are unrelated." + "reason": "Unrelated: Qwen3 MoE FP8 failure vs Strix Halo mmap/OOM behavior." }, { "left": "issue:42491", "right": "issue:43931", "accept": false, - "reason": "Different model-loading/training problems with no shared code-path evidence." + "reason": "Different issues: Qwen3 MoE LoRA loading vs Qwen3-VL weight-shape mismatch." }, { "left": "issue:43856", "right": "issue:44387", "accept": false, - "reason": "One is Qwen3 MoE memory use during training; the other is int4 CUDA reserved-memory/OOM under Transformers 5.x." + "reason": "Both are memory-related, but one is training inefficiency and the other is int4 quantization OOM; not the same bug." }, { "left": "issue:44661", "right": "issue:44683", "accept": false, - "reason": "Tokenizer-mapping registration bug and compiled flex_attention/torch-version failure are unrelated." + "reason": "Different areas: tokenizer-mapping registration vs compiled flex_attention on new torch." }, { "left": "issue:43644", "right": "issue:43645", "accept": false, - "reason": "Related to Transformers 5.0.0 regressions, but one is non-persistent buffers and the other is custom-model initialization in Jupyter; different bugs." + "reason": "Different regressions: junk non-persistent buffers vs custom-model initialization in notebooks." }, { "left": "issue:44910", "right": "issue:45588", "accept": false, - "reason": "Both are flash-attention crashes, but the root causes differ: packed-sequence misinterpretation vs s_aux=None handling." + "reason": "Both are flash-attention crashes, but one is a packed-sequence position_ids bug and the other is an s_aux=None AttributeError." }, { "left": "issue:45362", "right": "issue:45406", "accept": false, - "reason": "Different failures in different model stacks: Qwen3.5 chat crash vs Gemma4Processor missing _tokenizer." + "reason": "Different model/application paths: chat crash in Qwen3.5 vs serve crash in Gemma4Processor." }, { - "left": "issue:44488", - "right": "issue:44991", + "left": "issue:42491", + "right": "issue:45292", "accept": false, - "reason": "Both are tokenizer/model-loading complaints, but they concern different checkpoints and different failure modes." + "reason": "Unrelated: Qwen3 MoE LoRA loading vs resize_token_embeddings not updating output embeddings." }, { - "left": "issue:42491", - "right": "issue:45292", + "left": "issue:44488", + "right": "issue:44991", "accept": false, - "reason": "No concrete overlap; one is Qwen3 MoE LoRA compatibility, the other is output_embeddings not updating on resize." + "reason": "Both are tokenizer/loading problems, but they concern different models and different failure symptoms." }, { "left": "issue:43404", "right": "issue:43454", "accept": false, - "reason": "Both are weight-tying problems, but they affect different model classes and are not clearly the same code-path bug." + "reason": "Similar wording about tied lm_head weights, but these are separate model implementations and not clearly one duplicate bug." }, { "left": "issue:36010", "right": "issue:39404", "accept": false, - "reason": "ImportError for GenerationMixin and Whisper return_language pipeline regression are unrelated." + "reason": "Unrelated: missing GenerationMixin import vs Whisper pipeline return_language regression." }, { "left": "issue:45276", "right": "issue:45335", - "accept": false, - "reason": "Very similar symptom, but different model families and different embedding targets; not enough evidence of one shared bug." + "accept": true, + "reason": "Same underlying bug: resize_token_embeddings is not propagating to the relevant tied output/decoder embeddings in related model variants." }, { "left": "issue:44164", "right": "issue:44291", "accept": false, - "reason": "Saving/from_pretrained extra_state handling and init_empty_weights unexpected argument handling are distinct loading issues." + "reason": "Different serialization/loading failures: extra_state handling vs init_empty_weights/_is_hf_initialized TypeError." }, { "left": "issue:44402", "right": "issue:44488", "accept": false, - "reason": "Tokenizer vocab-size mismatch and a model-loading failure for cjvt/sleng-bert are different defects." + "reason": "Both are model loading issues, but they involve different models and different failure descriptions." }, { "left": "issue:39692", "right": "issue:44960", "accept": false, - "reason": "Different model families and different problem types; no shared underlying bug." + "reason": "Different model families and unrelated failure modes." }, { - "left": "issue:42947", - "right": "issue:45704", + "left": "issue:43425", + "right": "issue:43957", "accept": false, - "reason": "Both touch training stability/memory, but PEFT gradient checkpointing and apex FusedRMSNorm memory leak are unrelated." + "reason": "Torch version incompatibility is unrelated to meta-device loading breakage." }, { - "left": "issue:43425", - "right": "issue:43957", + "left": "issue:42947", + "right": "issue:45704", "accept": false, - "reason": "Torch 2.10 incompatibility and meta-device loading regressions are separate problems." + "reason": "Different concerns: LoRA gradient checkpointing vs an apex RMSNorm memory leak." }, { "left": "issue:43335", "right": "issue:45663", "accept": false, - "reason": "SwitchTransformers config construction and Gemma-4 FSDP2 shared_kv_states behavior are unrelated." + "reason": "Unrelated model-specific bugs: sparse-layer config creation vs FSDP2 shared_kv_states handling." }, { "left": "issue:45663", "right": "issue:45759", "accept": false, - "reason": "Gemma-4 FSDP2 attention state bug and Qwen multimodal config unwrapping are different issues." + "reason": "Different bugs in different model/config paths; no shared concrete code-path problem." }, { "left": "issue:43873", "right": "issue:44387", "accept": false, - "reason": "Offloading with quantization and int4 CUDA reserved-memory/OOM are related at a high level but not the same concrete bug." + "reason": "Both are memory-related, but one is quantization offloading behavior and the other is int4 CUDA reserved-memory growth." }, { "left": "issue:44387", "right": "issue:44756", "accept": false, - "reason": "Different performance/memory regressions: int4 quantization OOM vs disabling mmap on Strix Halo." + "reason": "Different memory/OOM issues with no shared root cause." }, { "left": "issue:43854", "right": "issue:45216", "accept": false, - "reason": "Different Qwen3/VL loading and save_pretrained regression issues." + "reason": "Different model-loading/save regressions for different models." + }, + { + "left": "issue:44291", + "right": "issue:44756", + "accept": false, + "reason": "Different issues: init_empty_weights argument mismatch vs mmap-induced OOM." } ] }, "evaluator_result": { "accept": true, - "feedback": "The summary is grounded in the packet and stays conservative on the soft edges. The only accepted pair, issue:43644 / issue:44534, is well-supported by the titles and is a reasonable canonical duplicate thread. The rejected pairs are all framed cautiously with no obvious overstatement." + "feedback": "The summary stays grounded in the packet and the duplicate calls are conservative. The two accepted soft edges are strongly supported by the titles/body-level similarity, and the rejected pairs are appropriately cautious about model- or failure-specific differences. The only mild stretch is the phrasing about \"broad impact\" for issue:43644, but it does not materially overstate the evidence." }, "evaluator_used": true, "retried": false, @@ -2057,8 +1950,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -2084,7 +1977,6 @@ "issue:44155", "issue:44164", "issue:44265", - "issue:44291", "issue:44292", "issue:44402", "issue:44451", @@ -2113,7 +2005,6 @@ ], "soft_pairs": [ "issue:44292|issue:44912", - "issue:44291|issue:44756", "issue:42947|issue:43381", "issue:45684|issue:45698", "issue:45276|issue:45292", @@ -2121,8 +2012,8 @@ "issue:44164|issue:44756", "issue:41720|issue:44155", "issue:45588|issue:45663", - "issue:44479|issue:44877", "issue:43646|issue:43708", + "issue:44479|issue:44877", "issue:43854|issue:44960", "issue:43957|issue:44756", "issue:45292|issue:45335", @@ -2137,268 +2028,256 @@ "issue:39404|issue:45701", "issue:43611|issue:43645", "issue:44530|issue:45084", - "issue:43931|issue:44863", "issue:43425|issue:43716", + "issue:43931|issue:44863", "issue:43901|issue:44509", "issue:39404|issue:44912", "issue:44863|issue:45663", "issue:44265|issue:44898", "issue:43299|issue:43931", "issue:44479|issue:44805", - "issue:43856|issue:44756", "issue:43425|issue:44918", + "issue:43856|issue:44756", "issue:44589|issue:45084" ], - "prepared_review_unit_hash": "dcafaf54bee22a60bbca48dfba4a0e916a5ba837cf28b58330a9f6f5494bcc32", + "prepared_review_unit_hash": "1b38e4851ea3336280727f64fbf49915264e4aa012bb64d2496b1de1c7b7ce10", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { - "node_count": 48, - "item_count": 48, - "soft_pair_count": 36, - "serialized_chars": 38455, - "estimated_input_tokens": 9614, - "estimated_eval_tokens": 19484 + "node_count": 47, + "item_count": 47, + "soft_pair_count": 35, + "serialized_chars": 37526, + "estimated_input_tokens": 9382, + "estimated_eval_tokens": 19020 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is mostly a grab-bag of unrelated Transformers regressions, with one clear duplicate subcluster around `resize_token_embeddings` not propagating to downstream embedding/output modules.", - "confidence": 0.78, - "canonical_issue_reason": "issue:45292 is the broadest, cleanest description of the resize-token-embedding propagation bug and can serve as the anchor for the small duplicate subcluster.", + "summary": "This cluster is mostly heterogeneous: most items are separate regressions across loading, quantization, training, and docs. The only clearly duplicate-worthy subcluster is the resize_token_embeddings regression around output embeddings/decoder embeddings (45276, 45292, 45335).", + "confidence": 0.69, + "canonical_issue_reason": "issue:45292 is the broadest, most central description of the resize_token_embeddings regression affecting output embeddings; issue:45276 and issue:45335 look like model-specific manifestations of the same bug.", "canonical_pr_reason": null, - "best_issue_reason": "issue:45292 is the best representative issue because it states the shared failure mode without being overly model-specific.", + "best_issue_reason": "issue:45292 is the best anchor for deduping the resize_token_embeddings bug because it is generic rather than model-specific and matches the shared underlying failure mode.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:44292", "right": "issue:44912", "accept": false, - "reason": "Both are quantization/loading failures, but for different models and different failure modes; no clear same underlying bug." - }, - { - "left": "issue:44291", - "right": "issue:44756", - "accept": false, - "reason": "`init_empty_weights` argument handling and Strix Halo mmap/OOM are unrelated bugs." + "reason": "Both are quantization-related load failures, but they involve different models and symptoms (NVFP4 run error vs MXFP4 fallback to bf16), so they do not look like the same bug." }, { "left": "issue:42947", "right": "issue:43381", "accept": false, - "reason": "Both mention gradient checkpointing, but one is a PEFT/LoRA ineffectiveness bug and the other is an eval-mode restriction." + "reason": "Both mention gradient checkpointing, but one is about PEFT/LoRA effectiveness and the other is an eval-mode restriction; different code paths and failure modes." }, { "left": "issue:45684", "right": "issue:45698", "accept": false, - "reason": "Both involve save/load custom-model workflows, but one is about read-only file permissions and the other is about loading the wrong module." + "reason": "One is about read-only permissions during save_pretrained, the other about loading the wrong custom module after save_pretrained; related area, but distinct bugs." }, { "left": "issue:45276", "right": "issue:45292", "accept": true, - "reason": "Same underlying bug: `resize_token_embeddings` is not propagating updates to output-side embeddings; 45276 is the gemma4-specific manifestation." + "reason": "Both report resize_token_embeddings not propagating to output embeddings; issue:45276 is a Gemma4-specific manifestation of the broader bug in issue:45292." }, { "left": "issue:43381", "right": "issue:44928", "accept": false, - "reason": "Unrelated gradient-checkpointing behavior vs RLHF/SDPA/BF16 instability." + "reason": "These are unrelated: eval-mode gradient checkpointing restriction versus RLHF NaN/gradient explosion from 3D position_ids and SDPA fallback." }, { "left": "issue:44164", "right": "issue:44756", "accept": false, - "reason": "`extra_state` serialization and Strix Halo mmap/OOM are different problems." + "reason": "Save/from_pretrained extra_state handling and mmap OOM on Strix Halo are unrelated problems." }, { "left": "issue:41720", "right": "issue:44155", "accept": false, - "reason": "Different models and different failures: device-map assertion vs batched inference token leakage." + "reason": "Different subsystems and symptoms: device-mapping cuda assert on Qwen3 versus batched AudioFlamingo3 inference leakage." }, { "left": "issue:45588", "right": "issue:45663", "accept": false, - "reason": "Different code paths and symptoms: flash-attention aux-state handling vs Gemma-4 FSDP shared-kv-state rebuilding." + "reason": "Flash-attention s_aux=None crash and Gemma4 FSDP2 KeyError are different concrete bugs despite both touching model internals." }, { - "left": "issue:44479", - "right": "issue:44877", + "left": "issue:43646", + "right": "issue:43708", "accept": false, - "reason": "Video-input regression for Qwen models and strict-config loading for granite_speech are unrelated." + "reason": "Custom model initialization breakage and Trainer max_steps miscalculation on checkpoint resume are unrelated." }, { - "left": "issue:43646", - "right": "issue:43708", + "left": "issue:44479", + "right": "issue:44877", "accept": false, - "reason": "Custom model initialization and Trainer checkpoint-resume step calculation are unrelated." + "reason": "Video-input regression for Qwen models and strict config blocking granite_speech loading are different failures." }, { "left": "issue:43854", "right": "issue:44960", "accept": false, - "reason": "GLM-4.7-Flash test loading and GLM5 are not the same bug." + "reason": "GLM-4.7-Flash loading issues and GLM5 are related only by family name; the reports describe different model/version problems." }, { "left": "issue:43957", "right": "issue:44756", "accept": false, - "reason": "`torch.device('meta')` loading regressions and Strix Halo mmap/OOM are different issues." + "reason": "Meta-device loading failures and Strix Halo mmap OOM are unrelated." }, { "left": "issue:45292", "right": "issue:45335", "accept": true, - "reason": "Same resize-token-embedding propagation bug, with 45335 describing the decoder embedding side specifically." + "reason": "Both describe the same resize_token_embeddings propagation bug: issue:45292 mentions output_embeddings generally, while issue:45335 is the decoder.embed_tokens manifestation in t5gemma." }, { "left": "issue:44402", "right": "issue:44451", "accept": false, - "reason": "Tokenizer vocab mismatch and ScandiBERT loading failure are too vague to treat as the same bug." + "reason": "A tokenizer vocab mismatch and a generic load failure for ScandiBERT are not clearly the same underlying bug." }, { "left": "issue:43645", "right": "issue:44164", "accept": false, - "reason": "Notebook custom-model init and `extra_state` save/load handling are different failures." + "reason": "Notebook custom-model initialization and extra_state save/load handling are different issues." }, { "left": "issue:43828", "right": "issue:43957", "accept": false, - "reason": "Autocast dtype mismatch and meta-device loading failure are unrelated." + "reason": "Autocast dtype mismatch in Phi-tiny-MoE and meta-device loading breakage are separate bugs." }, { "left": "issue:43257", "right": "issue:44863", "accept": false, - "reason": "Different model families and different loading/conversion failures." + "reason": "Both are loading/checkpoint problems, but one is Qwen3 MoE conversion with accelerate+deepspeed and the other is NemotronH checkpoint compatibility; not the same bug." }, { "left": "issue:43425", "right": "issue:43828", "accept": false, - "reason": "Torch version incompatibility is too broad to merge with a specific autocast dtype-mismatch bug." + "reason": "Torch version incompatibility and autocast dtype mismatch are different failure modes." }, { "left": "issue:44589", "right": "issue:44756", "accept": false, - "reason": "Float8 storage lookup failure and Strix Halo mmap/OOM are unrelated." + "reason": "Float8 storage deserialization and mmap OOM are unrelated." }, { "left": "issue:44479", "right": "issue:45663", "accept": false, - "reason": "Qwen video-input regression and Gemma-4 FSDP cache/state bug are unrelated." + "reason": "Qwen video regression and Gemma4 FSDP2 shared_kv_states KeyError are unrelated." }, { "left": "issue:43906", "right": "issue:44530", "accept": false, - "reason": "A reproduction of an earlier issue and a PagedAttentionCache/Qwen3.5 crash are not the same bug." + "reason": "An isolated reproduction of another issue and a PagedAttentionCache linear_attention crash are not clearly the same bug." }, { "left": "issue:39404", "right": "issue:45701", "accept": false, - "reason": "Whisper pipeline `return_language` regression and tokenizer-version changes are unrelated." + "reason": "Whisper pipeline return_language and version-dependent tokenization changes are unrelated." }, { "left": "issue:43611", "right": "issue:43645", "accept": false, - "reason": "Base-model-prefix loading regression and notebook-defined custom-model initialization are different code paths." + "reason": "Both are Transformers 5.0 custom-model regressions, but one is about base_model_prefix loading and the other about notebook-based initialization; too different to merge as duplicates." }, { "left": "issue:44530", "right": "issue:45084", "accept": false, - "reason": "PagedAttentionCache linear_attention crash and template-node compilation error are unrelated." + "reason": "PagedAttentionCache linear_attention crash and compile non-template-nodes TypeError are unrelated." }, { - "left": "issue:43931", - "right": "issue:44863", + "left": "issue:43425", + "right": "issue:43716", "accept": false, - "reason": "Different model families and different loading errors." + "reason": "Torch 2.10 incompatibility and Mistral-3 image-preprocessor/model dtype mismatch are different issues." }, { - "left": "issue:43425", - "right": "issue:43716", + "left": "issue:43931", + "right": "issue:44863", "accept": false, - "reason": "Torch version compatibility and Mistral-3 image-preprocessor dtype mismatch are unrelated." + "reason": "Qwen3-VL weight-shape mismatch and NemotronH checkpoint loading are different model-specific load failures." }, { "left": "issue:43901", "right": "issue:44509", "accept": false, - "reason": "Both are docs issues, but they cover different pipeline deprecations and should not be merged." + "reason": "Both are docs issues, but they cover different pipeline/task deprecations and do not describe the same change." }, { "left": "issue:39404", "right": "issue:44912", "accept": false, - "reason": "Whisper pipeline language-return regression and MXFP4 quantization fallback are unrelated." + "reason": "Whisper pipeline language return and quantization loading/fallback issues are unrelated." }, { "left": "issue:44863", "right": "issue:45663", "accept": false, - "reason": "NemotronH checkpoint loading and Gemma-4 FSDP state rebuilding are unrelated." + "reason": "NemotronH checkpoint loading and Gemma4 FSDP2 KeyError are distinct model-loading/training bugs." }, { "left": "issue:44265", "right": "issue:44898", "accept": false, - "reason": "`torch.export`/`torch_compilable_check` failure and Perceiver interpolation failure are different bugs." + "reason": "torch.export.export with torch_compilable_check and Perceiver resize/interpolation failure are different problems." }, { "left": "issue:43299", "right": "issue:43931", "accept": false, - "reason": "Both involve Qwen3-VL loading, but one is a v5 regression for MoE models and the other is a specific checkpoint shape mismatch." + "reason": "Both involve Qwen3-VL, but one is a dev-version MoE loading regression and the other is a checkpoint shape mismatch; not clearly the same underlying bug." }, { "left": "issue:44479", "right": "issue:44805", "accept": false, - "reason": "Video-input regression and mask/tensor shape mismatch are unrelated." + "reason": "Video-input regression and mask/index shape mismatch are not the same issue." }, { "left": "issue:43856", "right": "issue:44756", "accept": false, - "reason": "Qwen3 MoE memory usage and Strix Halo mmap/OOM are unrelated." - }, - { - "left": "issue:43425", - "right": "issue:44918", - "accept": false, - "reason": "Torch compatibility and TRL SFT input-embedding unpacking are different bugs." + "reason": "Qwen3 MoE training memory inefficiency and Strix Halo mmap OOM are unrelated performance/memory issues." }, { "left": "issue:44589", "right": "issue:45084", "accept": false, - "reason": "Float8 storage lookup failure and non-template-node compilation failure are unrelated." + "reason": "Float8 storage lookup failure and compile-on-template-nodes error are unrelated." } ] }, "evaluator_result": { "accept": true, - "feedback": "The summary is grounded in the packet: most pairs are conservatively rejected as different bugs, and the two accepted edges around `resize_token_embeddings` are well-supported by the issue titles and high similarity. The canonical issue choice is also reasonable for the shared embedding-propagation bug." + "feedback": "The summary is well grounded in the packet and stays mostly conservative. The resize_token_embeddings trio (45276, 45292, 45335) is the only clearly duplicate-looking subcluster from the titles, and the other pairwise rejections are cautious rather than overstated." }, "evaluator_used": true, "retried": false, @@ -2414,8 +2293,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -2444,7 +2323,6 @@ "issue:43646", "issue:43716", "issue:43883", - "issue:43957", "issue:43994", "issue:44079", "issue:44164", @@ -2452,6 +2330,7 @@ "issue:44220", "issue:44451", "issue:44466", + "issue:44479", "issue:44488", "issue:44530", "issue:44671", @@ -2475,19 +2354,19 @@ "issue:39404|issue:44671", "issue:43994|issue:44079", "issue:44671|issue:44991", - "issue:36032|issue:43452", "issue:44164|issue:45759", + "issue:36032|issue:43452", "issue:44991|issue:45081", "issue:32090|issue:39404", - "issue:43957|issue:44164", + "issue:44164|issue:44479", "issue:44488|issue:45081", "issue:40990|issue:43421", "issue:43582|issue:43632", "issue:43883|issue:44488", "issue:43611|issue:43644", "issue:44206|issue:44220", - "issue:44811|issue:45584", "issue:43716|issue:45084", + "issue:44811|issue:45584", "issue:44488|issue:44987", "issue:34567|issue:40444", "issue:43643|issue:44530", @@ -2495,8 +2374,8 @@ "issue:43883|issue:44991", "issue:44451|issue:44987", "issue:36032|issue:42915", - "issue:43452|issue:43482", "issue:43541|issue:45084", + "issue:43452|issue:43482", "issue:42915|issue:44530", "issue:43317|issue:43482", "issue:44849|issue:45325", @@ -2504,244 +2383,244 @@ "issue:36032|issue:43645", "issue:45084|issue:45507" ], - "prepared_review_unit_hash": "19ba22a1c0cccd5e491c58eec52701712f504e9e46f66cf8882549d7691a42b0", + "prepared_review_unit_hash": "9e21ceda20d06098cbb57ef658b06518e23060997fc7ba19d0709546217658a3", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 34, - "serialized_chars": 37904, - "estimated_input_tokens": 9476, - "estimated_eval_tokens": 19208 + "serialized_chars": 37891, + "estimated_input_tokens": 9473, + "estimated_eval_tokens": 19202 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is mostly a loose collection of unrelated Transformers regressions. The only likely duplicate-style overlap is the GGUF loading pair; the rest look like separate bugs affecting different code paths or models.", - "confidence": 0.79, - "canonical_issue_reason": "issue:43452 is the strongest anchor because it describes a concrete, broad GGUF loading regression affecting both tokenizer and CausalLM from_pretrained paths.", + "summary": "This cluster is mostly a grab-bag of unrelated Transformers regression reports, with one clear duplicate-like GGUF loading pair. The most representative issue is the generic v5 model-loading regression around `base_model_prefix`/core loading behavior.", + "confidence": 0.69, + "canonical_issue_reason": "Issue 43611 is the broadest, most central regression report in the set: a Transformers v5 model-loading failure affecting the core loading path, which matches the overall cluster theme better than the more model-specific reports.", "canonical_pr_reason": null, - "best_issue_reason": "issue:43452 is the best standalone issue to keep because it is the most general and actionable description of the GGUF-loading failure in this set.", + "best_issue_reason": "43611 is the best single issue to anchor this cluster because it is generic, high-signal, and closest to the shared v5 loading-regression pattern seen across several nearby issues.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43596", "right": "issue:44683", "accept": false, - "reason": "Different failure surfaces: DeepSpeed ZeRO3 initialization vs compiled flex_attention on newer torch." + "reason": "Different failures in different stacks: DeepSpeed ZeRO-3 initialization vs compiled flex_attention on torch>=2.9." }, { "left": "issue:40990", "right": "issue:42947", "accept": false, - "reason": "Unrelated bugs: perplexity quality issue vs gradient checkpointing/LoRA behavior." + "reason": "Unrelated bugs: perplexity on a specific model/dataset vs gradient-checkpointing behavior with PEFT LoRA." }, { "left": "issue:43425", "right": "issue:45685", "accept": false, - "reason": "Torch version compatibility on one side and an MPS histogram-op error on the other; not the same bug." + "reason": "No shared code path; one is a Torch version incompatibility, the other is an MPS-specific unsupported op in a MoE path." }, { "left": "issue:39404", "right": "issue:44671", "accept": false, - "reason": "Whisper pipeline language return bug and CamemBERT masked-LM predictions are separate model-specific issues." + "reason": "Different model families and symptoms: Whisper pipeline return_language vs CamemBERT masked-LM predictions." }, { "left": "issue:43994", "right": "issue:44079", "accept": false, - "reason": "SigLIP inference quality regression vs ModelOutput key assignment logic; different code paths." + "reason": "Both mention incorrect outputs, but the concrete bugs differ and there is no clear shared code-path evidence." }, { "left": "issue:44671", "right": "issue:44991", "accept": false, - "reason": "Different models and failures: CamemBERT prediction bug vs tokenizer loading for est-roberta." + "reason": "Separate tokenizer/model loading issues for different models; no indication they share the same underlying regression." }, { - "left": "issue:36032", - "right": "issue:43452", + "left": "issue:44164", + "right": "issue:45759", "accept": false, - "reason": "Tokenizer special-token method conflict is unrelated to GGUF loading failure." + "reason": "save/from_pretrained extra_state handling is unrelated to `from_config` unwrapping of `text_config` in composite configs." }, { - "left": "issue:44164", - "right": "issue:45759", + "left": "issue:36032", + "right": "issue:43452", "accept": false, - "reason": "Save/load extra_state handling and from_config text_config unwrapping are distinct loader issues." + "reason": "T5 tokenizer special-token method conflict is unrelated to GGUF loading failures." }, { "left": "issue:44991", "right": "issue:45081", "accept": false, - "reason": "Both involve tokenizers, but one is a tokenizer loading regression and the other is a Mistral regex patch crash; not the same bug." + "reason": "Both are tokenizer-loading problems, but for different models and different failure mechanisms." }, { "left": "issue:32090", "right": "issue:39404", "accept": false, - "reason": "Trainer GPU broadcast NoneType error does not match Whisper pipeline return_language behavior." + "reason": "Trainer GPU broadcast TypeError is unrelated to Whisper pipeline return_language behavior." }, { - "left": "issue:43957", - "right": "issue:44164", + "left": "issue:44164", + "right": "issue:44479", "accept": false, - "reason": "Meta-device model loading and extra_state save/load handling are different regressions." + "reason": "Save/load extra_state handling and Qwen VL video input regression are unrelated." }, { "left": "issue:44488", "right": "issue:45081", "accept": false, - "reason": "Loading a specific BERT model and crashing in Mistral regex patching are unrelated." + "reason": "Different model loaders and different root causes; only a broad tokenizer-loading theme is shared." }, { "left": "issue:40990", "right": "issue:43421", "accept": false, - "reason": "Perplexity regression vs runtime special-token post-processor update behavior." + "reason": "Perplexity quality issue vs runtime post-processor update bug; not the same bug or fix." }, { "left": "issue:43582", "right": "issue:43632", "accept": false, - "reason": "Apple Silicon allocator warmup TypeError and _is_hf_initialized flag breakage are separate issues." + "reason": "Apple Silicon allocator TypeError and `_is_hf_initialized` regression are unrelated." }, { "left": "issue:43883", "right": "issue:44488", "accept": false, - "reason": "Missing tied-weights key attribute vs model loading failure for a specific BERT repo are not the same underlying bug." + "reason": "Missing tied-weights key attribute in Molmo is unrelated to a BERT model loading failure." }, { "left": "issue:43611", "right": "issue:43644", "accept": false, - "reason": "Both are v5 regressions, but one is base_model_prefix loading and the other is non-persistent buffer serialization; different fixes." + "reason": "Both are v5 regressions, but one is model-loading/base_model_prefix and the other is non-persistent buffer initialization; different bugs." }, { "left": "issue:44206", "right": "issue:44220", "accept": false, - "reason": "Likely related audio feature-extraction area, but the titles point to different failures and there is not enough evidence of one shared bug." + "reason": "The titles suggest the same audio area, but the concrete failures are not clearly the same code-path problem." }, { - "left": "issue:44811", - "right": "issue:45584", + "left": "issue:43716", + "right": "issue:45084", "accept": false, - "reason": "Both are Whisper token-handling issues, but one is batch_decode skipping special tokens and the other is empty-transcription generation after align_special_tokens." + "reason": "Image preprocessor dtype mismatch is unrelated to a template-node compilation error." }, { - "left": "issue:43716", - "right": "issue:45084", + "left": "issue:44811", + "right": "issue:45584", "accept": false, - "reason": "Image preprocessor dtype mismatch and Jinja/template compilation failure are unrelated." + "reason": "Both involve Whisper and special tokens, but one is batch_decode skipping tokens and the other is empty-transcription generation after alignment; too different to merge." }, { "left": "issue:44488", "right": "issue:44987", "accept": false, - "reason": "Different loading failures for different models; no shared code-path evidence." + "reason": "Different model-loading failures with no shared concrete regression evidence." }, { "left": "issue:34567", "right": "issue:40444", "accept": false, - "reason": "Trainer token-count tracking and multi-image IterableDataset fine-tuning are unrelated." + "reason": "Trainer token-count tracking is unrelated to multi-image IterableDataset fine-tuning failure." }, { "left": "issue:43643", "right": "issue:44530", "accept": false, - "reason": "AutoConfig trust_remote_code field loss is unrelated to PagedAttentionCache's linear_attention group-type crash." + "reason": "AutoConfig trust_remote_code field loss is unrelated to PagedAttentionCache crashing on Qwen3.5 linear_attention." }, { "left": "issue:43644", "right": "issue:43646", "accept": false, - "reason": "Both mention v5 model initialization problems, but one is non-persistent buffers and the other is custom model init semantics; not the same bug." + "reason": "A non-persistent buffer initialization regression is not the same as custom model initialization in Jupyter notebooks." }, { "left": "issue:43883", "right": "issue:44991", "accept": false, - "reason": "Different symptoms and subsystems: tied-weights attribute vs tokenizer load regression." + "reason": "Molmo tied-weights attribute error and tokenizer loading for EMBEDDIA/est-roberta are unrelated." }, { "left": "issue:44451", "right": "issue:44987", "accept": false, - "reason": "Model-specific loading failure vs a broad transformers>=5.1.0 regression loading a different repo." + "reason": "Different model-specific loading failures with no clear shared implementation bug." }, { "left": "issue:36032", "right": "issue:42915", "accept": false, - "reason": "Tokenizer class conflict and FineGrainedFP8Config failure are unrelated." - }, - { - "left": "issue:43452", - "right": "issue:43482", - "accept": true, - "reason": "Both point to the same GGUF loading regression in transformers v5, affecting from_pretrained loading of GGUF models/tokenizers." + "reason": "Tokenizer method-name conflict is unrelated to Qwen3Moe FineGrainedFP8Config failure." }, { "left": "issue:43541", "right": "issue:45084", "accept": false, - "reason": "Torch dynamo tracing grouped_mm failure is unrelated to template compilation errors." + "reason": "TorchDynamo grouped_mm tracing error is unrelated to template-node compilation." + }, + { + "left": "issue:43452", + "right": "issue:43482", + "accept": true, + "reason": "Both are GGUF loading regressions in Transformers v5 affecting `from_pretrained`; they plausibly come from the same GGUF loader-path bug." }, { "left": "issue:42915", "right": "issue:44530", "accept": false, - "reason": "FineGrainedFP8Config failure and PagedAttentionCache group-type crash are separate MoE/model-loading issues." + "reason": "Different issues in different subsystems: FineGrainedFP8Config vs PagedAttentionCache linear_attention crash." }, { "left": "issue:43317", "right": "issue:43482", "accept": false, - "reason": "Dequantized model offload/device_map loading and GGUF model loading are different bugs." + "reason": "GPU+CPU offload dequantized-model loading and GGUF loading in v5 are distinct failures." }, { "left": "issue:44849", "right": "issue:45325", "accept": false, - "reason": "Qwen3.5 hidden-state output bug and Qwen2.5-VL rope-index scaling bug are distinct model-specific regressions." + "reason": "Qwen hidden-states behavior and Qwen2.5-VL rope-index scaling are different bugs." }, { "left": "issue:43404", "right": "issue:44466", "accept": false, - "reason": "Tied lm_head weight bug in Mistral3 and serialization inconsistency across devices are related only loosely, not the same concrete defect." + "reason": "Both mention tied weights, but one is a Mistral3 model definition bug and the other is device-dependent serialization; not clearly the same fix." }, { "left": "issue:36032", "right": "issue:43645", "accept": false, - "reason": "Tokenizer special-token conflict is unrelated to custom model initialization in Jupyter notebooks." + "reason": "Tokenizer special-token conflict is unrelated to custom model initialization in notebooks." }, { "left": "issue:45084", "right": "issue:45507", "accept": false, - "reason": "Template compilation failure and GraniteMoEHybrid invalid-method calls are unrelated code paths." + "reason": "Template compilation failure and GraniteMoEHybrid invalid method call are unrelated." } ] }, "evaluator_result": { "accept": true, - "feedback": "The summary is grounded in the packet and stays conservative about the soft-edge pairs. The GGUF overlap is the only clearly supported duplicate-style link, and the other false verdicts are adequately cautious given the titles alone." + "feedback": "The cluster read is mostly grounded in the packet: the chosen anchor issue 43611 is a reasonable broad v5 loading-regression representative, and the rejected soft pairs are generally conservative and tied to the issue titles. The only accepted GGUF pair is phrased as plausible rather than certain, which keeps it within bounds. No mergeability concerns apply here because these are issues, not PR pairs." }, "evaluator_used": true, "retried": false, @@ -2757,8 +2636,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -2819,10 +2698,10 @@ "issue:43322|issue:43540", "issue:43377|issue:43819", "issue:44220|issue:44683", - "issue:44451|issue:45081", "issue:45216|issue:45357", - "issue:43645|issue:44534", + "issue:44451|issue:45081", "issue:42915|issue:45084", + "issue:43645|issue:44534", "issue:44265|issue:44610", "issue:44987|issue:45081", "issue:43404|issue:45127", @@ -2836,26 +2715,26 @@ "issue:43611|issue:44534", "issue:42915|issue:44589", "issue:40990|issue:44811", - "issue:43317|issue:43856", "issue:44479|issue:45325", + "issue:43317|issue:43856", "issue:42175|issue:45542", "issue:43425|issue:44403", "issue:43818|issue:45357" ], - "prepared_review_unit_hash": "6396ecb9b1a351d457ebd541364f3cd2c1c39731e5e1093122e5fa1e7b25d38e", + "prepared_review_unit_hash": "4e64e49d7f90b6fe2920451b49ddb20f97a5039dd2df1390722f78b44cef57d1", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, "soft_pair_count": 30, - "serialized_chars": 36044, + "serialized_chars": 36043, "estimated_input_tokens": 9011, "estimated_eval_tokens": 18278 }, @@ -2864,196 +2743,202 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is mostly a grab bag of unrelated Transformers v5/model-loading regressions. I only see two likely duplicate subpairs: Qwen3.5 save_pretrained checkpoint corruption and TensorFlow backend packaging/import failures.", - "confidence": 0.79, - "canonical_issue_reason": "issue:43611 is the broadest representative of the dominant theme here: Transformers v5 loading regressions in model initialization/loading paths.", + "summary": "This cluster is mostly unrelated issue reports with broad Transformers/v5 overlap. The Qwen3.5 save_pretrained pair is the only close match, but it should be treated as likely related rather than definitively the same bug from titles alone.", + "confidence": 0.69, + "canonical_issue_reason": "issue:45216 is the broadest Qwen3.5 save_pretrained regression report and the best anchor for that mini-cluster, but 45357 should only be considered likely related without stronger body-level evidence.", "canonical_pr_reason": null, - "best_issue_reason": "issue:43611 is the best single anchor for the cluster because it captures the core loading-path regression theme more generally than the narrower model-specific reports.", + "best_issue_reason": "issue:45216 is the best issue anchor because it names the core Qwen3.5 save_pretrained regression and is broader than the later, more specific report.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43262", "right": "issue:43377", "accept": false, - "reason": "Audio chat-template sampling-rate behavior and MIMI batch padding-mask behavior are different model/component bugs." + "reason": "Different audio bugs: sampling-rate defaulting in apply_chat_template vs batched/single mismatch from missing padding-mask support." }, { "left": "issue:43646", "right": "issue:44534", "accept": false, - "reason": "Custom model initialization and non-persistent buffer corruption are different v5 regressions with different code paths." + "reason": "Different v5 regressions: custom model initialization vs non-persistent buffers being filled with junk." }, { "left": "issue:43994", "right": "issue:44610", "accept": false, - "reason": "SigLIP2 nonsensical outputs and OmDet-Turbo processor size mismatch are unrelated model/preprocessing issues." + "reason": "Different models and failure modes: SigLIP2 bad outputs vs OmDet-Turbo processor/model input-size mismatch." }, { "left": "issue:44987", "right": "issue:44991", "accept": false, - "reason": "Both are loading failures, but for different models and likely different root causes; not one concrete bug." + "reason": "Different affected models and code paths: loading physical-intelligence/fast vs tokenizer loading for EMBEDDIA/est-roberta." }, { "left": "issue:44403", "right": "issue:45081", "accept": false, - "reason": "Generic loading noise and the Mistral tokenizer backend_tokenizer crash are different problems." + "reason": "Unrelated problems: generic loading noise vs a Mistral tokenizer regex patch crash." }, { "left": "issue:43322", "right": "issue:43540", "accept": false, - "reason": "Both are multimodal crashes, but one is Llava Next segfaulting and the other is Qwen3OmniMoe video ValueError; different bugs." + "reason": "Different multimodal models and bugs: Llava Next segfault vs Qwen3OmniMoe video-processing ValueError." }, { "left": "issue:43377", "right": "issue:43819", "accept": false, - "reason": "MIMI padding-mask mismatch and DAC.from_latents STE mismatch are unrelated audio-model correctness bugs." + "reason": "Both are audio/model correctness issues, but they affect different models and distinct mechanisms: padding masks vs missing STE in DAC." }, { "left": "issue:44220", "right": "issue:44683", "accept": false, - "reason": "Fbank feature extraction and compiled flex_attention on newer torch are unrelated issues." - }, - { - "left": "issue:44451", - "right": "issue:45081", - "accept": false, - "reason": "Specific model loading failure and Mistral tokenizer patch crash are unrelated." + "reason": "Different subsystems: audio fbank extraction vs compiled flex_attention on newer torch." }, { "left": "issue:45216", "right": "issue:45357", - "accept": true, - "reason": "Both describe the same Qwen3.5 save_pretrained checkpoint corruption/incorrect saved keys on the same code path." + "accept": false, + "reason": "Likely related Qwen3.5 save_pretrained regressions, but the titles alone do not prove the same root cause or that 45357 is a direct manifestation of 45216." }, { - "left": "issue:43645", - "right": "issue:44534", + "left": "issue:44451", + "right": "issue:45081", "accept": false, - "reason": "Notebook custom-model initialization and junk-filled non-persistent buffers are different regressions." + "reason": "Different models and failures: ScandiBERT load failure vs Mistral regex-patching crash." }, { "left": "issue:42915", "right": "issue:45084", "accept": false, - "reason": "Qwen3Moe FineGrainedFP8Config failure and template-node compilation error are unrelated." + "reason": "Different bug classes: FineGrainedFP8Config load failure vs template-node compilation TypeError." + }, + { + "left": "issue:43645", + "right": "issue:44534", + "accept": false, + "reason": "Different regressions: notebook/custom-model initialization issues vs corrupted non-persistent buffers." }, { "left": "issue:44265", "right": "issue:44610", "accept": false, - "reason": "torch.export/torch_compilable_check failure and OmDet input-size mismatch are unrelated." + "reason": "Different code paths: torch.export with torch_compilable_check vs processor output size mismatch." }, { "left": "issue:44987", "right": "issue:45081", "accept": false, - "reason": "Different model-loading/tokenizer failure modes; too little to treat as one bug." + "reason": "Different affected components: model loading for physical-intelligence/fast vs Mistral tokenizer regex patching." }, { "left": "issue:43404", "right": "issue:45127", "accept": false, - "reason": "lm_head tying in Mistral3ForConditionalGeneration and LoRA merge collapse are different problems." + "reason": "Related to weight tying in broad terms, but one is Mistral3 lm_head tying and the other is LoRA merge with extended vocab." }, { "left": "issue:41553", "right": "issue:43582", "accept": false, - "reason": "Tokenizer-loading error and Apple Silicon allocator warmup TypeError are unrelated." + "reason": "Different domains: AutoTokenizer error messaging for Voxtral vs Apple Silicon warmup TypeError." }, { "left": "issue:43632", "right": "issue:44756", "accept": false, - "reason": "_is_hf_initialized regression and mmap/OOM behavior are unrelated." + "reason": "Different regressions: _is_hf_initialized handling vs mmap OOM behavior on Strix Halo." }, { "left": "issue:43296", "right": "issue:43531", "accept": false, - "reason": "PaddleOCR-VL load failure and Qwen3-MoE sliding_window issue are different bugs." + "reason": "Different models and bugs: PaddleOCR-VL load failure in vLLM vs Qwen3-MoE sliding_window issue." }, { "left": "issue:44062", "right": "issue:44987", "accept": false, - "reason": "AddedToken special-argument conflict and model loading failure for a specific repo are unrelated." + "reason": "Tokenizer AddedToken keyword duplication is unrelated to loading physical-intelligence/fast." }, { "left": "issue:42915", "right": "issue:44291", "accept": false, - "reason": "Both involve v5 initialization, but one is FineGrainedFP8Config and the other is init_empty_weights/_is_hf_initialized; different concrete failures." + "reason": "Different failures in different paths: Qwen3Moe FP8 config vs unexpected init_empty_weights argument." }, { "left": "issue:36032", "right": "issue:43742", "accept": false, - "reason": "T5 tokenizer special-token conflict and MobileLLM key error are unrelated loading issues." + "reason": "Different model/tokenizer failures: T5Tokenizer special-token conflict vs MobileLLM key error." }, { "left": "issue:43531", "right": "issue:44877", "accept": false, - "reason": "Qwen3-MoE sliding_window and granite_speech strict config loading are different bugs." + "reason": "Different areas and symptoms: Qwen3-MoE sliding-window behavior vs granite_speech strict-config loading." }, { "left": "issue:43611", "right": "issue:44534", "accept": false, - "reason": "Both are v5 loading regressions, but one is base_model_prefix handling and the other is non-persistent buffer initialization; not the same concrete bug." + "reason": "Different v5 regressions: base_model_prefix loading breakage vs junk values in non-persistent buffers." }, { - "left": "issue:40990", - "right": "issue:44811", + "left": "issue:42915", + "right": "issue:44589", "accept": false, - "reason": "High perplexity on gpt-oss and Whisper batch_decode skip_special_tokens behavior are unrelated." + "reason": "Different failures in different paths: Qwen3Moe FP8 config vs missing Float8 storage type." }, { - "left": "issue:43317", - "right": "issue:43856", + "left": "issue:40990", + "right": "issue:44811", "accept": false, - "reason": "device_map/offload loading failure and Qwen3 MoE training memory use are related only broadly by memory concerns, not the same bug." + "reason": "Unrelated issues: perplexity regression on gpt-oss-20b vs Whisper batch_decode special-token handling." }, { "left": "issue:44479", "right": "issue:45325", "accept": false, - "reason": "Both touch Qwen video/VL models, but one is a video-input regression and the other is a rope-index scaling bug for still images; different code paths." + "reason": "Both involve video/VL behavior, but the affected models and concrete bugs differ: video-input regression vs rope-index scaling." + }, + { + "left": "issue:43317", + "right": "issue:43856", + "accept": false, + "reason": "Different problems: device_map/offload loading failure vs Qwen3 MoE training memory usage." }, { "left": "issue:42175", "right": "issue:45542", - "accept": true, - "reason": "Both point to the same TensorFlow-backend packaging/import problem where TF is missing or only partially installed under non-TF extras." + "accept": false, + "reason": "Related TensorFlow packaging themes, but these are different packaging/backend failures and not the same bug." }, { "left": "issue:43425", "right": "issue:44403", "accept": false, - "reason": "Torch version incompatibility and generic loading noise are unrelated." + "reason": "Torch version incompatibility is unrelated to generic transformer-loading noise." }, { "left": "issue:43818", "right": "issue:45357", "accept": false, - "reason": "Video-LLaVA temporal-attention/weight-sharing bug and Qwen3.5 save_pretrained regression are unrelated model families and code paths." + "reason": "Different models and code paths: Video-LLaVA weight/temporal-attention issues vs Qwen3.5 save_pretrained serialization." } ] }, "evaluator_result": { - "accept": true, - "feedback": "Grounded and conservative overall. The summary does not overclaim beyond the packet, and the two accepted pairs (45216/45357, 42175/45542) look plausibly duplicate enough from the titles/descriptions. The other soft-edge rejections stay appropriately cautious." + "accept": false, + "feedback": "The cluster summary is mostly grounded, but the accepted 45216\u219445357 linkage is stated too strongly for the packet. The titles support that both are Qwen3.5 save_pretrained regressions, but not that they are definitively the same underlying bug or that 45357 is proven to be a more specific manifestation of 45216. Please soften that verdict to something like 'likely related' unless you have body-level evidence." }, "evaluator_used": true, - "retried": false, - "accepted_nontrivial_soft_edge": true, + "retried": true, + "accepted_nontrivial_soft_edge": false, "error_kind": null, "error_message": null }, @@ -3065,8 +2950,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -3132,8 +3017,8 @@ "issue:44829|issue:45479", "issue:40444|issue:45325", "issue:43482|issue:43525", - "issue:43377|issue:45491", "issue:44291|issue:44589", + "issue:43377|issue:45491", "issue:41720|issue:45081", "issue:44265|issue:45072", "issue:42175|issue:43296", @@ -3155,248 +3040,245 @@ "issue:35707|issue:40444", "issue:44291|issue:44877" ], - "prepared_review_unit_hash": "52207d60e831ca0d8a25123e19455468cbb6d0d4e543ba7ad4e5f9875651462f", + "prepared_review_unit_hash": "d9e3f7ba65062e2ca027a88be24c1586969e49e92cc5f3e2ac8ee32fa5d53156", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 34, - "serialized_chars": 38078, - "estimated_input_tokens": 9520, - "estimated_eval_tokens": 19296 + "serialized_chars": 38070, + "estimated_input_tokens": 9518, + "estimated_eval_tokens": 19292 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This set is mostly a grab bag of unrelated Transformers/model-specific bugs. The only soft edge that looks plausibly like the same underlying defect is the Qwen2.5-VL position-id / rope-index regression pair.", - "confidence": 0.78, - "canonical_issue_reason": "issue:45325 is the clearest root-cause report: it points to `get_rope_index` incorrectly scaling still-image temporal `position_ids` by `tokens_per_second` in Qwen2.5-VL. That is more specific and more actionable than the companion symptom report.", + "summary": "This cluster is mostly a collection of separate Transformers v5 / model-loading regressions, not one duplicate set. The closest center is a broad initialization/loading regression, but most pairs are only thematically similar.", + "confidence": 0.38, + "canonical_issue_reason": "issue:43632 is the strongest concrete anchor for the cluster: a core Transformers v5 initialization regression (`_is_hf_initialized`) that fits the many nearby model-loading breakages better than the narrower, model-specific reports.", "canonical_pr_reason": null, - "best_issue_reason": "issue:45325 is the best representative of the duplicate-like pair because it names the concrete failing code path and likely underlying fix. The rest of the set does not form one coherent duplicate cluster.", + "best_issue_reason": "issue:45020 is the best cluster representative overall because it is the broadest umbrella for the shared theme of recent Transformers versions breaking model loading, especially for remote-code models.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43421", "right": "issue:44568", "accept": false, - "reason": "Both are tokenizer issues, but one is about runtime post-processor updates and the other about BOS/EOS insertion in a specific tokenizer; different bugs." + "reason": "Both are tokenizer/special-token regressions, but they describe different failure modes and code paths: runtime post-processor updates vs BOS/EOS insertion." }, { "left": "issue:44079", "right": "issue:45663", "accept": false, - "reason": "`ModelOutput` key handling and Gemma-4 FSDP2 `shared_kv_states` rebuilding are unrelated code paths." + "reason": "One is a ModelOutput key assignment bug; the other is a Gemma-4 FSDP2 attention state KeyError. They are unrelated." }, { "left": "issue:44683", "right": "issue:45020", "accept": false, - "reason": "A torch 2.9 `flex_attention` compile failure is not the same as broad remote-code loading regressions." + "reason": "torch >= 2.9 flex_attention compilation failure is a different problem from remote_code model loading regressions." }, { "left": "issue:36032", "right": "issue:41720", "accept": false, - "reason": "Tokenizer method-name conflict and Qwen3 auto device-mapping CUDA assert are unrelated." + "reason": "Tokenizer method-name conflict and Qwen3 auto device mapping CUDA assert are unrelated." }, { "left": "issue:43632", "right": "issue:44534", "accept": false, - "reason": "Both mention Transformers v5 regressions, but one is `_is_hf_initialized` and the other is non-persistent buffers; different failures." + "reason": "Both are v5 regressions, but one concerns `_is_hf_initialized` during init and the other concerns non-persistent buffer contents; different bugs." }, { "left": "issue:43792", "right": "issue:44589", "accept": false, - "reason": "Whisper model loading and Float8 storage lookup are different error modes." + "reason": "Whisper loading failure and missing Float8 storage are distinct load/runtime issues." }, { "left": "issue:44403", "right": "issue:44991", "accept": false, - "reason": "Generic loading noise is not the same as tokenizer loading failure for a specific model." + "reason": "Generic load noise and a specific tokenizer load failure for EMBEDDIA/est-roberta are not the same bug." }, { "left": "issue:43482", "right": "issue:45538", "accept": false, - "reason": "GGUF loading failure and CLIPTokenizer `model_max_length` behavior are unrelated." + "reason": "GGUF loading under v5 and CLIPTokenizer max length behavior are unrelated." }, { "left": "issue:33666", "right": "issue:40444", "accept": false, - "reason": "Both involve Qwen2-VL fine-tuning, but multi-GPU training and multi-image iterable dataset failures are different bugs." + "reason": "Both involve Qwen VL training, but one is multi-GPU training and the other is iterable-dataset/multi-image prompt handling; not the same issue." }, { "left": "issue:44829", "right": "issue:45479", "accept": false, - "reason": "Flash-attention training degeneration and zero-loss single-label classification are distinct issues." + "reason": "Degenerate training with flash_attention_3 is a different defect from the single-label-classification zero-loss bug." }, { "left": "issue:40444", "right": "issue:45325", "accept": false, - "reason": "Multi-image prompt handling and still-image temporal position-id scaling are different Qwen2.5-VL failures." + "reason": "Both are Qwen2.5-VL problems, but they affect different paths: dataset/multi-image finetuning vs still-image temporal position-id scaling." }, { "left": "issue:43482", "right": "issue:43525", "accept": false, - "reason": "Model loading failure for Qwen2.5-GGUF and missing `pad_token_id` on Llama4Config are unrelated." + "reason": "Qwen2.5-GGUF loading failure and missing pad_token_id on Llama4Config are unrelated loading errors." }, { - "left": "issue:43377", - "right": "issue:45491", + "left": "issue:44291", + "right": "issue:44589", "accept": false, - "reason": "Batched-vs-single MIMI padding-mask mismatch and Gemma3 all-padding-window NaNs are different bugs." + "reason": "Unexpected `_is_hf_initialized` during init-empty-weights is not the same as the Float8 storage lookup failure." }, { - "left": "issue:44291", - "right": "issue:44589", + "left": "issue:43377", + "right": "issue:45491", "accept": false, - "reason": "Unexpected `_is_hf_initialized` init arg and Float8 storage lookup are unrelated." + "reason": "Missing padding-mask support in MIMI encoder and Gemma3 sliding-window NaNs are separate batching/attention bugs." }, { "left": "issue:41720", "right": "issue:45081", "accept": false, - "reason": "Qwen3 device-mapping crash and Mistral tokenizer regex patch crash are different code paths." + "reason": "Qwen3 device-mapping CUDA asserts and a Mistral tokenizer regex crash are unrelated." }, { "left": "issue:44265", "right": "issue:45072", "accept": false, - "reason": "`torch.export`/`torch_compilable_check` and bfloat16 dtype mismatches are separate problems." + "reason": "torch.export/torch_compilable_check failure is distinct from dtype mismatches in SwitchTransformers/TimmWrapperModel." }, { "left": "issue:42175", "right": "issue:43296", "accept": false, - "reason": "Backend packaging for pip extras and PaddleOCR-VL load failure in vLLM are unrelated." + "reason": "Packaging/backend inclusion issue and PaddleOCR-VL loading in vLLM are unrelated." }, { "left": "issue:43632", "right": "issue:45072", "accept": false, - "reason": "Initializer flag regression and dtype mismatch/OOM behavior are not the same bug." + "reason": "Initialization-flag regression and bf16 dtype mismatch are different code-path problems." }, { "left": "issue:43611", "right": "issue:43646", "accept": false, - "reason": "Both are v5 model-init regressions, but one is about `base_model_prefix` loading and the other about custom initialization; not clearly the same defect." + "reason": "Both are broad v5 loading regressions, but one is about `base_model_prefix` and the other about custom model initialization; not the same concrete bug." }, { "left": "issue:44079", "right": "issue:44534", "accept": false, - "reason": "`ModelOutput` field assignment and non-persistent buffer junk values are unrelated." + "reason": "ModelOutput key handling and junk-filled non-persistent buffers are unrelated." }, { "left": "issue:43531", "right": "issue:44291", "accept": false, - "reason": "Qwen3-MoE sliding-window behavior and `init_empty_weights` argument handling are different issues." + "reason": "Qwen3-MoE sliding_window behavior and init-empty-weights `_is_hf_initialized` are separate issues." }, { "left": "issue:43994", "right": "issue:44898", "accept": false, - "reason": "SigLIP2 nonsensical outputs and Perceiver resolution/interpolation failure are not the same bug." + "reason": "SigLIP2 wrong outputs and Perceiver resolution/interpolation failure are different model-specific bugs." }, { "left": "issue:43454", "right": "issue:44466", "accept": false, - "reason": "Both mention tied weights/lm_head, but one is a broken tie at model config time and the other is serialization inconsistency; different code paths." + "reason": "Tied lm_head weights not being set and serialization differences of lm_head.weight are related only thematically, not the same failure." }, { "left": "issue:43881", "right": "issue:44488", "accept": false, - "reason": "Separate model-loading failures for different models." + "reason": "glm-4v-9b loading failure and cjvt/sleng-bert tokenizer loading failure are unrelated." }, { "left": "issue:43828", "right": "issue:44683", "accept": false, - "reason": "Autocast dtype mismatch and compiled flex_attention failure are unrelated." + "reason": "Autocast dtype mismatch and compiled flex_attention failure are different problems." }, { "left": "issue:42915", "right": "issue:43482", "accept": false, - "reason": "Fine-grained FP8 config failure and GGUF loading failure are different problems." + "reason": "FineGrainedFP8Config failure and GGUF loading under v5 are unrelated." }, { "left": "issue:44079", "right": "issue:45072", "accept": false, - "reason": "`ModelOutput` key assignment and bfloat16 dtype mismatch are unrelated." + "reason": "ModelOutput key assignment bug and bf16 dtype mismatch are not the same underlying issue." }, { "left": "issue:43792", "right": "issue:43957", "accept": false, - "reason": "Whisper loading failure and `torch.device(\"meta\")` loading regressions are different classes of bugs." + "reason": "Whisper load failure and meta-device loading regression are distinct." }, { "left": "issue:43644", "right": "issue:44079", "accept": false, - "reason": "Non-persistent buffer junk values and `ModelOutput` key assignment are unrelated." + "reason": "Non-persistent buffer serialization junk and ModelOutput key assignment are different bugs." }, { "left": "issue:40444", "right": "issue:45381", "accept": false, - "reason": "Multiple-images-per-prompt failure and video `vision_position_ids` bug are distinct Qwen2.5-VL issues." + "reason": "Qwen2.5-VL multi-image iterable training failure and video vision_position_ids bug affect different paths." }, { "left": "issue:45325", "right": "issue:45381", - "accept": true, - "reason": "Both reports point to the Qwen2.5-VL position-id/rope-index path; one is the root-cause description and the other is a symptom on video inputs, so they plausibly collapse to one fix." + "accept": false, + "reason": "Both are Qwen2.5-VL position-id complaints, but one is still-image temporal scaling and the other is video vision_position_ids; close, but not clearly the same concrete bug." }, { "left": "issue:44387", "right": "issue:44617", "accept": false, - "reason": "Int4 quantization memory growth and Sam3Video OOM are not the same underlying issue." + "reason": "v5 int4 memory growth/OOM and Sam3Video OOM are different resource issues." }, { "left": "issue:35707", "right": "issue:40444", "accept": false, - "reason": "Progressive generation with `inputs_embeds`/`past_key_values` and Qwen2.5-VL multi-image finetuning failures are unrelated." + "reason": "Progressive generation with inputs_embeds/past_key_values is unrelated to Qwen2.5-VL multi-image iterable finetuning." }, { "left": "issue:44291", "right": "issue:44877", "accept": false, - "reason": "Unexpected init argument handling and strict config loading for `granite_speech` are different problems." + "reason": "Unexpected `_is_hf_initialized` argument and strict config blocking granite_speech are different loading failures." } ] }, - "evaluator_result": { - "accept": true, - "feedback": "The summary stays grounded in the packet: it treats the set as mostly unrelated, and the only accepted soft edge (45325/45381) is plausibly supported by the issue titles around the same Qwen2.5-VL position-id/rope-index path. The other pair judgments are conservative and do not overstate evidence." - }, - "evaluator_used": true, + "evaluator_result": null, + "evaluator_used": false, "retried": false, - "accepted_nontrivial_soft_edge": true, + "accepted_nontrivial_soft_edge": false, "error_kind": null, "error_message": null }, @@ -3408,8 +3290,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -3481,9 +3363,9 @@ "issue:43377|issue:44898", "issue:39692|issue:40444", "issue:44466|issue:45357", + "issue:43792|issue:44683", "issue:43296|issue:44991", "issue:43296|issue:44877", - "issue:43792|issue:44683", "issue:43454|issue:43720", "issue:43873|issue:45701", "issue:41720|issue:44977", @@ -3496,20 +3378,20 @@ "issue:35707|issue:41720", "issue:43873|issue:44756" ], - "prepared_review_unit_hash": "2eded9273d7dc0f0f44cc945ade98ef2c3bcaf343aa321dd9654af051897cec6", + "prepared_review_unit_hash": "83a9dcad506494e60c7e5ecedfa7d64dda1259d04ebb768fdcd1a5015bc50775", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 32, - "serialized_chars": 37537, + "serialized_chars": 37539, "estimated_input_tokens": 9385, "estimated_eval_tokens": 19026 }, @@ -3518,186 +3400,204 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This set is mostly a loose thematic bundle of unrelated Transformers bugs, not a true duplicate cluster. None of the soft edges look safe to merge; 43873 is the broadest and most active issue, but the overall cluster is too heterogeneous for a strong canonical.", - "confidence": 0.88, - "canonical_issue_reason": "issue:43873 is the broadest open report here, with the most discussion/inbound references and an umbrella framing around offloading/quantization problems; it is the least-bad canonical issue choice for triage.", + "summary": "Mostly unrelated closed issues with only a loose shared theme around loading/offloading, config, and multimodal runtime regressions. None of the soft pairs look like true duplicates, so I reject them all and center the cluster on the broad offloading/quantization report.", + "confidence": 0.71, + "canonical_issue_reason": "issue:43873 is the broadest and most central report in the small offloading/quantization subset, with additional discussion and inbound references.", "canonical_pr_reason": null, - "best_issue_reason": "issue:43873 is the best available issue to anchor the cluster because it is open, active, and general enough to serve as a catch-all for nearby offloading complaints, even though the rest of the set is not truly duplicate-related.", + "best_issue_reason": "issue:43873 is the best representative issue because it captures the widest symptom area among the related loading/offloading reports.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43526", "right": "issue:43883", "accept": false, - "reason": "Different bugs: one is a BEiT image processor label-reduction error, the other is a Molmo tied-weights attribute error." + "reason": "Different components and failures: BeitImageProcessorFast label reduction vs Molmo missing attribute during loading." }, { "left": "issue:43317", "right": "issue:43873", "accept": false, - "reason": "Both mention offloading/quantization, but one is a specific gpu+cpu offload load failure for a dequantized model while the other is a broader offloading complaint." + "reason": "Both mention offloading/quantization, but one is a dequantized-model load failure and the other is a generic offloading complaint; not the same bug." }, { "left": "issue:43606", "right": "issue:44265", "accept": false, - "reason": "Bark CPU-offload device mismatch and torch.export failures with torch_compilable_check are unrelated code paths." + "reason": "CPU offload device mismatch is unrelated to torch.export failures around torch_compilable_check." }, { "left": "issue:43881", "right": "issue:44991", "accept": false, - "reason": "A model load failure and a tokenizer loading regression are different failure modes." + "reason": "PaddleOCR-VL model loading in vLLM and tokenizer loading for est-roberta are different failure modes." }, { "left": "issue:43701", "right": "issue:45325", "accept": false, - "reason": "Checkpoint key mismatch is unrelated to Qwen2.5-VL rope-index scaling on still images." + "reason": "Checkpoint-resume key mismatch is unrelated to Qwen2.5-VL rope-index scaling." }, { "left": "issue:44186", "right": "issue:44265", "accept": false, - "reason": "Tokenizer crashes on NER/padding are not the same issue as export failures with torch_compilable_check." + "reason": "Tokenizer crashes on NER/padding are unrelated to torch.export compatibility issues." }, { "left": "issue:43632", "right": "issue:45563", "accept": false, - "reason": "Transformers v5 flag regression and a stale generate warning are unrelated bugs." + "reason": "_is_hf_initialized flag breakage is unrelated to a stale generate warning." }, { "left": "issue:43262", "right": "issue:43746", "accept": false, - "reason": "Sampling-rate defaulting in audio chat templates is unrelated to GraniteSpeech PEFT checkpoint loading." + "reason": "Audio sampling-rate defaulting and PEFT adapter local-checkpoint loading are different code paths." }, { "left": "issue:43606", "right": "issue:43819", "accept": false, - "reason": "CPU offload device mismatch and DAC latent/STE mismatch are different audio-model bugs." + "reason": "CPU offload device mismatch and DAC.from_latents/STE mismatch are unrelated model bugs." }, { "left": "issue:44610", "right": "issue:45325", "accept": false, - "reason": "Processor output resolution mismatch and Qwen2.5-VL temporal position scaling affect different paths." + "reason": "Processor image size mismatch is unrelated to temporal position-id scaling in Qwen2.5-VL." }, { "left": "issue:43454", "right": "issue:43550", "accept": false, - "reason": "Tied lm_head weights and torch.compile SDPA failures are not the same underlying defect." + "reason": "lm_head weight tying and torch.compile SDPA failures are not the same underlying issue." }, { "left": "issue:43377", "right": "issue:44610", "accept": false, - "reason": "Missing padding-mask support in MIMI is unrelated to OmDet-Turbo\u2019s image-size mismatch." + "reason": "MIMI batching/padding-mask behavior and OmDet processor resolution mismatch are unrelated." }, { "left": "issue:43454", "right": "issue:45072", "accept": false, - "reason": "A tied-weights serialization bug is not the same as bfloat16 dtype mismatches in CI." + "reason": "Weight-tying bugs are unrelated to dtype mismatch issues in bfloat16 inference." }, { "left": "issue:40990", "right": "issue:45584", "accept": false, - "reason": "High perplexity on gpt-oss-20b and Whisper empty-transcription failures are unrelated." + "reason": "High perplexity on GPT-OSS and Whisper empty-transcription failure are unrelated generation problems." }, { "left": "issue:43572", "right": "issue:45759", "accept": false, - "reason": "Missing pad_token_idx in StableLmConfig is unrelated to AutoModelForCausalLM.from_config text_config unwrapping." + "reason": "Missing pad_token_idx in StableLmConfig is unrelated to unwrapping text_config for Qwen multimodal configs." }, { "left": "issue:43377", "right": "issue:44898", "accept": false, - "reason": "MIMI batch-padding behavior and Perceiver non-default-resolution behavior do not share the same bug." + "reason": "Missing padding-mask support in MIMI encoder is unrelated to Perceiver positional encoding interpolation." }, { "left": "issue:39692", "right": "issue:40444", "accept": false, - "reason": "SigLIP2 doc-example errors are not the same as Qwen2.5-VL IterableDataset multi-image fine-tuning failures." + "reason": "SigLIP2 doc/example errors and Qwen2.5-VL iterable-dataset finetuning failures are different multimodal issues." }, { "left": "issue:44466", "right": "issue:45357", "accept": false, - "reason": "lm_head.weight serialization inconsistency and incorrect visual-encoder keys are different save_pretrained problems." + "reason": "lm_head.weight serialization inconsistency and incorrect visual encoder keys are both save_pretrained-related but not the same bug." }, { - "left": "issue:43296", - "right": "issue:44991", + "left": "issue:43792", + "right": "issue:44683", "accept": false, - "reason": "PaddleOCR-VL loading in vLLM and tokenizer loading for est-roberta are unrelated." + "reason": "Whisper runtime failure and compiled flex_attention failure are unrelated." }, { "left": "issue:43296", - "right": "issue:44877", + "right": "issue:44991", "accept": false, - "reason": "A vLLM model load failure is not the same as strict config blocking granite_speech config loading." + "reason": "PaddleOCR-VL load failure in vLLM is unrelated to tokenizer loading for est-roberta." }, { - "left": "issue:43792", - "right": "issue:44683", + "left": "issue:43296", + "right": "issue:44877", "accept": false, - "reason": "Whisper runtime failure and compiled flex_attention failing on torch>=2.9 are different issues." + "reason": "Strict config rejection for granite_speech is a different loading/config bug than PaddleOCR-VL model loading." }, { "left": "issue:43454", "right": "issue:43720", "accept": false, - "reason": "Tied lm_head weights and unpacking packed BitNet weights during accelerate loading are separate bugs." + "reason": "lm_head weight tying is unrelated to BitNet packed-weight unpacking during accelerate loading." }, { "left": "issue:43873", "right": "issue:45701", "accept": false, - "reason": "Offloading/quantization behavior is unrelated to version-dependent tokenization changes." + "reason": "Generic offloading/quantization issues do not match a tokenization-version regression." }, { "left": "issue:41720", "right": "issue:44977", "accept": false, - "reason": "Qwen3 auto device-map cuda asserts and Qwen3.5 flash-attention generation failures are different code paths." + "reason": "Both are Qwen-family runtime bugs, but auto device mapping cuda asserts and flash-attention generation failures are different code paths." }, { "left": "issue:45632", "right": "issue:45684", "accept": false, - "reason": "Cache-path collisions for local models and read-only permission propagation in save_pretrained are unrelated." + "reason": "Cache-path collisions for trust_remote_code are unrelated to permission propagation in save_pretrained." }, { "left": "issue:43404", "right": "issue:45357", "accept": false, - "reason": "Mistral3 lm_head tying and Qwen3.5 visual-encoder key serialization are different model-specific bugs." + "reason": "A missing tie_weights call and incorrect visual encoder keys are different serialization/model-init bugs." }, { "left": "issue:44423", "right": "issue:45406", "accept": false, - "reason": "Continuous-batching multimodal serve crashes and Gemma4Processor missing _tokenizer are distinct failures." + "reason": "Continuous-batching multimodal crash and Gemma4Processor missing _tokenizer are related to serve, but not the same concrete defect." }, { - "left": "issue:43632", - "right": "issue:45563", + "left": "issue:44186", + "right": "issue:45072", + "accept": false, + "reason": "Tokenizer padding/truncation crashes are unrelated to bfloat16 dtype mismatches." + }, + { + "left": "issue:43577", + "right": "issue:43708", + "accept": false, + "reason": "dtype propagation for BLIP2 is unrelated to Trainer resume max_steps calculation." + }, + { + "left": "issue:34634", + "right": "issue:35707", "accept": false, - "reason": "Transformers v5 initialization flag breakage and a stale paged-generate warning are unrelated." + "reason": "Bark voice_preset failure and progressive generation with inputs_embeds/past_key_values are unrelated." + }, + { + "left": "issue:35707", + "right": "issue:41720", + "accept": false, + "reason": "Progressive generation state handling is unrelated to Qwen3 auto-device-map cuda asserts." }, { "left": "issue:43873", "right": "issue:44756", "accept": false, - "reason": "Quantization/offloading complaints and disabling mmap on Strix Halo to avoid OOM are not the same defect." + "reason": "Generic offloading/quantization problems are not the same as disabling mmap to avoid OOM on Strix Halo." } ] }, @@ -3716,8 +3616,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -3784,17 +3684,17 @@ "issue:36331|issue:42617", "issue:43316|issue:44913", "issue:42915|issue:43317", - "issue:41720|issue:43322", "issue:40990|issue:45701", - "issue:43550|issue:43606", + "issue:41720|issue:43322", "issue:43295|issue:45701", + "issue:43550|issue:43606", "issue:41720|issue:45161", "issue:44671|issue:44743", "issue:43531|issue:43572", "issue:44292|issue:44479", "issue:41720|issue:42617", - "issue:44186|issue:44361", "issue:36010|issue:42617", + "issue:44186|issue:44361", "issue:43540|issue:43782", "issue:45406|issue:45464", "issue:41720|issue:44292", @@ -3805,238 +3705,238 @@ "issue:42915|issue:43322", "issue:43606|issue:43720" ], - "prepared_review_unit_hash": "0ae3c7367f13b2bf9837e92fb3cecb99bd43c72c7c1b8b30b4a99d24eddd666d", + "prepared_review_unit_hash": "8604d3d40b1a2b07cac2e63c172aeb3200af909eea5af0ad884b7905620ed748", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, "soft_pair_count": 34, - "serialized_chars": 36943, - "estimated_input_tokens": 9236, - "estimated_eval_tokens": 18728 + "serialized_chars": 36948, + "estimated_input_tokens": 9237, + "estimated_eval_tokens": 18730 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is a broad semantic bucket of unrelated Transformers issues, mostly around model loading, tokenizers, processors, and device/offload behavior. None of the soft pairs look like the same underlying bug/change, so I would not merge any of them.", - "confidence": 0.96, - "canonical_issue_reason": null, + "summary": "The soft pairs are only superficially similar; they span unrelated models, APIs, and failure modes. I would not merge any of these pairs as duplicates.", + "confidence": 0.94, + "canonical_issue_reason": "No clear canonical issue: this set is heterogeneous, with distinct bugs across loading, tokenization, multimodal processing, quantization, and serving paths.", "canonical_pr_reason": null, - "best_issue_reason": null, + "best_issue_reason": "No global best issue candidate; none of the items subsumes the others or represents the shared underlying defect.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43749", "right": "issue:44534", "accept": false, - "reason": "FSDP CPU/RAM-efficient loading vs non-persistent buffer corruption; different failure modes." + "reason": "FSDP CPU/RAM-efficient loading vs non-persistent buffer initialization are different code paths and symptoms." }, { "left": "issue:43526", "right": "issue:43881", "accept": false, - "reason": "BEiT label reduction bug vs GLM-4V loading failure; unrelated components." + "reason": "A BEiT image-processor label bug and a GLM-4V loading failure are unrelated model-specific issues." }, { "left": "issue:44877", "right": "issue:44991", "accept": false, - "reason": "Config strictness vs tokenizer-loading regression; same broad area, not same bug." + "reason": "Both involve loading regressions, but one is strict config handling and the other is tokenizer loading; not the same defect." }, { "left": "issue:42617", "right": "issue:42915", "accept": false, - "reason": "3d_parallel runtime failure vs Qwen3-MoE FineGrainedFP8 loading issue." + "reason": "Different failures: a 3d_parallel.py runtime problem versus a Qwen3MoE FP8 config issue." }, { "left": "issue:43450", "right": "issue:45701", "accept": false, - "reason": "Video batching shape bug vs tokenization-version regression; different subsystems." + "reason": "A batched video-shape bug is unrelated to a general tokenization-version regression." }, { "left": "issue:43550", "right": "issue:43720", "accept": false, - "reason": "SDPA torch.compile bug vs BitNet packed-weight unpacking bug; different code paths." + "reason": "These are different model/runtime bugs: torch.compile+SDPA for Bamba versus packed-weight loading for BitNet." }, { "left": "issue:36331", "right": "issue:42915", "accept": false, - "reason": "CustomTrainer argument mismatch vs Qwen3-MoE FP8 loading; unrelated." + "reason": "Trainer API keyword mismatch and Qwen3MoE FP8 loading are distinct problems." }, { "left": "issue:44610", "right": "issue:44898", "accept": false, - "reason": "OmDet input-size mismatch vs Perceiver resize/interpolation issue; different models." + "reason": "Both mention image sizes, but they affect different models and different resize/interpolation logic." }, { "left": "issue:44186", "right": "issue:44610", "accept": false, - "reason": "LayoutLMv2 tokenizer crash vs OmDet processor resolution mismatch; unrelated." + "reason": "LayoutLMv2 tokenizer crashes and OmDet-Turbo image-size mismatch are unrelated." }, { "left": "issue:44387", "right": "issue:45704", "accept": false, - "reason": "Int4 CUDA memory regression vs apex RMSNorm leak; different root causes." + "reason": "CUDA reserved-memory growth under int4 quantization is not the same as an apex RMSNorm memory leak." }, { "left": "issue:44479", "right": "issue:44918", "accept": false, - "reason": "Video-input regression vs embedding unpacking/TRL trainer issue; different paths." + "reason": "One is a video-input regression in Qwen VL models; the other is embedding unpacking with TRL SFT." }, { "left": "issue:36331", "right": "issue:42617", "accept": false, - "reason": "Trainer loss signature bug vs 3d_parallel runtime failure." + "reason": "Trainer compute_loss signature breakage is unrelated to the 3d_parallel.py failure." }, { "left": "issue:43316", "right": "issue:44913", "accept": false, - "reason": "Gemma3 config API mismatch vs GPTNeoX rotary_pct reload bug." + "reason": "Config API discrepancy in Gemma3TextConfig is not the same as GPTNeoX rotary_pct reload behavior." }, { "left": "issue:42915", "right": "issue:43317", "accept": false, - "reason": "Qwen3-MoE FP8 loading vs dequantized model offload loading; distinct loading issues." + "reason": "Both are loading-related, but one is FP8 Qwen3MoE and the other is dequantized model offload with device_map=auto." + }, + { + "left": "issue:40990", + "right": "issue:45701", + "accept": false, + "reason": "High perplexity on a model/dataset pair is not the same as a tokenizer behavior change." }, { "left": "issue:41720", "right": "issue:43322", "accept": false, - "reason": "Qwen3 auto-device-map CUDA assert vs Llava Next segfault; unrelated failures." + "reason": "Qwen3 auto device-map assertions and Llava Next loading segfaults are different failures." }, { - "left": "issue:40990", + "left": "issue:43295", "right": "issue:45701", "accept": false, - "reason": "Perplexity/evaluation issue vs tokenizer-version behavior change." + "reason": "Processor/tokenizer API regression and general tokenization-version changes are not the same code path." }, { "left": "issue:43550", "right": "issue:43606", "accept": false, - "reason": "Bamba compile/SDPA bug vs Bark CPU-offload device mismatch." - }, - { - "left": "issue:43295", - "right": "issue:45701", - "accept": false, - "reason": "Processor.tokenizer regression vs tokenization-version change; different concrete breakages." + "reason": "Different models and different offload/compile issues; only the broad 'device mismatch' theme overlaps." }, { "left": "issue:41720", "right": "issue:45161", "accept": false, - "reason": "Qwen3 auto-device-map assert vs GPT-OSS MoE tensor-parallel issue." + "reason": "Auto device mapping on Qwen3 is not the same as tensor-parallel GPT-OSS MoE behavior." }, { "left": "issue:44671", "right": "issue:44743", "accept": false, - "reason": "CamemBERT MLM output issue vs Qwen3 recurrent-state caching bug." + "reason": "CamemBERT MLM prediction error and Qwen3.5 recurrent-state reset are unrelated." }, { "left": "issue:43531", "right": "issue:43572", "accept": false, - "reason": "Sliding-window MoE issue vs missing pad_token_idx config field." + "reason": "Sliding-window behavior in Qwen3-MoE and missing pad_token_idx in StableLmConfig are different bugs." }, { "left": "issue:44292", "right": "issue:44479", "accept": false, - "reason": "Qwen-3 NVFP4 runtime error vs Qwen3.5 video-input regression." + "reason": "NVFP4 run failure and Qwen VL video regression are unrelated model-specific issues." }, { "left": "issue:41720", "right": "issue:42617", "accept": false, - "reason": "Auto device mapping assert vs 3d_parallel execution failure; different components." + "reason": "Auto device mapping assert and 3d_parallel.py runtime failure are distinct problems." }, { - "left": "issue:44186", - "right": "issue:44361", + "left": "issue:36010", + "right": "issue:42617", "accept": false, - "reason": "LayoutLMv2 tokenizer padding crash vs MLukeTokenizer task AttributeError." + "reason": "GenerationMixin import failure is unrelated to the 3d_parallel.py issue." }, { - "left": "issue:36010", - "right": "issue:42617", + "left": "issue:44186", + "right": "issue:44361", "accept": false, - "reason": "GenerationMixin import break vs 3d_parallel runtime failure; unrelated." + "reason": "Two different tokenizer bugs affecting different models and tasks." }, { "left": "issue:43540", "right": "issue:43782", "accept": false, - "reason": "Video-processing ValueError vs Qwen3VL weight-only loading error." + "reason": "Qwen3OmniMoe video-processing ValueError and Qwen3VL weight_only load error are different code paths." }, { "left": "issue:45406", "right": "issue:45464", "accept": false, - "reason": "Gemma4Processor missing _tokenizer vs Qwen3.5 streaming chat failure; different models and paths." + "reason": "Gemma4Processor missing _tokenizer and Qwen3.5 streaming inference failure are separate serving bugs." }, { "left": "issue:41720", "right": "issue:44292", "accept": false, - "reason": "Auto device-map CUDA assert vs NVFP4 runtime error; different quantization/device issues." + "reason": "Auto device mapping on Qwen3 and Qwen-3-8B-NVFP4 runtime failure are not the same defect." }, { "left": "issue:43421", "right": "issue:45701", "accept": false, - "reason": "Runtime special-token/post-processor update bug vs tokenizer-version behavior change." + "reason": "Runtime special-token post-processor updates and general tokenization-version changes overlap only loosely; not a duplicate." }, { "left": "issue:30064", "right": "issue:34634", "accept": false, - "reason": "Void segmentation-map processing vs Bark voice_preset issue; unrelated processors." + "reason": "Void segmentation-map processing and Bark voice_preset issues are unrelated." }, { "left": "issue:44492", "right": "issue:44509", "accept": false, - "reason": "Cache-strategy typo vs docs mentioning removed pipeline tasks." + "reason": "A cache-strategy typo and stale docs for removed pipeline tasks are different documentation issues." }, { "left": "issue:43388", "right": "issue:45059", "accept": false, - "reason": "Mask2Former gather_for_metrics bug vs SAM3 text/bbox behavior; different data paths." + "reason": "Metric label truncation and SAM3 text/box behavior are unrelated." }, { "left": "issue:42915", "right": "issue:43322", "accept": false, - "reason": "Qwen3-MoE FP8 loading vs Llava Next segmentation fault; unrelated failures." + "reason": "Qwen3MoE FP8 loading and Llava Next segfaults are different model-loading failures." }, { "left": "issue:43606", "right": "issue:43720", "accept": false, - "reason": "Bark CPU-offload device mismatch vs BitNet packed-weight unpacking bug." + "reason": "CPU offload device mismatch for Bark and packed-weight unpacking for BitNet are distinct bugs." } ] }, @@ -4055,8 +3955,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -4118,8 +4018,8 @@ "issue:34567|issue:35707", "issue:44368|issue:45357", "issue:44898|issue:45072", - "issue:43873|issue:45704", "issue:44451|issue:44991", + "issue:43873|issue:45704", "issue:41720|issue:43746", "issue:44062|issue:44451", "issue:43421|issue:43653", @@ -4133,8 +4033,8 @@ "issue:39692|issue:44610", "issue:44918|issue:44936", "issue:45081|issue:45356", - "issue:44625|issue:45479", "issue:43525|issue:43531", + "issue:44625|issue:45479", "issue:43782|issue:45381", "issue:43994|issue:44671", "issue:34634|issue:41720", @@ -4142,14 +4042,14 @@ "issue:43596|issue:44220", "issue:39692|issue:43873" ], - "prepared_review_unit_hash": "8cb002688b153384541072a1b40781a2eb83e31cd3bb28dfe75fa08ad92e3d13", + "prepared_review_unit_hash": "fa6969c618e3bee768fcbf33b464f80887d58bb27f95a72ed22e7c1411fa5b99", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, @@ -4164,204 +4064,204 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is mostly heterogeneous tokenizer/model/runtime bug reports with one clear duplicate pair and a few loose semantic similarities, so only the exact duplicate should be merged.", - "confidence": 0.71, - "canonical_issue_reason": "Issue 44514 is the earliest and clearest report of the exact Qwen2_5_VLProcessor.apply_chat_template batched-input/padding=False crash, making it the best canonical issue for the confirmed duplicate pair.", + "summary": "The cluster is mostly heterogeneous; only the two Qwen2_5_VLProcessor `apply_chat_template` reports are the same underlying bug. All other suggested pairs appear to be unrelated issues in different models/code paths.", + "confidence": 0.94, + "canonical_issue_reason": "issue:44514 is the earlier, more specific report of the Qwen2_5_VLProcessor batched `apply_chat_template` crash; issue:44545 looks like the later duplicate of the same failure.", "canonical_pr_reason": null, - "best_issue_reason": "44514 is the best representative because it precisely captures the concrete failure mode and is directly duplicated by 44545.", + "best_issue_reason": "issue:44514 is the clearest representative of the only true duplicate pair in this set.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:44514", "right": "issue:44545", "accept": true, - "reason": "Same Qwen2_5_VLProcessor.apply_chat_template crash on batched input with padding=False; identical underlying bug." + "reason": "Same component, same batched `apply_chat_template` crash when `padding=False`; titles describe the same bug." }, { "left": "issue:43550", "right": "issue:43653", "accept": false, - "reason": "Different subsystems and failures: torch.compile+SDPA on Bamba vs tokenizer special-token registration." + "reason": "Different subsystems: torch.compile/SDPA model bug vs tokenizer special-token registration." }, { "left": "issue:43531", "right": "issue:45759", "accept": false, - "reason": "Both involve Qwen3-family configs, but one is sliding_window behavior and the other is text_config unwrapping; not the same bug." + "reason": "Qwen3-MoE sliding-window behavior is unrelated to composite Qwen multimodal config unwrapping." }, { "left": "issue:43606", "right": "issue:43653", "accept": false, - "reason": "CPU offload device mismatch is unrelated to BigBirdTokenizer mask-token decode output." + "reason": "CPU offload device mismatch for Bark is unrelated to BigBirdTokenizer special-token decoding." }, { "left": "issue:43421", "right": "issue:43927", "accept": false, - "reason": "Both touch token/special-token handling, but one is runtime post-processor updates and the other is save/load of custom token IDs." + "reason": "Both touch tokenizer state, but one is runtime post-processor refresh and the other is config persistence/load; not the same bug." }, { "left": "issue:34567", "right": "issue:35707", "accept": false, - "reason": "Trainer token-count tracking and progressive generation with inputs_embeds/past_key_values are unrelated." + "reason": "Trainer token counting and progressive generation with embeddings/past key values are unrelated failures." }, { "left": "issue:44368", "right": "issue:45357", "accept": false, - "reason": "A warning about tie_word_embeddings during LoRA fine-tuning is unrelated to incorrect visual encoder keys on save_pretrained." + "reason": "A warning about `tie_word_embeddings` is not the same as incorrect visual-encoder keys being saved." }, { "left": "issue:44898", "right": "issue:45072", "accept": false, - "reason": "Perceiver interpolation bug and bfloat16 dtype mismatch in SwitchTransformers/TimmWrapperModel are different issues." + "reason": "Perceiver resolution handling and bfloat16 dtype mismatch in different models are unrelated." }, { - "left": "issue:43873", - "right": "issue:45704", + "left": "issue:44451", + "right": "issue:44991", "accept": false, - "reason": "Quantization/offloading expectations do not match a T5 RMSNorm memory-leak issue." + "reason": "Both are tokenizer-loading regressions, but they affect different models and likely different code paths." }, { - "left": "issue:44451", - "right": "issue:44991", + "left": "issue:43873", + "right": "issue:45704", "accept": false, - "reason": "Both are loading regressions, but for different models and likely different tokenizer code paths." + "reason": "Quantization/offloading behavior is unrelated to a T5 apex RMSNorm memory leak." }, { "left": "issue:41720", "right": "issue:43746", "accept": false, - "reason": "Qwen3 auto device mapping cuda assert is unrelated to GraniteSpeech PEFT checkpoint loading." + "reason": "Qwen3 auto device-map CUDA assert is unrelated to GraniteSpeech PEFT checkpoint loading." }, { "left": "issue:44062", "right": "issue:44451", "accept": false, - "reason": "AddedToken constructor argument clash is a different failure from loading ScandiBERT." + "reason": "AddedToken construction error is unrelated to loading ScandiBERT." }, { "left": "issue:43421", "right": "issue:43653", "accept": false, - "reason": "Runtime special-token post-processor updates are unrelated to BigBirdTokenizer empty decode output." + "reason": "Special-token post-processing updates are different from BigBirdTokenizer mask-token registration/decoding." }, { "left": "issue:42886", "right": "issue:44843", "accept": false, - "reason": "Both involve HF_HUB_OFFLINE, but 44843 is a specific _patch_mistral_regex/model_info regression while 42886 is a broader cache-load failure." + "reason": "Both involve offline mode, but one is generic cache loading while the other is a specific `_patch_mistral_regex` hub call." }, { "left": "issue:44987", "right": "issue:45042", "accept": false, - "reason": "Loading physical-intelligence/fast and PIL-backend torchvision requirements are not the same bug." + "reason": "Loading a specific model in 5.1.0 is unrelated to PIL image processors incorrectly requiring torchvision." }, { "left": "issue:45588", "right": "issue:45759", "accept": false, - "reason": "Flash-attention aux-state crash and Qwen multimodal text_config unwrapping are unrelated." + "reason": "Flash-attention sinkless model crash is unrelated to Qwen multimodal config unwrapping." }, { "left": "issue:33666", "right": "issue:35707", "accept": false, - "reason": "Multi-GPU training for Qwen2-VL and progressive generation with inputs_embeds are unrelated topics." + "reason": "Qwen2-VL multi-GPU training and progressive generation are different problems." }, { "left": "issue:43531", "right": "issue:44589", "accept": false, - "reason": "Qwen3-MoE sliding_window behavior is unrelated to the missing Float8 storage type error." + "reason": "Qwen3-MoE sliding-window behavior is unrelated to the Float8 storage lookup error." }, { "left": "issue:43421", "right": "issue:44186", "accept": false, - "reason": "TokenizersBackend special-token updates and LayoutLMv2 tokenizer NER/padding crashes are different code paths." + "reason": "Runtime special-token updates are not the same as LayoutLMv2Tokenizer NER/padding crashes." }, { "left": "issue:43381", "right": "issue:43844", "accept": false, - "reason": "Eval-mode gradient checkpointing limitation is unrelated to gradient blow-up under HfDeepSpeedConfig + ZeRO-3." + "reason": "Eval-mode gradient checkpointing restriction is unrelated to abnormal gradient growth under ZeRO-3." }, { "left": "issue:39692", "right": "issue:44610", "accept": false, - "reason": "SigLIP2 doc/example errors and OmDet-Turbo processor size mismatch are different model-specific issues." + "reason": "Both mention processor/model mismatches, but they involve different models and distinct preprocessing bugs." }, { "left": "issue:44918", "right": "issue:44936", "accept": false, - "reason": "TRL SFT unpacking of Qwen3.5 embeddings is unrelated to trainer.evaluate failing after trainer.train()." + "reason": "TRL SFT embedding unpacking and trainer evaluate-after-train failures are different training issues." }, { "left": "issue:45081", "right": "issue:45356", "accept": false, - "reason": "Both mention fix_mistral_regex/tokenizer regression, but one is a hard crash in Mistral loading and the other is a Kimi-K2.5 codec-handling regression with a warning issue; not clearly the same bug." + "reason": "Both reference mistral-regex-related tokenization, but one is an AttributeError crash and the other is a codec-handling regression." }, { - "left": "issue:44625", - "right": "issue:45479", + "left": "issue:43525", + "right": "issue:43531", "accept": false, - "reason": "Missing num_labels propagation in Qwen3.5 configs is unrelated to zero-loss single-label classification behavior across models." + "reason": "Missing `pad_token_id` on Llama4Config is unrelated to Qwen3-MoE sliding-window behavior." }, { - "left": "issue:43525", - "right": "issue:43531", + "left": "issue:44625", + "right": "issue:45479", "accept": false, - "reason": "Llama4Config pad_token_id missing and Qwen3-MoE sliding_window are different config bugs." + "reason": "Qwen3.5 config propagation and degenerate single-label classification loss are separate model-classification issues." }, { "left": "issue:43782", "right": "issue:45381", "accept": false, - "reason": "Qwen3VL weight_only loading and qwen2.5-vl video vision_position_ids are different multimodal bugs." + "reason": "Qwen3VL weight-only loading failure is unrelated to Qwen2.5-VL video vision-position IDs." }, { "left": "issue:43994", "right": "issue:44671", "accept": false, - "reason": "SigLIP2 AutoModel/pipeline output mismatch and CamemBERT masked LM predictions are unrelated." + "reason": "Incorrect SigLIP2 outputs and CamemBERT MLM regression are different model-specific failures." }, { "left": "issue:34634", "right": "issue:41720", "accept": false, - "reason": "BarkProcessor voice_preset and Qwen3 auto device mapping failures are unrelated." + "reason": "BarkProcessor voice preset handling is unrelated to Qwen3 auto device-mapping CUDA asserts." }, { "left": "issue:43381", "right": "issue:45704", "accept": false, - "reason": "Gradient-checkpointing eval-mode behavior is unrelated to T5 apex.FusedRMSNorm memory leak." + "reason": "Gradient checkpointing in eval mode is unrelated to T5\u2019s apex RMSNorm leak." }, { "left": "issue:43596", "right": "issue:44220", "accept": false, - "reason": "DeepSpeed ZeRO-3/BertModel index error and _torch_extract_fbank_features() are different failures." + "reason": "DeepSpeed ZeRO-3 init with BertModel and `_torch_extract_fbank_features()` are unrelated errors." }, { "left": "issue:39692", "right": "issue:43873", "accept": false, - "reason": "SigLIP2 documentation/example errors are unrelated to quantization offloading behavior." + "reason": "SigLIP2 doc-example errors are unrelated to quantization/offloading behavior." } ] }, "evaluator_result": { "accept": true, - "feedback": "The summary is grounded in the packet and stays conservative: it identifies only the clearly duplicated Qwen2_5_VLProcessor.apply_chat_template batched-input/padding=False pair (44514/44545) and rejects the rest as distinct or only loosely similar. No overstatement or mergeability issue is apparent." + "feedback": "The summary is grounded in the packet: the 44514/44545 pair is the only clearly duplicated issue based on the near-identical titles, and the other pair rejections stay conservative rather than overstating sameness." }, "evaluator_used": true, "retried": false, @@ -4377,8 +4277,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -4436,10 +4336,10 @@ "issue:45183|issue:45479", "issue:43618|issue:43697", "issue:43540|issue:44008", - "issue:44077|issue:44625", "issue:43720|issue:43819", - "issue:43575|issue:45704", + "issue:44077|issue:44625", "issue:30064|issue:36010", + "issue:43575|issue:45704", "issue:44756|issue:45759", "issue:43716|issue:44008", "issue:43335|issue:43756", @@ -4450,8 +4350,8 @@ "issue:44843|issue:45081", "issue:43550|issue:45072", "issue:44186|issue:44898", - "issue:43421|issue:44811", "issue:43550|issue:43819", + "issue:43421|issue:44811", "issue:43577|issue:43818", "issue:43643|issue:44877", "issue:44062|issue:44488", @@ -4462,208 +4362,208 @@ "issue:43782|issue:44977", "issue:44561|issue:45020" ], - "prepared_review_unit_hash": "efbfab781e704ddb17df5a177a2a93e46ff461c6a346818669fb1889e40e35c3", + "prepared_review_unit_hash": "d816910637bf2f00fcef59a0745aa91171c2dedc220b02ad3503082186406b1e", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, "soft_pair_count": 30, - "serialized_chars": 36368, - "estimated_input_tokens": 9092, - "estimated_eval_tokens": 18440 + "serialized_chars": 36372, + "estimated_input_tokens": 9093, + "estimated_eval_tokens": 18442 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "These issues are mostly unrelated model- or subsystem-specific bugs. A few are loosely adjacent (notably around tokenizer/Mistral helpers and remote_code regressions), but none of the soft pairs look like true duplicates of the same concrete bug/change.", + "summary": "The set is a mix of unrelated model-, tokenizer-, config-, and runtime-specific bugs. None of the soft-pair candidates look like true duplicates or the same fixable code-path problem, and there are no PRs in the input.", "confidence": 0.86, - "canonical_issue_reason": "If a canonical issue is needed, issue:45020 is the broadest umbrella-style report in the set (recent versions breaking remote_code models). It is not an exact duplicate of the others, but it is the closest general anchor for the cluster\u2019s theme of version-related regressions.", + "canonical_issue_reason": "issue:43873 is the broadest, most active open report here, with multiple inbound references and discussion around a runtime behavior regression.", "canonical_pr_reason": null, - "best_issue_reason": "Issue:45020 is the best representative issue because it is open, broadly scoped, and captures a recurring compatibility/regression class rather than a narrow model-specific symptom.", + "best_issue_reason": "issue:43873 is the strongest triage anchor because it is active, broadly scoped, and already attracting follow-up references.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:42915", "right": "issue:45759", "accept": false, - "reason": "Both mention Qwen configs, but one is FineGrainedFP8Config failure and the other is missing text_config unwrapping in composite multimodal configs; different code paths and fixes." + "reason": "Different problems: Qwen3MoE FP8 config failure vs. AutoModelForCausalLM config unwrapping for composite Qwen multimodal configs." }, { "left": "issue:44220", "right": "issue:44661", "accept": false, - "reason": "Different areas entirely: audio feature extraction vs tokenizer mapping/model registration logic." + "reason": "Different code paths: audio feature extraction failure vs. tokenizer-model mapping logic in add-new-model-like." }, { "left": "issue:45183", "right": "issue:45479", "accept": false, - "reason": "One is input validation/XOR error messaging, the other is a loss-function/label semantics bug; not the same defect." + "reason": "One is input validation/error messaging, the other is a classification loss bug with num_labels=1." }, { "left": "issue:43618", "right": "issue:43697", "accept": false, - "reason": "CLIP output attentions regression and RTDetrV2 output drift are unrelated model bugs." + "reason": "Unrelated regressions in different models: CLIP output attentions vs. RTDetrV2 output differences in v5." }, { "left": "issue:43540", "right": "issue:44008", "accept": false, - "reason": "Video input handling in Qwen3OmniMoe is unrelated to the Gemma3n variable-name collision in forward()." - }, - { - "left": "issue:44077", - "right": "issue:44625", - "accept": false, - "reason": "PatchTSMixer post_init policy and Qwen3.5 num_labels propagation are unrelated configuration issues." + "reason": "Different multimodal failures in different models; no shared concrete bug or fix path." }, { "left": "issue:43720", "right": "issue:43819", "accept": false, - "reason": "Packed-weight loading in BitNet and DAC latent/STE mismatch are different concrete bugs." + "reason": "Different mechanisms: accelerate loading/unpacking packed weights vs. DAC latent conversion/STE mismatch." }, { - "left": "issue:43575", - "right": "issue:45704", + "left": "issue:44077", + "right": "issue:44625", "accept": false, - "reason": "TP OOM when loading a Qwen model and T5 apex RMSNorm memory leak are distinct memory problems in different paths." + "reason": "One is a model post_init restriction, the other is config propagation for Qwen3.5 labels." }, { "left": "issue:30064", "right": "issue:36010", "accept": false, - "reason": "Image segmentation map processing and a GenerationMixin import error are unrelated." + "reason": "Completely unrelated: image processor segmentation-map handling vs. an import error for GenerationMixin." + }, + { + "left": "issue:43575", + "right": "issue:45704", + "accept": false, + "reason": "Different issues: tensor-parallel OOM on Qwen2-57B vs. T5 using a leaky apex RMSNorm." }, { "left": "issue:44756", "right": "issue:45759", "accept": false, - "reason": "Strix Halo mmap/OOM behavior is unrelated to Qwen multimodal config unwrapping." + "reason": "Different scope: mmap/OOM on Strix Halo vs. config unwrapping for Qwen multimodal models." }, { "left": "issue:43716", "right": "issue:44008", "accept": false, - "reason": "Mistral image-preprocessor dtype mismatch and Gemma3n tensor attribute collision are unrelated." + "reason": "Different models and failures: Mistral-3 dtype mismatch vs. Gemma 3n variable name collision." }, { "left": "issue:43335", "right": "issue:43756", "accept": false, - "reason": "Sparse encoder layer creation and Smollm3 RoPE layer dropping are different model-configuration mistakes." + "reason": "Unrelated model architecture bugs: sparse layer creation vs. RoPE layer retention." }, { "left": "issue:43653", "right": "issue:43720", "accept": false, - "reason": "Tokenizer special-token registration and BitNet packed-weight loading do not share the same underlying bug." + "reason": "Tokenizer special-token registration bug vs. BitNet packed-weight loading bug." }, { "left": "issue:42915", "right": "issue:43746", "accept": false, - "reason": "Fine-grained FP8 config failure and GraniteSpeech PEFT checkpoint loading are separate issues." + "reason": "Different models and failure modes: Qwen3MoE FP8 config vs. GraniteSpeech PEFT checkpoint loading." }, { "left": "issue:43746", "right": "issue:45381", "accept": false, - "reason": "PEFT local checkpoint loading and Qwen2.5-VL video position ids are unrelated." + "reason": "Checkpoint loading with PEFT adapters is unrelated to Qwen2.5-VL video vision_position_ids." }, { "left": "issue:43746", "right": "issue:43873", "accept": false, - "reason": "GraniteSpeech checkpoint loading and quantization offloading behavior are different code paths." + "reason": "Unrelated: GraniteSpeech checkpoint loading vs. quantization/offloading behavior." }, { "left": "issue:44843", "right": "issue:45081", "accept": false, - "reason": "Both involve _patch_mistral_regex, but one is offline model_info access and the other is an AttributeError on backend_tokenizer; same helper, different failures and fixes." + "reason": "Both touch _patch_mistral_regex, but they are different bugs: offline model_info call vs. backend_tokenizer AttributeError during regex fixing." }, { "left": "issue:43550", "right": "issue:45072", "accept": false, - "reason": "Bamba torch.compile+SDPA and SwitchTransformers/TimmWrapper dtype mismatches are unrelated." + "reason": "Different subsystems and symptoms: torch.compile+SDPA failure vs. dtype mismatches in bfloat16 inference." }, { "left": "issue:44186", "right": "issue:44898", "accept": false, - "reason": "LayoutLMv2 tokenizer crashes and Perceiver interpolation failure are different model-specific bugs." + "reason": "Distinct preprocessing/tokenization bug vs. image classification failure at non-default resolution." }, { - "left": "issue:43421", - "right": "issue:44811", + "left": "issue:43550", + "right": "issue:43819", "accept": false, - "reason": "Runtime special-token/post-processor updates and Whisper batch_decode skip_special_tokens handling are related to tokenization, but not the same concrete bug." + "reason": "Different model/runtime issues: compile/SDPA failure vs. DAC latent/forward mismatch." }, { - "left": "issue:43550", - "right": "issue:43819", + "left": "issue:43421", + "right": "issue:44811", "accept": false, - "reason": "torch.compile SDPA failure and DAC forward/from_latents mismatch are unrelated." + "reason": "Post-processor updating for special tokens is not the same as Whisper batch_decode skip_special_tokens handling." }, { "left": "issue:43577", "right": "issue:43818", "accept": false, - "reason": "BLIP2 dtype propagation and Video-LLaVA temporal-attention/weight-sharing issues are unrelated." + "reason": "Different models and bugs: BLIP2 dtype preservation vs. Video-LLaVA missing temporal attention/weight sharing." }, { "left": "issue:43643", "right": "issue:44877", "accept": false, - "reason": "trust_remote_code missing fields and strict granite_speech config loading are distinct configuration-loading problems." + "reason": "Both involve config loading, but one is missing fields from trust_remote_code AutoConfig and the other is strict granite_speech config rejection." }, { "left": "issue:44062", "right": "issue:44488", "accept": false, - "reason": "AddedToken special-argument error and sleng-bert model loading failure are not the same bug." + "reason": "Different tokenizer/loading failures: AddedToken keyword collision vs. a model load regression for cjvt/sleng-bert." }, { "left": "issue:43531", "right": "issue:43643", "accept": false, - "reason": "Qwen3-MoE sliding_window behavior and trust_remote_code config field loss are unrelated." + "reason": "Qwen3-MoE sliding_window behavior is unrelated to AutoConfig trust_remote_code field loss." }, { "left": "issue:44448", "right": "issue:44671", "accept": false, - "reason": "Both are v5 output regressions, but they affect different models and likely different changes." + "reason": "Both are v5 regressions, but they affect different models and concrete outputs; not the same underlying bug." }, { "left": "issue:43653", "right": "issue:43819", "accept": false, - "reason": "Tokenizer special-token registration and DAC latent conversion mismatch do not share a concrete underlying defect." + "reason": "Tokenizer special-token registration and DAC STE mismatch are unrelated." }, { "left": "issue:43782", "right": "issue:44977", "accept": false, - "reason": "Qwen3VL weight_only loading failure and Qwen3.5 flash-attention generation issues are separate multimodal/runtime problems." + "reason": "Different Qwen variants and failure modes: from_pretrained weight_only error vs. flash-attention generation bug." }, { "left": "issue:44561", "right": "issue:45020", "accept": false, - "reason": "Removal of is_torch_fx_available is one specific remote_code breakage, while 45020 is a broad umbrella report that likely covers multiple unrelated failures." + "reason": "44561 is specifically the removal of is_torch_fx_available breaking remote_code models; 45020 is a broader report about recent versions breaking remote_code models, not the same concrete bug." } ] }, @@ -4682,8 +4582,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -4737,8 +4637,8 @@ "issue:45701" ], "soft_pairs": [ - "issue:43322|issue:45685", "issue:43825|issue:43994", + "issue:43322|issue:45685", "issue:44821|issue:45084", "issue:43825|issue:43827", "issue:30064|issue:36331", @@ -4761,24 +4661,24 @@ "issue:43262|issue:43819", "issue:43526|issue:43697", "issue:45084|issue:45538", - "issue:43644|issue:43749", "issue:43329|issue:45600", + "issue:43644|issue:43749", "issue:43653|issue:44912" ], - "prepared_review_unit_hash": "e613805d1a505ef1b774c3671c3e6432f2501f054f9c88849c7066afb63a15ce", + "prepared_review_unit_hash": "fbff41e56a679b16bda0d3690fce5af578a8dee00dba3fdef0be3728830f6735", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 27, - "serialized_chars": 35991, + "serialized_chars": 35989, "estimated_input_tokens": 8998, "estimated_eval_tokens": 18252 }, @@ -4787,174 +4687,174 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This is a very noisy, mostly false-positive similarity cluster. The items span many unrelated bugs across processors, tokenizers, loading, quantization, training, and docs; none of the soft pairs look like the same underlying defect.", - "confidence": 0.97, - "canonical_issue_reason": "Issue 43825 is the clearest representative of the only coherent subtheme in the set: v5 pipeline-removal behavior/error messaging. It is concrete and has inbound references, while most other items are unrelated point bugs.", + "summary": "This cluster is heterogeneous: the items span unrelated bugs in processors, tokenizers, training, loading, and docs. I found no convincing duplicate issue/PR pairings.", + "confidence": 0.9, + "canonical_issue_reason": "No single issue is a good canonical representative because the cluster does not center on one underlying bug; the titles point to many distinct code paths and models.", "canonical_pr_reason": null, - "best_issue_reason": "43825 is the best single issue to anchor this cluster, but only as a loose representative of the v5 pipeline-removal/docs subgroup; the overall bucket is too heterogeneous for a true canonical duplicate.", + "best_issue_reason": "No issue is a strong cluster anchor here. The most discussed items are still about different failures, so using any one as canonical would be misleading.", "best_pr_reason": null, "soft_edge_verdicts": [ { - "left": "issue:43322", - "right": "issue:45685", + "left": "issue:43825", + "right": "issue:43994", "accept": false, - "reason": "Different subsystems and failure modes: Llava Next load-time segfault vs an MPS histogram op unsupported for Int in moe." + "reason": "Both mention pipeline behavior, but one is about a wrong v5 error message and the other about SigLIP2 producing bad outputs; different bugs." }, { - "left": "issue:43825", - "right": "issue:43994", + "left": "issue:43322", + "right": "issue:45685", "accept": false, - "reason": "One is a v5 pipeline error-message/docs issue; the other is incorrect SigLIP2 outputs. Same broad area, but not the same bug." + "reason": "Different subsystems and failure modes: Llava Next loading segfault vs an MPS histogram/int dtype issue." }, { "left": "issue:44821", "right": "issue:45084", "accept": false, - "reason": "AutoImageProcessor URL loading and template-node compilation are unrelated code paths." + "reason": "Unrelated: AutoImageProcessor URL loading vs a template compilation TypeError." }, { "left": "issue:43825", "right": "issue:43827", "accept": false, - "reason": "Both mention removed pipeline tasks, but one is runtime error messaging and the other is documentation cleanup." + "reason": "Closely related theme (v5 pipeline removals), but one is runtime error messaging and the other is docs cleanup; not the same concrete bug." }, { "left": "issue:30064", "right": "issue:36331", "accept": false, - "reason": "ImageProcessor void segmentation maps and CustomTrainer compute_loss kwargs are unrelated issues." + "reason": "Different areas: image segmentation-map processing vs CustomTrainer compute_loss signature mismatch." }, { "left": "issue:44485", "right": "issue:44509", "accept": false, - "reason": "GLM-5 RoPE implementation discussion is unrelated to pipeline-removal documentation." + "reason": "GLM-5 RoPE implementation discussion is unrelated to pipeline docs references." }, { "left": "issue:43577", "right": "issue:43582", "accept": false, - "reason": "Blip2 dtype loading bug and AppleSilicon allocator TypeError are separate problems." + "reason": "Blip2 dtype loading bug vs Apple Silicon allocator warmup TypeError are unrelated." }, { "left": "issue:44186", "right": "issue:44442", "accept": false, - "reason": "Two tokenizer bugs, but with different models and failure mechanisms." + "reason": "Both are tokenizer bugs, but they affect different tokenizers with different behaviors and fixes." }, { "left": "issue:43606", "right": "issue:45072", "accept": false, - "reason": "CPU-offload device mismatch and bfloat16 dtype mismatch are different model-loading/runtime defects." + "reason": "Both involve device/dtype issues, but they are separate models and separate code paths." }, { "left": "issue:43452", "right": "issue:44991", "accept": false, - "reason": "gguf_file loading breakage and est-roberta tokenizer loading failure are not the same underlying issue." + "reason": "Tokenizer/model loading regressions, but not the same concrete failure; one is gguf_file handling, the other is a specific tokenizer load breakage." }, { "left": "issue:34634", "right": "issue:36010", "accept": false, - "reason": "BarkProcessor voice_preset bug and GenerationMixin import error are unrelated." + "reason": "BarkProcessor voice_preset bug is unrelated to the GenerationMixin import error." }, { "left": "issue:44805", "right": "issue:44936", "accept": false, - "reason": "Mask-shape IndexError and trainer.evaluate() failure after train() do not share a concrete root cause." + "reason": "Mask shape IndexError and trainer.evaluate() after train() are different problems." }, { "left": "issue:42175", "right": "issue:45081", "accept": false, - "reason": "Packaging/backend dependency issue and a Mistral tokenizer regex patch crash are unrelated." + "reason": "Package/backend install issue vs Mistral regex patch AttributeError; no shared underlying bug." }, { "left": "issue:44610", "right": "issue:45072", "accept": false, - "reason": "OmDet-Turbo input-size mismatch and dtype mismatches are different bug classes." + "reason": "OmDet-Turbo input-size mismatch is unrelated to dtype mismatch inference issues." }, { "left": "issue:45081", "right": "issue:45406", "accept": false, - "reason": "Both involve tokenizer/processor internals, but one is a Mistral regex patch crash and the other is a Gemma4Processor attribute error." + "reason": "Both are AttributeErrors in tokenizer/processor code, but on different attributes and different components." }, { "left": "issue:44060", "right": "issue:44368", "accept": false, - "reason": "Both mention tied-weight warnings, but they refer to different models and different warning logic." + "reason": "Related to Qwen tied-weights warnings, but the reported failures are not the same concrete bug or fix." }, { "left": "issue:43650", "right": "issue:43906", "accept": false, - "reason": "'ADD THE DATA' is not a concrete duplicate candidate for the isolated reproduction issue." + "reason": "Placeholder issue title vs an isolated reproduction of a different issue; no substantive overlap." }, { "left": "issue:43720", "right": "issue:45072", "accept": false, - "reason": "Packed-weight loading failure in BitNet and bfloat16 dtype mismatch are unrelated." + "reason": "Packed-weight loading bug vs bfloat16 dtype mismatch; unrelated." }, { "left": "issue:44671", "right": "issue:45701", "accept": false, - "reason": "CamemBERT MLM regression and version-dependent tokenization changes are not the same bug." + "reason": "CamemBERT MLM regression in v5 is not the same as the broad statement that version changes tokenization." }, { "left": "issue:43404", "right": "issue:43782", "accept": false, - "reason": "Mistral3 tied-weight bug and Qwen3VL weight_only loading error are separate multimodal loading issues." + "reason": "Different model families and different loading/tied-weight failures." }, { "left": "issue:43582", "right": "issue:44857", "accept": false, - "reason": "AppleSilicon allocator TypeError and CUDA AMP loss crash are different platform-specific problems." + "reason": "Apple Silicon allocator warmup TypeError is unrelated to a CUDA AMP loss crash." }, { "left": "issue:43262", "right": "issue:43819", "accept": false, - "reason": "Audio chat-template sampling-rate defaulting and DAC from_latents/STE mismatch are unrelated audio bugs." + "reason": "Audio chat-template sampling-rate default bug vs DAC.from_latents mismatch; different functionality." }, { "left": "issue:43526", "right": "issue:43697", "accept": false, - "reason": "BeitImageProcessorFast label reduction and RTDetrV2 output differences are distinct vision bugs." + "reason": "Different models and different bugs: label reduction in BeitImageProcessorFast vs RTDetrV2 output drift." }, { "left": "issue:45084", "right": "issue:45538", "accept": false, - "reason": "Template compilation TypeError and CLIPTokenizer model_max_length behavior are unrelated." + "reason": "Template compilation error is unrelated to CLIPTokenizer model_max_length behavior." }, { - "left": "issue:43644", - "right": "issue:43749", + "left": "issue:43329", + "right": "issue:45600", "accept": false, - "reason": "Non-persistent buffer initialization junk and FSDP CPU RAM-efficient loading are different low-level loading bugs." + "reason": "Multimodal token counting bug vs auto_mappings referencing removed configs; no shared underlying defect." }, { - "left": "issue:43329", - "right": "issue:45600", + "left": "issue:43644", + "right": "issue:43749", "accept": false, - "reason": "Undefined variables in multimodal token counting and stale auto_mappings references are unrelated." + "reason": "Non-persistent buffer junk in v5 and broken FSDP CPU RAM efficient loading are different loading problems." }, { "left": "issue:43653", "right": "issue:44912", "accept": false, - "reason": "BigBirdTokenizer special-token registration and MXFP4 quantization loading are different issues." + "reason": "BigBird special-token decoding bug is unrelated to MXFP4 quantization fallback." } ] }, @@ -4973,8 +4873,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -5032,13 +4932,13 @@ "issue:43746|issue:44811", "issue:38175|issue:44610", "issue:44589|issue:45042", - "issue:43854|issue:44977", "issue:43454|issue:43577", + "issue:43854|issue:44977", "issue:43526|issue:44554", "issue:36010|issue:43645", "issue:44568|issue:45488", - "issue:39692|issue:44912", "issue:34634|issue:36331", + "issue:39692|issue:44912", "issue:43257|issue:43329", "issue:43819|issue:43927", "issue:43720|issue:44610", @@ -5047,8 +4947,8 @@ "issue:43262|issue:44811", "issue:44568|issue:45356", "issue:43531|issue:45042", - "issue:36010|issue:36331", "issue:43643|issue:45759", + "issue:36010|issue:36331", "issue:34567|issue:40990", "issue:43329|issue:43450", "issue:43329|issue:44479", @@ -5056,210 +4956,217 @@ "issue:43257|issue:43931", "issue:43697|issue:45357", "issue:43452|issue:43582", - "issue:41720|issue:43540" + "issue:41720|issue:43540", + "issue:44291|issue:45042" ], - "prepared_review_unit_hash": "efa187047fa892e688b66e665d351be5e263451715dcf1968e82788c51be2033", + "prepared_review_unit_hash": "dba4fec7fd2248aaac6fba3f39fc3f599bb0fd097a37cdcb185898679487d297", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, - "soft_pair_count": 29, - "serialized_chars": 36755, - "estimated_input_tokens": 9189, - "estimated_eval_tokens": 18634 + "soft_pair_count": 30, + "serialized_chars": 37051, + "estimated_input_tokens": 9263, + "estimated_eval_tokens": 18782 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "The items form a loose collection of unrelated Transformers bugs. The suggested soft pairs share broad subsystems at best, but not the same concrete root cause, so none should be merged as duplicates.", - "confidence": 0.91, - "canonical_issue_reason": "issue:43749 is the strongest representative of the cluster because it is a broad, high-impact loading regression with multiple references and discussion, making it the best anchor issue if one must be chosen.", + "summary": "This cluster is a grab bag of mostly unrelated Transformers regressions. The soft pairs share broad themes like loading, tokenization, video, or multimodal processing, but they do not look like the same concrete bug/change.", + "confidence": 0.88, + "canonical_issue_reason": "issue:43749 is the broadest, highest-impact loading/initialization regression in the set and has the strongest discussion/reference signal; the rest are mostly narrower model- or feature-specific bugs.", "canonical_pr_reason": null, - "best_issue_reason": "issue:43749 is the most central and widely impacted bug report in the set; it is the best issue to keep as the representative anchor, though it does not subsume the others.", + "best_issue_reason": "issue:43749 best represents the cluster because it is a central infrastructure-loading failure rather than a model-specific edge case.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43638", "right": "issue:44936", "accept": false, - "reason": "Both mention trainer behavior, but one is a zero-size index error under ZeRO3 with a non-pretrained model and the other is evaluate() failing after train(); different failure modes and code paths." + "reason": "Different trainer failures: Deepspeed ZeRO-3 index error vs trainer.evaluate() after train()." }, { "left": "issue:43746", "right": "issue:44811", "accept": false, - "reason": "Local checkpoint loading with PEFT adapters and Whisper batch_decode skip_special_tokens are unrelated features and bugs." + "reason": "Different code paths: PEFT/local checkpoint loading vs Whisper batch_decode skip_special_tokens handling." }, { "left": "issue:38175", "right": "issue:44610", "accept": false, - "reason": "SigLIP2 zero probabilities and OmDet-Turbo processor/model size mismatch are different model-specific problems." + "reason": "Different models and bugs: SigLIP2 zero probabilities vs OmDet-Turbo processor size mismatch." }, { "left": "issue:44589", "right": "issue:45042", "accept": false, - "reason": "Float8 storage lookup failure and PIL backend requiring torchvision are separate runtime dependency issues." + "reason": "Unrelated failures: Float8 storage lookup vs PIL image processor torchvision dependency regression." }, { - "left": "issue:43854", - "right": "issue:44977", + "left": "issue:43454", + "right": "issue:43577", "accept": false, - "reason": "GLM-4.7 test loading failure and Qwen3.5 flash-attention generation issue do not share the same concrete defect." + "reason": "Both are model-loading issues, but one is lm_head tying in AyaVision and the other is Blip2 dtype propagation; not the same bug." }, { - "left": "issue:43454", - "right": "issue:43577", + "left": "issue:43854", + "right": "issue:44977", "accept": false, - "reason": "AyaVision weight tying and BLIP2 dtype propagation are distinct model-loading bugs." + "reason": "Different issues: GLM-4.7-Flash test loading failure vs Qwen3.5 flash-attention generation behavior." }, { "left": "issue:43526", "right": "issue:44554", "accept": false, - "reason": "BEiT label reduction and an MPS attention correctness issue are unrelated." + "reason": "Different subsystems: BEiT label reduction bug vs MPS attention correctness issue." }, { "left": "issue:36010", "right": "issue:43645", "accept": false, - "reason": "GenerationMixin import breakage and notebook custom-model initialization failures are both v5 regressions, but not the same bug." + "reason": "Both are API regressions, but one is GenerationMixin import resolution and the other is custom model initialization in notebooks." }, { "left": "issue:44568", "right": "issue:45488", "accept": false, - "reason": "Both are tokenizer regressions, but one is add_special_tokens/BOS-EOS handling and the other is tokenizer.json pre-tokenizer overriding; different root causes." + "reason": "Both concern tokenization, but they hit different tokenizer implementations and failure modes." }, { - "left": "issue:39692", - "right": "issue:44912", + "left": "issue:34634", + "right": "issue:36331", "accept": false, - "reason": "Documentation/example problems for SigLIP2 and MXFP4 quantization fallback in git-oss are unrelated." + "reason": "Different training/audio bugs: BarkProcessor voice_preset vs CustomTrainer.compute_loss signature change." }, { - "left": "issue:34634", - "right": "issue:36331", + "left": "issue:39692", + "right": "issue:44912", "accept": false, - "reason": "BarkProcessor voice_preset and CustomTrainer.compute_loss kwarg handling are unrelated API bugs." + "reason": "Unrelated: documentation/example errors for SigLIP2 vs MXFP4 quantization fallback for git-oss-20b." }, { "left": "issue:43257", "right": "issue:43329", "accept": false, - "reason": "Qwen3 MoE weight conversion with accelerate/deepspeed and multimodal video token counting use different code paths and symptoms." + "reason": "Different problems: Qwen3 MoE weight conversion/loading vs undefined helpers in multimodal video token counting." }, { "left": "issue:43819", "right": "issue:43927", "accept": false, - "reason": "DAC from_latents/STE mismatch and DiaConfig save-load token ID loss are different model/serialization bugs." + "reason": "Different model bugs: DAC latent reconstruction mismatch vs DiaConfig losing custom token IDs on save/load." }, { "left": "issue:43720", "right": "issue:44610", "accept": false, - "reason": "Packed-weight unpacking during accelerate loading and OmDet processor output-size mismatch are not the same underlying issue." + "reason": "Different failures: BitNet packed-weight unpacking during accelerate loading vs OmDet-Turbo processor output shape/size mismatch." }, { "left": "issue:43749", "right": "issue:44291", "accept": false, - "reason": "Both concern loading, but FSDP CPU RAM efficient loading and init_empty_weights/_is_hf_initialized are separate regressions." + "reason": "Both involve loading infrastructure, but one is FSDP CPU RAM efficient loading and the other is init_empty_weights/_is_hf_initialized signature breakage." }, { "left": "issue:43550", "right": "issue:44912", "accept": false, - "reason": "torch.compile with SDPA on Bamba and MXFP4 fallback on git-oss are unrelated model/runtime bugs." + "reason": "Different model/runtime issues: torch.compile SDPA bug vs MXFP4 quantization fallback." }, { "left": "issue:43262", "right": "issue:44811", "accept": false, - "reason": "Audio chat-template sample-rate defaults and Whisper batch_decode token skipping are separate processor behaviors." + "reason": "Different audio processor bugs: chat-template sampling rate default vs Whisper decode special-token handling." }, { "left": "issue:44568", "right": "issue:45356", "accept": false, - "reason": "Both are tokenizer regressions, but they affect different models and different parts of the tokenization pipeline." + "reason": "Different tokenizer regressions on different models and code paths; not the same defect." }, { "left": "issue:43531", "right": "issue:45042", "accept": false, - "reason": "Qwen3-MoE sliding_window and PIL backend torchvision dependency are unrelated." + "reason": "Unrelated: Qwen3-MoE sliding_window bug vs PIL backend torchvision requirement regression." }, { - "left": "issue:36010", - "right": "issue:36331", + "left": "issue:43643", + "right": "issue:45759", "accept": false, - "reason": "Importing GenerationMixin and passing num_items_in_batch to compute_loss are separate incompatibilities." + "reason": "Both are config/loading-related, but one is trust_remote_code returning missing fields and the other is AutoModelForCausalLM.from_config not unwrapping text_config." }, { - "left": "issue:43643", - "right": "issue:45759", + "left": "issue:36010", + "right": "issue:36331", "accept": false, - "reason": "Missing fields from trust_remote_code config loading and composite Qwen config unwrapping are both config-related, but not the same concrete defect." + "reason": "Different APIs: import path for GenerationMixin vs Trainer.compute_loss keyword-argument compatibility." }, { "left": "issue:34567", "right": "issue:40990", "accept": false, - "reason": "Trainer token accounting and high perplexity on gpt-oss-20b are unrelated." + "reason": "Unrelated: TrainerState token counter not updating vs GPT-OSS perplexity regression." }, { "left": "issue:43329", "right": "issue:43450", "accept": false, - "reason": "Undefined video-token helper names and batched video processor shapes are different video pipeline bugs." + "reason": "Both are video-related, but one is missing functions in token counting and the other is batched video processor shape handling." }, { "left": "issue:43329", "right": "issue:44479", "accept": false, - "reason": "The video-token helper bug and the v5.3 video-input regression overlap only at a high level; the concrete root causes differ." + "reason": "Shared video theme only; one is undefined helpers in a video branch, the other is a broader v5.3.0 video input regression." }, { "left": "issue:44849", "right": "issue:45701", "accept": false, - "reason": "Qwen3.5 hidden-states output and version-dependent tokenization changes are unrelated." + "reason": "Different regressions: Qwen3.5 hidden-state output bug vs version-dependent tokenization changes." }, { "left": "issue:43257", "right": "issue:43931", "accept": false, - "reason": "Qwen3 MoE accelerate/deepspeed weight conversion and Qwen3-VL weight-shape mismatch are different loading problems." + "reason": "Different Qwen-related loading bugs: MoE weight conversion vs VL weight shape mismatch." }, { "left": "issue:43697", "right": "issue:45357", "accept": false, - "reason": "RTDetrV2 output drift and Qwen3.5 save_pretrained key mismatch are separate regressions." + "reason": "Different issues: RTDetrV2 output drift in v5 vs save_pretrained storing incorrect visual encoder keys." }, { "left": "issue:43452", "right": "issue:43582", "accept": false, - "reason": "gguf_file/from_pretrained breakage and Apple Silicon caching_allocator_warmup TypeError are unrelated." + "reason": "Unrelated: gguf_file/from_pretrained loading breakage vs Apple Silicon caching_allocator_warmup TypeError." }, { "left": "issue:41720", "right": "issue:43540", "accept": false, - "reason": "Both involve Qwen3-family video/device issues, but one is CUDA assert under auto device mapping and the other is a video-input ValueError in Qwen3OmniMoe; not the same bug." + "reason": "Different Qwen3-family failures: auto device mapping cuda assert vs video input ValueError in Qwen3OmniMoe." + }, + { + "left": "issue:44291", + "right": "issue:45042", + "accept": false, + "reason": "Different bugs in different layers: init_empty_weights argument handling vs image processor torchvision dependency." } ] }, @@ -5278,8 +5185,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -5321,7 +5228,6 @@ "issue:44964", "issue:44987", "issue:44991", - "issue:45042", "issue:45072", "issue:45081", "issue:45137", @@ -5334,18 +5240,17 @@ ], "soft_pairs": [ "issue:42915|issue:43742", - "issue:44291|issue:45042", "issue:33357|issue:34567", "issue:43317|issue:45538", "issue:44987|issue:45538", "issue:44479|issue:45381", - "issue:43927|issue:44912", "issue:43540|issue:44560", + "issue:43927|issue:44912", "issue:42915|issue:44821", "issue:43257|issue:43756", "issue:43577|issue:45072", - "issue:42915|issue:45538", "issue:43540|issue:44479", + "issue:42915|issue:45538", "issue:43257|issue:43696", "issue:45137|issue:45405", "issue:38175|issue:44898", @@ -5355,8 +5260,8 @@ "issue:43906|issue:45381", "issue:43646|issue:43749", "issue:43749|issue:44756", - "issue:36010|issue:42175", "issue:44062|issue:45081", + "issue:36010|issue:42175", "issue:34567|issue:36331", "issue:43643|issue:45375", "issue:43493|issue:43994", @@ -5366,232 +5271,226 @@ "issue:44291|issue:44821", "issue:44792|issue:44964" ], - "prepared_review_unit_hash": "7a4afd73b2344db050c82714bcdb3496455ebeb3ea8d4f25666480d2fc2e0d36", + "prepared_review_unit_hash": "10475a892ee88241aff13a2d3e0b91d5fea5da0d52da5e458f69c5482d49f8e9", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { - "node_count": 48, - "item_count": 48, - "soft_pair_count": 32, - "serialized_chars": 37440, - "estimated_input_tokens": 9360, - "estimated_eval_tokens": 18976 + "node_count": 47, + "item_count": 47, + "soft_pair_count": 31, + "serialized_chars": 36567, + "estimated_input_tokens": 9142, + "estimated_eval_tokens": 18540 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is mostly heterogeneous and does not look like a single duplicate family. I only found two soft-edge pairs that plausibly describe the same underlying bug; the rest are different models, different subsystems, or different failure modes.", - "confidence": 0.94, - "canonical_issue_reason": "No single canonical issue fits: most items are unrelated bugs across different models/subsystems, with only a couple of tight duplicate-style pairs.", + "summary": "This cluster is mostly heterogeneous: most items are distinct bug reports across unrelated models/subsystems. The only strong duplicate-like matches are the Qwen video-input regression pair and the SigLIP2 correctness pair.", + "confidence": 0.66, + "canonical_issue_reason": "issue:44479 is the broadest, most central issue in the set around the shared Qwen video-input regression, and it has the strongest umbrella framing for the accepted duplicate-like report.", "canonical_pr_reason": null, - "best_issue_reason": "issue:44964 is the clearest representative user-facing loading regression, with a specific model and reproducible failure in the current transformers version.", + "best_issue_reason": "issue:44479 is the best single representative of the cluster\u2019s overlapping video-regression theme; it is broader than the narrower follow-up report and is more suitable as the canonical issue.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:42915", "right": "issue:43742", "accept": false, - "reason": "Different models and failures: Qwen3Moe FP8 config vs MobileLLM key error." - }, - { - "left": "issue:44291", - "right": "issue:45042", - "accept": false, - "reason": "Unrelated bugs: init_empty_weights arg handling vs PIL backend torchvision dependency." + "reason": "Different models and failure modes: Qwen3Moe FP8 config vs MobileLLM key error." }, { "left": "issue:33357", "right": "issue:34567", "accept": false, - "reason": "MacOS bus error in CLIP loading vs TrainerState token-count tracking bug." + "reason": "Unrelated areas: MacOS bus error with CLIP vs TrainerState token-count tracking." }, { "left": "issue:43317", "right": "issue:45538", "accept": false, - "reason": "Different areas: dequantized model offload failure vs tokenizer max_length constant." + "reason": "Different problems: offload/loading failure vs CLIPTokenizer max-length constant." }, { "left": "issue:44987", "right": "issue:45538", "accept": false, - "reason": "Model-loading regression for physical-intelligence/fast vs CLIPTokenizer max_length issue." + "reason": "Different loading bugs: model load regression vs tokenizer max-length issue." }, { "left": "issue:44479", "right": "issue:45381", "accept": true, - "reason": "Both describe the same Qwen video-input regression in transformers 5.3.0, centered on incorrect video/vision position handling." + "reason": "Both describe the same Qwen video-input regression; 45381 looks like the narrower symptom/diagnosis of the broader v5.3.0 video-input bug." }, { - "left": "issue:43927", - "right": "issue:44912", + "left": "issue:43540", + "right": "issue:44560", "accept": false, - "reason": "DiaConfig save/load corruption is unrelated to MXFP4 quantization fallback." + "reason": "Both involve video inputs, but they are different Qwen model variants with different errors; not enough to treat as the same bug." }, { - "left": "issue:43540", - "right": "issue:44560", + "left": "issue:43927", + "right": "issue:44912", "accept": false, - "reason": "Both are video-related but on different Qwen model paths and different errors." + "reason": "Different root causes: config token IDs being lost vs MXFP4 quantization fallback/loading behavior." }, { "left": "issue:42915", "right": "issue:44821", "accept": false, - "reason": "Qwen3Moe FP8 failure is unrelated to AutoImageProcessor URL loading." + "reason": "Different targets and failures: Qwen3Moe FP8 config vs AutoImageProcessor URL loading." }, { "left": "issue:43257", "right": "issue:43756", "accept": false, - "reason": "Different model families and bugs: weight conversion with accelerate/deepspeed vs RoPE-layer omission." + "reason": "Distinct model-specific issues: MOE weight conversion under accelerate/deepspeed vs Smollm3 RoPE layer mismatch." }, { "left": "issue:43577", "right": "issue:45072", "accept": false, - "reason": "Both mention dtype mismatches, but they affect different models and code paths." + "reason": "Both mention dtype, but they affect different models and code paths; too broad to merge." }, { - "left": "issue:42915", - "right": "issue:45538", + "left": "issue:43540", + "right": "issue:44479", "accept": false, - "reason": "Qwen3Moe quantization bug vs tokenizer max_length bug." + "reason": "Same broad topic of video inputs, but the model families and symptoms differ; not clearly the same concrete bug." }, { - "left": "issue:43540", - "right": "issue:44479", + "left": "issue:42915", + "right": "issue:45538", "accept": false, - "reason": "Similar domain, but different Qwen models and not clearly the same underlying bug." + "reason": "Completely different scopes: FP8/Qwen3Moe loading vs CLIP tokenizer max length." }, { "left": "issue:43257", "right": "issue:43696", "accept": false, - "reason": "Qwen3 MOE conversion issue is unrelated to GPT-oss OOM." + "reason": "Different underlying issue classes: MOE conversion/loading vs CUDA OOM." }, { "left": "issue:45137", "right": "issue:45405", "accept": false, - "reason": "DeepSpeed ZeRO3 deque error vs PEFT version metadata problem." + "reason": "DeepSpeed ZeRO3 deque failure is unrelated to PEFT version bump." }, { "left": "issue:38175", "right": "issue:44898", "accept": false, - "reason": "Both are vision-model bugs, but they are different models and different failure modes." + "reason": "Both are image-model correctness bugs, but they concern different architectures and failure modes." }, { "left": "issue:44737", "right": "issue:45588", "accept": false, - "reason": "XLNet CPU device bug is unrelated to flash-attention aux-state crash." + "reason": "XLNet CPU-device bug is unrelated to flash-attention sink-less model crash." }, { "left": "issue:44062", "right": "issue:44991", "accept": false, - "reason": "Tokenizer-related, but one is an AddedToken TypeError and the other is a model-specific tokenizer load failure without enough evidence they are the same bug." + "reason": "Tokenizer-related, but different models and insufficient evidence of the same special-token bug." }, { "left": "issue:43742", "right": "issue:44291", "accept": false, - "reason": "Different failures: MobileLLM key error vs unexpected _is_hf_initialized argument." + "reason": "Both are model-loading errors, but the specific causes are different (_is_hf_initialized vs MobileLLM key error)." }, { "left": "issue:43906", "right": "issue:45381", "accept": false, - "reason": "Isolated reproduction of a different issue vs Qwen2.5-VL video-position regression." + "reason": "Isolated reproduction of a different issue; not the same as the Qwen2.5-VL video regression." }, { "left": "issue:43646", "right": "issue:43749", "accept": false, - "reason": "Custom model initialization breakage is unrelated to FSDP CPU RAM efficient loading." + "reason": "Custom model init breakage is unrelated to FSDP CPU RAM-efficient loading." }, { "left": "issue:43749", "right": "issue:44756", "accept": false, - "reason": "Both concern loading/memory, but one is FSDP-specific and the other is mmap OOM on Strix Halo." + "reason": "Different memory/loading failures with different triggers; not the same bug." }, { - "left": "issue:36010", - "right": "issue:42175", + "left": "issue:44062", + "right": "issue:45081", "accept": false, - "reason": "ImportError for GenerationMixin is unrelated to missing TensorFlow in the torch extra." + "reason": "Tokenizer special-arg error vs Mistral regex/backend_tokenizer AttributeError are different failure points." }, { - "left": "issue:44062", - "right": "issue:45081", + "left": "issue:36010", + "right": "issue:42175", "accept": false, - "reason": "Tokenizer AddedToken error and Mistral regex patch crash are different code paths." + "reason": "Import path break for GenerationMixin is unrelated to TensorFlow being absent from a torch extra install." }, { "left": "issue:34567", "right": "issue:36331", "accept": false, - "reason": "TrainerState num_input_tokens_seen and custom compute_loss signature are separate trainer API regressions." + "reason": "TrainerState token accounting vs custom compute_loss signature mismatch are separate training issues." }, { "left": "issue:43643", "right": "issue:45375", "accept": false, - "reason": "Both involve config fields, but the failures are different and model-specific." + "reason": "Both concern config fields, but one is trust_remote_code config materialization and the other is a specific Qwen vision config field being dropped." }, { "left": "issue:43493", "right": "issue:43994", "accept": true, - "reason": "Both are SigLIP2 correctness bugs on the same model; one reports JAX/HF discrepancy and the other nonsensical outputs from the same implementation." + "reason": "Both report the same SigLIP2 model discrepancy/correctness problem: HF outputs disagree with expected behavior and produce nonsensical results." }, { "left": "issue:42175", "right": "issue:43824", "accept": false, - "reason": "Missing TensorFlow backend and missing Qwen2.5-VL export/import are unrelated." + "reason": "Package/dependency availability issue vs missing model class import; unrelated." }, { "left": "issue:43452", "right": "issue:44488", "accept": false, - "reason": "gguf_file/from_pretrained breakage is not the same as a specific model failing to load." + "reason": "Different loading breakages for different models and APIs." }, { "left": "issue:43526", "right": "issue:45479", "accept": false, - "reason": "BeitImageProcessorFast label reduction bug is unrelated to sequence-classification zero loss." + "reason": "Different image/label handling issues: BeitImageProcessorFast reduction bug vs sequence-classification loss degeneracy." }, { "left": "issue:44291", "right": "issue:44821", "accept": false, - "reason": "init_empty_weights argument handling is unrelated to AutoImageProcessor URL loading." + "reason": "General init_empty_weights loading bug vs AutoImageProcessor-from-URL failure are different code paths." }, { "left": "issue:44792", "right": "issue:44964", "accept": false, - "reason": "Janus test failure and Phi-4-multimodal load failure are different multimodal bugs." + "reason": "Janus generation test failure and Phi-4 multimodal loading failure are unrelated." } ] }, "evaluator_result": { "accept": true, - "feedback": "The summary is grounded in the packet and stays conservative. It correctly treats the cluster as mostly heterogeneous, and the two accepted soft-edge pairs are plausible same-bug families based on the issue titles and version/model overlap. I don\u2019t see a clear overstatement in the rejected pairs." + "feedback": "The summary is grounded in the packet and stays conservative. The two accepted duplicate-like pairs are reasonably supported by the issue titles/descriptions, and the remaining soft-edge rejections are appropriately cautious. No overstatement stands out." }, "evaluator_used": true, "retried": false, @@ -5607,8 +5506,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -5635,7 +5534,6 @@ "issue:44442", "issue:44466", "issue:44484", - "issue:44488", "issue:44514", "issue:44561", "issue:44568", @@ -5648,6 +5546,7 @@ "issue:44877", "issue:44898", "issue:44936", + "issue:44938", "issue:44977", "issue:45005", "issue:45042", @@ -5670,15 +5569,15 @@ "issue:43697|issue:45216", "issue:28282|issue:29127", "issue:44514|issue:45381", - "issue:43643|issue:44561", "issue:43531|issue:44821", - "issue:44625|issue:44877", + "issue:43643|issue:44561", "issue:43582|issue:44977", + "issue:44625|issue:44877", "issue:43452|issue:43577", "issue:43720|issue:44898", "issue:44514|issue:45290", "issue:43452|issue:44843", - "issue:42175|issue:44488", + "issue:42175|issue:44938", "issue:44466|issue:45005", "issue:43819|issue:44811", "issue:43295|issue:45356", @@ -5686,205 +5585,198 @@ "issue:43653|issue:44568", "issue:24643|issue:29127", "issue:38175|issue:43295", - "issue:44361|issue:44442", - "issue:43653|issue:44361" + "issue:44361|issue:44442" ], - "prepared_review_unit_hash": "67fe28b40af45649a84f4e03bcd7c5dccd48987266bce7857149413edca7114d", + "prepared_review_unit_hash": "4bca7397a15a2ff095c2d0091c6669a6d94828502a7a1e11cf3547d8af6d16cf", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, - "soft_pair_count": 27, - "serialized_chars": 35540, - "estimated_input_tokens": 8885, - "estimated_eval_tokens": 18026 + "soft_pair_count": 26, + "serialized_chars": 35227, + "estimated_input_tokens": 8807, + "estimated_eval_tokens": 17870 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is mostly a set of unrelated model-loading, tokenizer, config, and backend regressions. I only see one plausible duplicate edge around tied-weights serialization in v5; the rest should stay separate.", - "confidence": 0.74, - "canonical_issue_reason": "issue:45005 is the broadest statement of the tied-weights regression and best captures the shared underlying problem; issue:44466 looks like a more specific manifestation of the same v5 serialization bug.", + "summary": "This cluster is mostly heterogeneous and does not look like a single duplicate set. The only convincing duplicate pair is the tied-weights save_pretrained regression; the rest are unrelated tokenizer, config, model-loading, and vision bugs.", + "confidence": 0.88, + "canonical_issue_reason": "Issue 44466 is the most concrete anchor for the tied-weights regression: it describes the symptom, version boundary, and device-dependent serialization behavior more specifically than 45005.", "canonical_pr_reason": null, - "best_issue_reason": "issue:45005 is the best representative issue for the only likely duplicate subgroup because it describes the core tied-weights problem at a higher level.", + "best_issue_reason": "Issue 44466 is the strongest standalone issue in the cluster because it has the clearest repro target and the most actionable description of the underlying save_pretrained/tied-weights problem.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:28282", "right": "issue:34689", "accept": false, - "reason": "Both are model-loading failures, but one is a missing PyTorch dependency and the other is a specific Llama 3.2 Vision regression; different causes and fixes." + "reason": "Different failures: missing PyTorch import vs. model loading regression for a specific Llama vision model." }, { "left": "issue:45042", "right": "issue:45081", "accept": false, - "reason": "These are unrelated: one is a torchvision dependency regression in PIL image processors, the other is a Mistral regex/tokenizer attribute crash." + "reason": "Unrelated subsystems: image processor torchvision dependency vs. Mistral tokenizer regex crash." }, { "left": "issue:44484", "right": "issue:44704", "accept": false, - "reason": "One is about the default `max_shard_size` in `save_pretrained`, the other about forwarding kwargs in `AutoProcessor.from_pretrained`; not the same bug." + "reason": "One is a save_pretrained shard-size question; the other is a cached_file kwargs propagation bug." }, { "left": "issue:45405", "right": "issue:45600", "accept": false, - "reason": "Different maintenance issues: PEFT version pinning versus stale auto-mapping references to removed configs." + "reason": "Release/versioning problem vs. broken auto_mappings references; not the same bug." }, { "left": "issue:43756", "right": "issue:45325", "accept": false, - "reason": "Both involve position/rope-related behavior, but they affect different models and different code paths; not the same concrete defect." + "reason": "Both involve positional/RoPE logic, but they affect different models and different code paths." }, { "left": "issue:43742", "right": "issue:44589", "accept": false, - "reason": "A key error while loading MobileLLM is unrelated to the Float8 storage lookup failure." + "reason": "Different loading/runtime failures: missing key for MobileLLM vs. Float8 storage lookup error." }, { "left": "issue:43697", "right": "issue:45216", "accept": false, - "reason": "Different regressions in different model families: RTDetrV2 output changes versus Qwen3.5 checkpoint saving." + "reason": "Different regression surfaces: output drift in RTDetrV2 vs. incorrect checkpoint serialization for Qwen3.5." }, { "left": "issue:28282", "right": "issue:29127", "accept": false, - "reason": "Missing PyTorch in the environment is not the same as poor error messaging for LayoutLMv3 boxes." + "reason": "ImportError for PyTorch is unrelated to the LayoutLMv3 box-validation error message issue." }, { "left": "issue:44514", "right": "issue:45381", "accept": false, - "reason": "Both mention chat/processor pipelines, but one is a batched chat-template crash and the other is a Qwen2.5-VL video position-id bug." - }, - { - "left": "issue:43643", - "right": "issue:44561", - "accept": false, - "reason": "One is missing fields from `AutoConfig.from_pretrained` with `trust_remote_code`, the other is a v5 breakage from removing `is_torch_fx_available`." + "reason": "Both are Qwen2.5-VL chat/vision bugs, but one is batched padding handling and the other is wrong video position ids." }, { "left": "issue:43531", "right": "issue:44821", "accept": false, - "reason": "A Qwen3-MoE sliding-window bug and an AutoImageProcessor URL-loading issue are unrelated." + "reason": "Sliding-window behavior for Qwen3-MoE is unrelated to loading an AutoImageProcessor from a URL." }, { - "left": "issue:44625", - "right": "issue:44877", + "left": "issue:43643", + "right": "issue:44561", "accept": false, - "reason": "Different config propagation problems affecting different model families." + "reason": "Both involve trust_remote_code, but one is missing fields in AutoConfig and the other is a removed helper breaking imports." }, { "left": "issue:43582", "right": "issue:44977", "accept": false, - "reason": "Apple Silicon allocator warmup TypeError is unrelated to Qwen3.5 flash-attention generation behavior." + "reason": "AppleSilicon allocator TypeError is unrelated to Qwen3.5 flash-attention generation issues." + }, + { + "left": "issue:44625", + "right": "issue:44877", + "accept": false, + "reason": "Config propagation of num_labels is a different problem from strict config blocking granite_speech loading." }, { "left": "issue:43452", "right": "issue:43577", "accept": false, - "reason": "One is a gguf/loader regression, the other is dtype propagation on BLIP-2; they do not share the same underlying bug." + "reason": "gguf_file loading breakage and Blip2 dtype propagation are different model-loading bugs." }, { "left": "issue:43720", "right": "issue:44898", "accept": false, - "reason": "Packed-weight handling in accelerate loading is unrelated to Perceiver positional interpolation at non-default resolution." + "reason": "Packed-weight unpacking during accelerate loading is unrelated to Perceiver interpolation failure." }, { "left": "issue:44514", "right": "issue:45290", "accept": false, - "reason": "Same general API area, but the failures are different edge cases: batched padding behavior versus assistant tool-call messages without content." + "reason": "Both hit apply_chat_template, but the triggers and failure modes differ enough that they are not the same bug." }, { "left": "issue:43452", "right": "issue:44843", "accept": false, - "reason": "Both involve loading helpers, but one is gguf loading and the other is an offline-mode regression in `_patch_mistral_regex`." + "reason": "Different failure points: gguf_file loading vs. unconditional model_info() call breaking offline mode." }, { "left": "issue:42175", - "right": "issue:44488", + "right": "issue:44938", "accept": false, - "reason": "Packaging/backend selection and a specific model-load failure are not the same concrete issue." + "reason": "Packaging/backend dependency issue is unrelated to a Python 3.14 load failure." }, { "left": "issue:44466", "right": "issue:45005", "accept": true, - "reason": "Both point to the same tied-weights/save_pretrained regression in v5 for translation models; one is the broad issue, the other a specific manifestation." + "reason": "Both describe the same tied-weights/save_pretrained regression in v5, with 44466 giving the clearest device-dependent serialization symptom." }, { "left": "issue:43819", "right": "issue:44811", "accept": false, - "reason": "DAC latent conversion mismatch and Whisper batch_decode special-token handling are unrelated." + "reason": "DAC latent conversion mismatch and Whisper batch_decode skip_special_tokens behavior are unrelated." }, { "left": "issue:43295", "right": "issue:45356", "accept": false, - "reason": "Different regressions: processor/tokenizer access in custom model code versus codec and warning behavior in Kimi-K2.5 tokenization." + "reason": "Different tokenizer regressions with different symptoms and likely different code paths." }, { "left": "issue:41762", "right": "issue:44936", "accept": false, - "reason": "DeepSpeed ZeRO-3 loading failure for Gemma3 is unrelated to `trainer.evaluate()` failing after training." + "reason": "DeepSpeed ZeRO-3 loading IndexError is unrelated to trainer.evaluate() failing after trainer.train()." }, { "left": "issue:43653", "right": "issue:44568", "accept": false, - "reason": "BigBird special-token registration and mDeBERTa BOS/EOS insertion are different tokenizer bugs." + "reason": "Different tokenizer/special-token bugs affecting different models and behaviors." }, { "left": "issue:24643", "right": "issue:29127", "accept": false, - "reason": "DeepSpeed `'weight' must be 2-D` is a training/runtime error, not a LayoutLMv3 error-message clarity issue." + "reason": "DeepSpeed weight-dimension RuntimeError is unrelated to LayoutLMv3 box-information validation." }, { "left": "issue:38175", "right": "issue:43295", "accept": false, - "reason": "SigLIP2 zero probabilities and the processor.tokenizer regression are unrelated." + "reason": "Zero-probabilities output issue is unrelated to processor/tokenizer regression." }, { "left": "issue:44361", "right": "issue:44442", "accept": false, - "reason": "Both are tokenizer-loading failures, but they affect different tokenizer classes and do not appear to share a single code path." - }, - { - "left": "issue:43653", - "right": "issue:44361", - "accept": false, - "reason": "A special-token registration bug in BigBird and an AttributeError in MLukeTokenizer are different issues." + "reason": "Different tokenizer load failures for different model families." } ] }, "evaluator_result": { "accept": true, - "feedback": "The summary stays grounded in the packet and is appropriately conservative. The only accepted duplicate edge, issue:44466 \u2194 issue:45005, is plausibly the same tied-weights / v5 save_pretrained regression, while the other pair rejections avoid overstating similarity and mostly distinguish different model families or code paths." + "feedback": "The cluster summary is conservative and grounded in the packet: the tied-weights save_pretrained pair (44466/45005) is the only clearly mergeable duplicate, and the other pairwise rejections stay within the evidence suggested by the titles. No soft-edge verdict appears to overstate sameness." }, "evaluator_used": true, "retried": false, @@ -5900,8 +5792,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -5955,8 +5847,9 @@ "issue:45588" ], "soft_pairs": [ - "issue:44464|issue:44743", "issue:30064|issue:33666", + "issue:43653|issue:44361", + "issue:44464|issue:44743", "issue:43577|issue:43697", "issue:44843|issue:44991", "issue:41720|issue:43906", @@ -5977,237 +5870,243 @@ "issue:28282|issue:35141", "issue:43504|issue:44610", "issue:43452|issue:44451", - "issue:43688|issue:43749", + "issue:43756|issue:44913", "issue:43329|issue:45325", "issue:43742|issue:44821", - "issue:43756|issue:44913", - "issue:33666|issue:36010", + "issue:43688|issue:43749", "issue:43525|issue:43572", "issue:41720|issue:44464", + "issue:33666|issue:36010", "issue:43493|issue:43825", - "issue:44849|issue:45357", - "issue:43262|issue:45084" + "issue:43262|issue:45084", + "issue:44849|issue:45357" ], - "prepared_review_unit_hash": "126633ec37d95f89458e6c23d746d96ba3f643a00c520b6ec7377784d4003c45", + "prepared_review_unit_hash": "40735fe0efd11e59b73ef79c443cee48b614b9b4bb563b0e373d2f06ab3d48d7", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, - "soft_pair_count": 32, - "serialized_chars": 37415, - "estimated_input_tokens": 9354, - "estimated_eval_tokens": 18964 + "soft_pair_count": 33, + "serialized_chars": 37710, + "estimated_input_tokens": 9428, + "estimated_eval_tokens": 19112 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "These items are mostly unrelated Transformer bugs/regressions across different models and subsystems; I did not find any true duplicate pairs. I picked the broadest model-loading regression as the canonical issue.", - "confidence": 0.64, - "canonical_issue_reason": "issue:43452 is the broadest, most central regression report here: it affects both AutoTokenizer.from_pretrained and AutoModelForCausalLM.from_pretrained via gguf_file handling, making it a good cluster anchor.", + "summary": "This cluster is heterogeneous: it mixes unrelated tokenizer, trainer, vision, config, and generation regressions. None of the soft pairs look like the same underlying bug, so no dedup merges are justified.", + "confidence": 0.96, + "canonical_issue_reason": "No single canonical issue stands out; the items span multiple unrelated subsystems and model families, and the candidate pairs are not close enough to be duplicates.", "canonical_pr_reason": null, - "best_issue_reason": "issue:43452 is concrete, reproducible, and broadly scoped enough to represent the loading-path regressions in this set.", + "best_issue_reason": "Issue 44843 is the clearest self-contained regression report, but it is still too specific to serve as a true canonical for this mixed cluster.", "best_pr_reason": null, "soft_edge_verdicts": [ { - "left": "issue:44464", - "right": "issue:44743", + "left": "issue:30064", + "right": "issue:33666", "accept": false, - "reason": "Both involve generation/runtime behavior, but they describe different failures in different code paths: compiled forward inconsistency vs cache/reset behavior in modular_qwen3_5." + "reason": "Image processor segmentation-map failure vs Qwen2-VL multi-GPU training; different models and failure modes." }, { - "left": "issue:30064", - "right": "issue:33666", + "left": "issue:43653", + "right": "issue:44361", "accept": false, - "reason": "Completely different problems: segmentation-map preprocessing versus Qwen2-VL multi-GPU training." + "reason": "BigBirdTokenizer special-token decode bug vs MLukeTokenizer task AttributeError; not the same code path." + }, + { + "left": "issue:44464", + "right": "issue:44743", + "accept": false, + "reason": "Both affect generation, but one is compiled-forward inconsistency and the other is recurrent-state reset with cache; distinct bugs." }, { "left": "issue:43577", "right": "issue:43697", "accept": false, - "reason": "Different models and symptoms: BLIP2 dtype propagation vs RTDetrV2 output mismatch in v5." + "reason": "Blip2 dtype-loading issue vs RTDetrV2 output regression; unrelated subsystems and symptoms." }, { "left": "issue:44843", "right": "issue:44991", "accept": false, - "reason": "Both are tokenizer/loading regressions, but the titles point to different concrete failures and models; not enough to treat as the same bug." + "reason": "Both are tokenizer-load regressions, but one is an offline model_info call in Mistral regex patching and the other is a specific tokenizer load failure for a different model." }, { "left": "issue:41720", "right": "issue:43906", "accept": false, - "reason": "Different failures: Qwen3 auto device mapping CUDA assert vs an isolated reproduction of another issue." + "reason": "Insufficient evidence of the same defect; one is an A800 device-map cuda assert, the other is only a reproduction of another issue." }, { "left": "issue:44479", "right": "issue:44936", "accept": false, - "reason": "Video-input regression in VLMs versus trainer lifecycle failure; unrelated code paths." + "reason": "Video-input regression for several Qwen VL models vs trainer evaluate/train lifecycle failure; unrelated." }, { "left": "issue:44977", "right": "issue:45356", "accept": false, - "reason": "Flash-attention generation bug in Qwen3.5 versus tokenizer codec/regex regression in Kimi-K2.5." + "reason": "Flash-attention generation bug vs tokenizer codec/regex regression; different layers of the stack." }, { "left": "issue:43688", "right": "issue:45440", "accept": false, - "reason": "Both are MoE-related at a high level, but they target different models and different implementation mismatches." + "reason": "Both involve MoE models, but one is auxiliary-loss normalization and the other is native-vs-remote implementation divergence; not mergeable into one fix." }, { "left": "issue:43577", "right": "issue:45588", "accept": false, - "reason": "BLIP2 dtype issue and flash_attention crash on s_aux=None are unrelated bugs." + "reason": "Dtype propagation bug vs flash_attention crash on s_aux=None; different failure mechanisms." }, { "left": "issue:43531", "right": "issue:43742", "accept": false, - "reason": "Sliding-window behavior in Qwen3-MoE is unrelated to the MobileLLM loading KeyError." + "reason": "Qwen3-MoE sliding_window issue vs MobileLLM load KeyError; unrelated." }, { "left": "issue:43653", "right": "issue:43927", "accept": false, - "reason": "Tokenizer special-token registration versus config custom-token persistence; different underlying defects." + "reason": "Both are token-related, but one is mask-token registration and the other is saved custom token IDs causing generation IndexError; distinct bugs." }, { "left": "issue:44464", "right": "issue:44977", "accept": false, - "reason": "Different generation failures with different triggers: compiled forward versus flash-attention." + "reason": "Compiled-forward inconsistency vs flash-attention generation failure; different concrete problems." }, { "left": "issue:43550", "right": "issue:43927", "accept": false, - "reason": "torch.compile/SDPA failure is unrelated to DiaConfig save/load token-ID loss." + "reason": "torch.compile/SDPA failure vs config save/load token-ID loss; unrelated." }, { "left": "issue:43550", "right": "issue:44361", "accept": false, - "reason": "Different subsystems and failures: compile-time SDPA issue versus tokenizer AttributeError on tasks." + "reason": "Bamba compile+SDPA bug vs MLukeTokenizer AttributeError; no shared underlying issue." }, { "left": "issue:43122", "right": "issue:43906", "accept": false, - "reason": "Tokenization-difference regression and an isolated reproduction issue are not clearly the same bug." + "reason": "Different tokenization regression vs isolated reproduction of another issue; no evidence they are the same bug." }, { "left": "issue:43504", "right": "issue:43720", "accept": false, - "reason": "Legacy-field model-loading bug versus packed-weight handling in accelerate loading; different code paths." + "reason": "BEiT legacy-field load bug vs BitNet packed-weight unpacking bug; different loading paths." }, { "left": "issue:32090", "right": "issue:34567", "accept": false, - "reason": "Trainer broadcast TypeError and num_input_tokens_seen not updating are distinct trainer bugs." + "reason": "Trainer broadcast NoneType error vs token-count bookkeeping not updating; both trainer-related but not the same defect." }, { "left": "issue:43504", "right": "issue:43550", "accept": false, - "reason": "Unrelated model-loading preset bug versus compile/SDPA failure." + "reason": "Legacy-field pretrained-model load bug vs Bamba torch.compile+SDPA failure; unrelated." }, { "left": "issue:32090", "right": "issue:33666", "accept": false, - "reason": "Trainer NoneType broadcast error is unrelated to Qwen2-VL multi-GPU training." + "reason": "Trainer _gpu_broadcast_one TypeError vs Qwen2-VL multi-GPU training; unrelated." }, { "left": "issue:28282", "right": "issue:35141", "accept": false, - "reason": "ImportError for missing PyTorch and embedding reinitialization after resize are unrelated." + "reason": "Missing PyTorch import vs embedding resize reinitialization after post_init; different concerns entirely." }, { "left": "issue:43504", "right": "issue:44610", "accept": false, - "reason": "Different models and failure modes: legacy field on load versus processor/model input-size mismatch." + "reason": "BEiT preset legacy-field bug vs OmDet-Turbo processor size mismatch; different models and code paths." }, { "left": "issue:43452", "right": "issue:44451", "accept": false, - "reason": "Both are loading issues, but one is gguf_file handling for AutoTokenizer/AutoModelForCausalLM and the other is a model-specific load failure." + "reason": "gguf_file loading breakage vs a specific model load failure; both are load issues but not the same one." }, { - "left": "issue:43688", - "right": "issue:43749", + "left": "issue:43756", + "right": "issue:44913", "accept": false, - "reason": "Different MoE/train-loading issues: auxiliary-loss normalization versus FSDP CPU RAM efficient loading." + "reason": "RoPE layer-dropping bug vs rotary_pct reload bug; both config-related but clearly different regressions." }, { "left": "issue:43329", "right": "issue:45325", "accept": false, - "reason": "Both are multimodal token/position bugs, but they affect different branches and different concrete calculations." + "reason": "Undefined video-branch helpers in multimodal token counting vs still-image temporal position scaling in Qwen2.5-VL; distinct code paths." }, { "left": "issue:43742", "right": "issue:44821", "accept": false, - "reason": "Model KeyError on load versus AutoImageProcessor-from-URL failure; unrelated." - }, - { - "left": "issue:43756", - "right": "issue:44913", - "accept": false, - "reason": "RoPE-layer dropping in a specific model versus a config field resetting on reload." + "reason": "Model KeyError on load vs AutoImageProcessor URL-loading failure; unrelated." }, { - "left": "issue:33666", - "right": "issue:36010", + "left": "issue:43688", + "right": "issue:43749", "accept": false, - "reason": "Multi-GPU training bug versus missing GenerationMixin import; unrelated." + "reason": "Aux-loss normalization bug vs FSDP CPU RAM efficient loading breakage; different distributed-training issues." }, { "left": "issue:43525", "right": "issue:43572", "accept": false, - "reason": "Both are missing config fields after 5.0, but they concern different models and different fields, so not the same underlying bug." + "reason": "Missing pad_token_id in Llama4Config vs missing pad_token_idx in StableLmConfig; similar theme, but different model/config regressions." }, { "left": "issue:41720", "right": "issue:44464", "accept": false, - "reason": "CUDA assert in Qwen3 auto device mapping is unrelated to chunked generation with compiled forward." + "reason": "A800 auto device-map cuda assert vs chunked generation inconsistency with compiled forward; unrelated." }, { - "left": "issue:43493", - "right": "issue:43825", + "left": "issue:33666", + "right": "issue:36010", "accept": false, - "reason": "SigLIP2 implementation discrepancy and pipeline translation-message regression are unrelated." + "reason": "Qwen2-VL multi-GPU training issue vs GenerationMixin import error; unrelated." }, { - "left": "issue:44849", - "right": "issue:45357", + "left": "issue:43493", + "right": "issue:43825", "accept": false, - "reason": "Qwen3.5 hidden-states bug versus incorrect visual encoder keys on save_pretrained; different issues." + "reason": "SigLIP2 implementation discrepancy vs pipeline translation-task error message; different bugs." }, { "left": "issue:43262", "right": "issue:45084", "accept": false, - "reason": "Audio processor chat-template sampling-rate default bug is unrelated to template compilation TypeError." + "reason": "Audio processor sampling-rate default bug vs template compilation TypeError; unrelated." + }, + { + "left": "issue:44849", + "right": "issue:45357", + "accept": false, + "reason": "Qwen3.5 output_hidden_states regression vs save_pretrained writing wrong visual-encoder keys; different parts of the model lifecycle." } ] }, @@ -6226,8 +6125,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -6291,8 +6190,8 @@ "issue:43688|issue:45704", "issue:39692|issue:43493", "issue:33666|issue:36331", - "issue:43720|issue:44361", "issue:43334|issue:45759", + "issue:43720|issue:44361", "issue:44964|issue:45405", "issue:44479|issue:44625", "issue:43540|issue:44977", @@ -6304,8 +6203,8 @@ "issue:43278|issue:43844", "issue:43824|issue:44589", "issue:43646|issue:44466", - "issue:44589|issue:45520", "issue:39692|issue:43295", + "issue:44589|issue:45520", "issue:43688|issue:45701", "issue:33357|issue:33666", "issue:43504|issue:43606", @@ -6313,241 +6212,241 @@ "issue:43572|issue:44291", "issue:43927|issue:44361", "issue:43504|issue:43653", - "issue:43575|issue:45405", - "issue:44625|issue:45759" + "issue:44625|issue:45759", + "issue:43575|issue:45405" ], - "prepared_review_unit_hash": "f6594de271c2ecde4c4f34328ada6d53608fd59986ce71910646e74a82dc2128", + "prepared_review_unit_hash": "0cf7a325194358c5398ae78978d12c74257fd0d106a55629ae6611e78624f5bc", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 34, - "serialized_chars": 38212, - "estimated_input_tokens": 9553, - "estimated_eval_tokens": 19362 + "serialized_chars": 38205, + "estimated_input_tokens": 9552, + "estimated_eval_tokens": 19360 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is mostly a grab-bag of unrelated Transformers issues spanning Qwen multimodal loading, dtype/device regressions, tokenizers, training, and docs. I do not see a single underlying bug that unifies the set, and none of the soft-edge pairs look mergeable as true duplicates.", - "confidence": 0.9, - "canonical_issue_reason": "No strong canonical issue: the items cover many distinct code paths and model families, so one issue cannot represent the whole cluster cleanly.", + "summary": "This cluster is a heterogeneous mix of unrelated bugs (model loading, tokenizer/config regressions, dtype/device issues, training/serving failures). None of the soft pairs looks like the same underlying defect under a strict duplicate standard.", + "confidence": 0.93, + "canonical_issue_reason": "No single canonical issue: the cluster is too heterogeneous for duplicate consolidation, with only a few loose Qwen/config-loading themes.", "canonical_pr_reason": null, - "best_issue_reason": "If a representative must be chosen, issue 43299 is the closest to a recurring theme around Qwen3/Qwen3-VL MoE loading regressions, but it is still only a partial fit for the mixed cluster.", + "best_issue_reason": "No global best issue; at most, issue 45759 is the closest representative of a small Qwen multimodal config-loading subgroup, but it does not cover the rest of the cluster.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43540", "right": "issue:43716", "accept": false, - "reason": "Different bugs: Qwen3OmniMoe video-processing failure vs Mistral-3 image-preprocessor/model dtype mismatch." + "reason": "Different models and failures: Qwen3OmniMoe video-processing ValueError vs Mistral-3 image-preprocessor dtype mismatch." }, { "left": "issue:43720", "right": "issue:43927", "accept": false, - "reason": "Unrelated failures: BitNet packed-weight loading vs DiaConfig token-ID persistence." + "reason": "Unrelated code paths: packed-weight loading in BitNet vs DiaConfig token-id persistence/generation IndexError." }, { "left": "issue:43606", "right": "issue:43927", "accept": false, - "reason": "CPU-offload device mismatch is unrelated to DiaConfig save/load token-ID loss." + "reason": "CPU offload device mismatch is unrelated to DiaConfig save/load token ID loss." }, { "left": "issue:43334", "right": "issue:43572", "accept": false, - "reason": "Both are config attribute issues, but for different models and different missing fields; not the same bug." + "reason": "Both mention missing pad token fields, but they affect different models/configs and likely different fixes." }, { "left": "issue:28282", "right": "issue:32090", "accept": false, - "reason": "ImportError for missing PyTorch is unrelated to Trainer GPU broadcast rejecting NoneType." + "reason": "ImportError about missing PyTorch is unrelated to Trainer broadcast TypeError." }, { "left": "issue:28282", "right": "issue:33357", "accept": false, - "reason": "Different symptoms and code paths: missing PyTorch vs MacOS bus error on a CLIP model." + "reason": "Different failures: missing PyTorch ImportError vs MacOS bus error with a CLIP model." }, { "left": "issue:43577", "right": "issue:44554", "accept": false, - "reason": "BLIP2 dtype propagation bug is unrelated to the MPS attention correctness issue." + "reason": "Different bug classes: BLIP2 dtype propagation vs MPS attention correctness when head dims differ." }, { "left": "issue:43688", "right": "issue:45704", "accept": false, - "reason": "Aux-loss normalization in MoE models is unrelated to T5/apex memory leak behavior." + "reason": "Auxiliary-loss normalization in MoE models is unrelated to T5/apex RMSNorm memory leak." }, { "left": "issue:39692", "right": "issue:43493", "accept": false, - "reason": "Both touch SigLIP2, but one is a docs example problem and the other is an implementation discrepancy; not the same concrete bug." + "reason": "Doc example errors are not the same as HF-vs-JAX implementation discrepancy for SigLIP2." }, { "left": "issue:33666", "right": "issue:36331", "accept": false, - "reason": "Qwen2-VL multi-GPU training request is unrelated to the custom trainer compute_loss signature regression." - }, - { - "left": "issue:43720", - "right": "issue:44361", - "accept": false, - "reason": "BitNet packed-weight loading and MLukeTokenizer task failure are unrelated." + "reason": "Multi-GPU training request is not the same as CustomTrainer compute_loss signature regression." }, { "left": "issue:43334", "right": "issue:45759", "accept": false, - "reason": "Both involve Qwen multimodal configs, but one is missing pad_token_id and the other is from_config not unwrapping text_config; distinct bugs." + "reason": "Both are Qwen multimodal config issues, but one is missing pad_token_id and the other is from_config text_config unwrapping; not the same bug." + }, + { + "left": "issue:43720", + "right": "issue:44361", + "accept": false, + "reason": "BitNet accelerate-loading issue is unrelated to MLukeTokenizer AttributeError." }, { "left": "issue:44964", "right": "issue:45405", "accept": false, - "reason": "Model loading failure for Phi-4 multimodal is unrelated to the PEFT version bump issue." + "reason": "Model loading regression for Phi-4 multimodal is unrelated to an unreleased PEFT minimum-version bump." }, { "left": "issue:44479", "right": "issue:44625", "accept": false, - "reason": "Video-input regression and num_labels propagation are different issues." + "reason": "Video-input regression across Qwen variants is distinct from num_labels propagation in Qwen3.5 config." }, { "left": "issue:43540", "right": "issue:44977", "accept": false, - "reason": "Different Qwen models and different failures: video processing vs flash-attention generation." + "reason": "Different Qwen3.5/Qwen3OmniMoe video-processing failure vs flash-attention generation bug." }, { "left": "issue:44488", "right": "issue:44843", "accept": false, - "reason": "Loading a specific model failing is unrelated to the offline-mode mistral regex patch issue." + "reason": "Loading a specific model is unrelated to offline-mode breakage in AutoTokenizer._patch_mistral_regex." }, { "left": "issue:43493", "right": "issue:43844", "accept": false, - "reason": "SigLIP2 implementation mismatch and ZeRO-3 gradient inflation are unrelated." + "reason": "SigLIP2 implementation mismatch is unrelated to ZeRO-3 gradient inflation." }, { "left": "issue:43756", "right": "issue:45083", "accept": false, - "reason": "Smollm3 RoPE-layer discrepancy is unrelated to the qwen3_omni_moe helper function behavior." + "reason": "Smollm3 RoPE-layer mismatch is unrelated to a helper-function behavior issue in qwen3_omni_moe." }, { "left": "issue:45200", "right": "issue:45229", "accept": false, - "reason": "Defaulting mm_token_type_ids and a Gemma4 multi-GPU OOM are not the same bug." + "reason": "Text-only fine-tuning token-id defaulting is unrelated to Gemma4 multi-GPU CUDA OOM." }, { "left": "issue:43257", "right": "issue:43299", "accept": false, - "reason": "Both are Qwen3 MoE loading regressions, but they affect different paths and failure modes; too broad to treat as one duplicate." + "reason": "Both are Qwen MOE-related, but one is weight conversion under accelerate/deepspeed and the other is composite model loading; not the same defect." }, { "left": "issue:43278", "right": "issue:43844", "accept": false, - "reason": "Embedding dtype drift during eval is unrelated to the ZeRO-3 gradient issue." + "reason": "Embedding dtype drift during eval is unrelated to gradient growth under HfDeepSpeedConfig + ZeRO-3." }, { "left": "issue:43824", "right": "issue:44589", "accept": false, - "reason": "Missing Qwen2_5_VL import and missing Float8 storage type are unrelated runtime errors." + "reason": "Missing Qwen2.5-VL export is unrelated to Float8 storage lookup failure." }, { "left": "issue:43646", "right": "issue:44466", "accept": false, - "reason": "Custom model initialization breakage is unrelated to inconsistent lm_head.weight serialization." + "reason": "Custom model init breakage in v5 is unrelated to lm_head.weight serialization inconsistency." }, { - "left": "issue:44589", - "right": "issue:45520", + "left": "issue:39692", + "right": "issue:43295", "accept": false, - "reason": "Float8 storage lookup failure and flash_attn KeyError on Python 3.13 are different compatibility issues." + "reason": "A broken doc example is not the same as a regression in custom model code using processor.tokenizer and image passing." }, { - "left": "issue:39692", - "right": "issue:43295", + "left": "issue:44589", + "right": "issue:45520", "accept": false, - "reason": "Docs example errors are unrelated to the processor.tokenizer / image-to-tokenizer regression." + "reason": "Different import/runtime problems: missing Float8 storage class vs flash_attn KeyError on Python 3.13." }, { "left": "issue:43688", "right": "issue:45701", "accept": false, - "reason": "Aux-loss normalization is unrelated to tokenization changes." + "reason": "MoE auxiliary-loss normalization is unrelated to tokenization differences across transformer versions." }, { "left": "issue:33357", "right": "issue:33666", "accept": false, - "reason": "MacOS CLIP bus error and Qwen2-VL multi-GPU training are unrelated." + "reason": "MacOS bus error in CLIP loading is unrelated to Qwen2-VL multi-GPU training." }, { "left": "issue:43504", "right": "issue:43606", "accept": false, - "reason": "Legacy-field model loading failure and CPU-offload device mismatch are unrelated." + "reason": "BEiT legacy-field loading failure is unrelated to suno/bark-small CPU offload device mismatch." }, { "left": "issue:44423", "right": "issue:44734", "accept": false, - "reason": "Both are serve crashes, but with different root causes and code paths; not mergeable as one PR." + "reason": "Both hit serve, but one is continuous-batching multimodal crash and the other is KV-cache continuation tensor indexing; different bugs." }, { "left": "issue:43572", "right": "issue:44291", "accept": false, - "reason": "Missing pad_token_idx in StableLmConfig and unexpected _is_hf_initialized argument are separate regression types." + "reason": "Missing pad_token_idx after 5.0 update is unrelated to init_empty_weights rejecting _is_hf_initialized." }, { "left": "issue:43927", "right": "issue:44361", "accept": false, - "reason": "DiaConfig token-ID persistence and MLukeTokenizer task AttributeError are unrelated." + "reason": "DiaConfig token ID persistence bug is unrelated to MLukeTokenizer task AttributeError." }, { "left": "issue:43504", "right": "issue:43653", "accept": false, - "reason": "Beit legacy-field loading and BigBirdTokenizer special-token registration are different bugs." + "reason": "BEiT legacy-field loading failure is unrelated to BigBirdTokenizer special-token registration." }, { - "left": "issue:43575", - "right": "issue:45405", + "left": "issue:44625", + "right": "issue:45759", "accept": false, - "reason": "Tensor-parallel OOM and a release-version pin issue are unrelated." + "reason": "Both concern Qwen3.5 config handling, but one is num_labels propagation and the other is from_config text_config unwrapping; not the same underlying issue." }, { - "left": "issue:44625", - "right": "issue:45759", + "left": "issue:43575", + "right": "issue:45405", "accept": false, - "reason": "Qwen3.5 num_labels propagation and AutoModelForCausalLM.from_config text_config unwrapping are related only broadly, not the same concrete defect." + "reason": "Tensor-parallel OOM for a large model is unrelated to a PEFT version requirement bump." } ] }, @@ -6566,8 +6465,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -6626,8 +6525,8 @@ "issue:43994|issue:44493", "issue:29942|issue:30333", "issue:44464|issue:45584", - "issue:45084|issue:45520", "issue:43844|issue:44964", + "issue:45084|issue:45520", "issue:44554|issue:45446", "issue:44568|issue:45701", "issue:43688|issue:43873", @@ -6639,28 +6538,28 @@ "issue:43278|issue:43701", "issue:44315|issue:44849", "issue:44554|issue:45356", - "issue:45200|issue:45325", "issue:43295|issue:44610", + "issue:45200|issue:45325", "issue:43334|issue:43525", "issue:44898|issue:45412", "issue:44466|issue:45127", "issue:43493|issue:43701", "issue:37428|issue:43957" ], - "prepared_review_unit_hash": "a523895b82e2194eef7029e3cfb6650db7ad33b16932b7880862f3cc31d8402c", + "prepared_review_unit_hash": "1e34de14281a8fa704959989494164d873a1d9df73535df32823468d9d310a7e", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, "soft_pair_count": 26, - "serialized_chars": 35090, + "serialized_chars": 35092, "estimated_input_tokens": 8773, "estimated_eval_tokens": 17802 }, @@ -6669,168 +6568,162 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "The set is mostly heterogeneous; the soft links are only surface-level similarities across tokenizer/chat-template, attention, or loading regressions. I did not find any pair that clearly represents the same underlying bug or change.", - "confidence": 0.68, - "canonical_issue_reason": "Issue 45701 is the broadest umbrella-like report in the set and the best anchor if one must be chosen, but the cluster itself is not a clean duplicate group.", + "summary": "This set is a loose grab bag of unrelated Transformers issues: tokenization regressions, generation bugs, multimodal processor failures, device/offload problems, and CI/test noise. The proposed soft links are mostly superficial theme matches rather than the same underlying bug.", + "confidence": 0.91, + "canonical_issue_reason": null, "canonical_pr_reason": null, - "best_issue_reason": "45701 is open, general, and has inbound references, so it is the most suitable representative issue despite weak evidence that the other items are true duplicates.", + "best_issue_reason": "If one issue must represent the set, issue:45701 is the broadest/meta report ('transformers version changes the tokenization') and best captures the recurring version-regression theme, though it is not a true duplicate of the others.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:44792", "right": "issue:45405", "accept": false, - "reason": "A test failure in Janus is unrelated to a PEFT version pin issue; no shared bug or code path." + "reason": "Janus image-generation test failure vs an unreleased PEFT-version constraint; unrelated bugs." }, { "left": "issue:43606", "right": "issue:44112", "accept": false, - "reason": "Both mention device-related CI failures, but they affect different models and likely different execution paths." + "reason": "Both are CI/device-related, but one is a CPU offload mismatch on Bark and the other is a stale device override test in GraniteSpeech." }, { "left": "issue:44521", "right": "issue:45290", "accept": false, - "reason": "Both touch apply_chat_template, but one is about zero assistant masks for multimodal inputs and the other is a crash on tool-call messages." + "reason": "Both involve apply_chat_template, but one is multimodal assistant masks and the other is a crash on tool-call assistant messages; different concrete failures." }, { "left": "issue:43994", "right": "issue:44493", "accept": false, - "reason": "These are different correctness problems: one is SigLIP2 inference quality, the other is unexpected position-id keys." + "reason": "SigLIP2 nonsensical outputs vs a broad unexpected position-id key issue; different code paths and symptoms." }, { "left": "issue:29942", "right": "issue:30333", "accept": false, - "reason": "Flash Attention CI failures and MLFlow job-reporting issues are unrelated." + "reason": "FlashAttention 2 test failures and MLflow job-reporting issues are unrelated." }, { "left": "issue:44464", "right": "issue:45584", "accept": false, - "reason": "Chunked generation with compiled forward and Whisper empty-transcription handling are different failures." + "reason": "Compiled-forward chunked generation inconsistency vs Whisper empty-transcription generation failure; not the same bug." }, { - "left": "issue:45084", - "right": "issue:45520", + "left": "issue:43844", + "right": "issue:44964", "accept": false, - "reason": "A compilation error in template nodes and a Python 3.13 import_utils KeyError are not the same bug." + "reason": "ZeRO-3 gradient growth and Phi-4 multimodal load failures are separate training/loading problems." }, { - "left": "issue:43844", - "right": "issue:44964", + "left": "issue:45084", + "right": "issue:45520", "accept": false, - "reason": "ZeRO-3 gradient inflation and a model-loading regression are unrelated." + "reason": "Non-template node compilation error vs flash_attn import KeyError on Python 3.13; unrelated failures." }, { "left": "issue:44554", "right": "issue:45446", "accept": false, - "reason": "MPS attention correctness and a PyTorch version-check bug in flex_attention are different issues." + "reason": "MPS attention correctness issue and an AuxRequest version-check bug in flex_attention are different problems." }, { "left": "issue:44568", "right": "issue:45701", "accept": false, - "reason": "Both concern tokenization behavior, but one is a specific BOS/EOS regression and the other is a broad version-change complaint; not the same concrete bug." + "reason": "Both are tokenizer regressions, but 44568 is a specific mdeberta BOS/EOS bug while 45701 is a broad version-change report; not one concrete duplicate." }, { "left": "issue:43688", "right": "issue:43873", "accept": false, - "reason": "Aux-loss normalization and quantization/offloading behavior are unrelated." + "reason": "Aux-loss normalization and quantization offloading are different model-training/runtime issues." }, { "left": "issue:43540", "right": "issue:45200", "accept": false, - "reason": "Both are multimodal/model-input issues, but one is video processing in Qwen3OmniMoe and the other is Gemma 4 token-type defaults." + "reason": "Qwen3OmniMoe video-processing error vs Gemma 4 mm_token_type_ids defaulting; different multimodal bugs." }, { "left": "issue:15354", "right": "issue:38175", "accept": false, - "reason": "TorchScript GeneratorExp support and SigLIP2 zero probabilities are unrelated." + "reason": "torch.jit.script GeneratorExp limitation vs SigLIP2 zero-probability outputs; unrelated." }, { "left": "issue:45072", "right": "issue:45198", "accept": false, - "reason": "These are different regressions: dtype mismatch in inference versus Wav2Vec2 save/tokenization failure." + "reason": "BFloat16 dtype mismatches in inference and Wav2Vec2 save/tokenization failures do not share a concrete code-path." }, { "left": "issue:43232", "right": "issue:44849", "accept": false, - "reason": "Generation kwargs update after sync_gpus and Qwen3.5 hidden-states output are distinct generation bugs." + "reason": "Generation kwargs sync issue and Qwen3.5 hidden-states bug are separate generation-path regressions." }, { "left": "issue:39401", "right": "issue:44568", "accept": false, - "reason": "Both are tokenizer issues, but offset_mapping and add_special_tokens are separate concrete behaviors." + "reason": "Both are tokenizer bugs, but offset_mapping handling in Qwen3 and BOS/EOS insertion in mdeberta are distinct failures." }, { "left": "issue:43278", "right": "issue:43701", "accept": false, - "reason": "A dtype drift between train/eval and a resume_from_checkpoint key mismatch do not describe the same bug." + "reason": "BF16-to-FP32 embedding dtype drift vs resume_from_checkpoint key mismatch; unrelated." }, { "left": "issue:44315", "right": "issue:44849", "accept": false, - "reason": "Liger Kernel application and Qwen3.5 hidden-states output are unrelated." + "reason": "Liger Kernel application with model_init and Qwen3.5 hidden-states output are not the same bug." }, { "left": "issue:44554", "right": "issue:45356", "accept": false, - "reason": "Attention correctness on MPS and Kimi-K2.5 tokenizer codec handling are unrelated." + "reason": "MPS attention correctness vs Kimi-K2.5 tokenizer codec handling regression; different subsystems." }, { - "left": "issue:45200", - "right": "issue:45325", + "left": "issue:43295", + "right": "issue:44610", "accept": false, - "reason": "Both involve multimodal metadata, but mm_token_type_ids defaults and RoPE position scaling are different code paths." + "reason": "Processor/tokenizer API regression and OmDet-Turbo image-size mismatch are both processor-facing but not the same underlying issue." }, { - "left": "issue:43295", - "right": "issue:44610", + "left": "issue:45200", + "right": "issue:45325", "accept": false, - "reason": "Processor/tokenizer regression and OmDet-Turbo image-size mismatch are not the same underlying bug." + "reason": "Gemma 4 token-type defaults and Qwen2.5-VL temporal position-id scaling are unrelated multimodal issues." }, { "left": "issue:43334", "right": "issue:43525", "accept": false, - "reason": "The errors look similar, but they affect different model configs and there is no evidence of a shared root cause beyond missing pad_token_id access." + "reason": "Both mention missing pad_token_id, but they affect different model families and likely different config paths; too broad to merge." }, { "left": "issue:44898", "right": "issue:45412", "accept": false, - "reason": "Perceiver resolution handling and RT-DETR memory release are unrelated." + "reason": "Perceiver resolution/interpolation failure and RT-DETR memory leak are unrelated." }, { "left": "issue:44466", "right": "issue:45127", "accept": false, - "reason": "Serialization of tied lm_head weights and LoRA merge collapse with extended vocab are different problems." - }, - { - "left": "issue:43493", - "right": "issue:43701", - "accept": false, - "reason": "SigLIP2 HF-vs-JAX discrepancy and resume_from_checkpoint key mismatch are unrelated." + "reason": "Serialization/tied-weights inconsistency and LoRA-merge collapse with extended vocab are different weight-handling bugs." }, { "left": "issue:37428", "right": "issue:43957", "accept": false, - "reason": "An import error for flash attention helpers and meta-device model loading failures are different issues." + "reason": "Flash-attention import error and meta-device model-loading breakage are separate loading/runtime issues." } ] }, @@ -6849,8 +6742,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -6916,8 +6809,8 @@ "issue:43278|issue:43856", "issue:43295|issue:43994", "issue:37428|issue:43824", - "issue:38175|issue:43493", "issue:42915|issue:43824", + "issue:38175|issue:43493", "issue:43827|issue:43901", "issue:43653|issue:44112", "issue:43450|issue:43493", @@ -6927,242 +6820,245 @@ "issue:45127|issue:45245", "issue:43819|issue:45083", "issue:44206|issue:45584", - "issue:43819|issue:44112", "issue:43262|issue:43450", + "issue:43819|issue:44112", "issue:37428|issue:43425", "issue:44610|issue:44857", - "issue:44857|issue:45593", "issue:43278|issue:43825", "issue:45290|issue:45381", + "issue:44857|issue:45593", "issue:43606|issue:45083" ], - "prepared_review_unit_hash": "083439898bcdce58b3e1b29b202f19a931da9d9c9607dda6491ad41862fa064c", + "prepared_review_unit_hash": "6c84a3c27a42b637b5bcee019b73325cd45a8bc6aaef65e34dbecf2b264f69b2", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, "soft_pair_count": 32, - "serialized_chars": 36646, - "estimated_input_tokens": 9162, - "estimated_eval_tokens": 18580 + "serialized_chars": 36641, + "estimated_input_tokens": 9161, + "estimated_eval_tokens": 18578 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is heterogeneous: it mixes unrelated bugs in import compatibility, model-specific inference regressions, processor/chat-template issues, training/memory problems, and docs/test cleanups. None of the soft pairs look like true duplicates or mergeable PR-equivalents.", - "confidence": 0.94, - "canonical_issue_reason": "No single issue is a safe canonical: the items span unrelated regressions across import paths, processors, training, and documentation, and the proposed soft pairs do not describe the same underlying bug.", + "summary": "This cluster is mostly a set of unrelated transformer bug reports that only share broad subsystem vocabulary. I only accept the SigLIP2 pair as likely the same underlying model discrepancy; the rest look like distinct issues.", + "confidence": 0.78, + "canonical_issue_reason": "issue:43493 is the best canonical issue because it describes the broad SigLIP2 implementation discrepancy at the root of the symptom-specific output failures, making it the most general and reusable duplicate target in this set.", "canonical_pr_reason": null, - "best_issue_reason": "If a representative is needed, issue 44521 is the closest umbrella for the chat-template/multimodal subset, but it should not be treated as a cluster-wide duplicate anchor.", + "best_issue_reason": "issue:43493 is the strongest representative of the only plausible duplicate chain here: it frames the underlying SigLIP2 HF-vs-JAX mismatch, whereas nearby reports are narrower symptoms.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43572", "right": "issue:44987", "accept": false, - "reason": "Both involve v5-related model/config loading regressions, but one is a missing StableLmConfig pad_token_idx and the other is a transformers>=5.1.0 load failure for a specific checkpoint; different bugs." + "reason": "Both are version/load failures, but one is a missing config field and the other is a model load OOM/regression; different code paths." }, { "left": "issue:43572", "right": "issue:43957", "accept": false, - "reason": "These are separate v5 loading/config issues: missing pad_token_idx vs meta-device model loading breakage." + "reason": "Different failure modes: missing StableLm config field vs meta-device model loading breakage." }, { "left": "issue:43873", "right": "issue:44464", "accept": false, - "reason": "Quantization/offload behavior and chunked generation with compiled forward are different code paths and failure modes." + "reason": "Quantization/offloading behavior and compiled chunked generation are unrelated bugs." }, { "left": "issue:43824", "right": "issue:43975", "accept": false, - "reason": "ImportError for a missing Qwen2.5-VL symbol is unrelated to DeepSeek detokenization output corruption." + "reason": "ImportError for a missing Qwen2.5-VL symbol vs DeepSeek detokenization issue; not the same bug." }, { "left": "issue:43295", "right": "issue:43450", "accept": false, - "reason": "Processor.tokenizer/image handling regression is distinct from batched video processor shape errors." + "reason": "Processor.tokenizer regression and batched video shape bug are different processor issues." }, { "left": "issue:43844", "right": "issue:44936", "accept": false, - "reason": "ZeRO-3 gradient explosion and trainer.evaluate() failing after train() are separate training-state problems." + "reason": "DeepSpeed/ZeRO gradient inflation and trainer.evaluate() after train() are distinct training-loop problems." }, { "left": "issue:43295", "right": "issue:43873", "accept": false, - "reason": "Processor API regression and quantization/offloading failure are not the same underlying bug." + "reason": "Custom processor regression and quantization offloading bug are unrelated." }, { "left": "issue:38175", "right": "issue:45412", "accept": false, - "reason": "SigLIP2 zero probabilities and RT-DETR memory not being released are unrelated." + "reason": "SigLIP2 zero probabilities and RT-DETR memory cleanup are unrelated models and symptoms." }, { "left": "issue:44514", "right": "issue:44521", "accept": false, - "reason": "Both touch apply_chat_template on multimodal inputs, but one is a batched padding=False crash while the other is assistant_mask generation returning zeros; different concrete failures." + "reason": "Both involve apply_chat_template, but one is a batched-padding crash and the other is incorrect assistant masks; not enough evidence of one underlying bug." }, { "left": "issue:44442", "right": "issue:45198", "accept": false, - "reason": "FastSpeech2ConformerTokenizer loading failure and Wav2Vec2 save/tokenization failure are distinct tokenizer bugs." + "reason": "Different tokenizer/model families with separate tokenizer/save_pretrained failures." }, { "left": "issue:43278", "right": "issue:43856", "accept": false, - "reason": "Embedding dtype drift in evaluate() and Qwen3 MoE memory inefficiency are unrelated training/runtime issues." + "reason": "Embedding dtype drift in eval vs Qwen3 MoE memory efficiency are unrelated." }, { "left": "issue:43295", "right": "issue:43994", "accept": false, - "reason": "Processor/tokenizer regression is not the same as SigLIP2 nonsensical model outputs." + "reason": "Processor.tokenizer regression is not the same as SigLIP2 nonsensical outputs." }, { "left": "issue:37428", "right": "issue:43824", "accept": false, - "reason": "Both are import errors, but they concern different missing symbols in different modules and unrelated fixes." - }, - { - "left": "issue:38175", - "right": "issue:43493", - "accept": false, - "reason": "SigLIP2 zero probabilities and HF-vs-JAX discrepancy are related by model family, but they are not the same concrete bug." + "reason": "Different missing imports from different modules and model families." }, { "left": "issue:42915", "right": "issue:43824", "accept": false, - "reason": "Qwen3Moe FineGrainedFP8Config failure and missing Qwen2.5-VL import are unrelated." + "reason": "FineGrainedFP8Config failure and missing Qwen2.5-VL export are unrelated." + }, + { + "left": "issue:38175", + "right": "issue:43493", + "accept": true, + "reason": "Both point to the same SigLIP2 implementation problem: HF outputs are wrong/non-sensical relative to expected behavior, with 38175 showing a symptom and 43493 describing the broader mismatch." }, { "left": "issue:43827", "right": "issue:43901", "accept": false, - "reason": "Both are docs around pipeline deprecations, but they cover different task pages and different stale references." + "reason": "Both are docs-related, but they concern different pipeline APIs and different stale-deprecation issues." }, { "left": "issue:43653", "right": "issue:44112", "accept": false, - "reason": "BigBirdTokenizer special-token registration and GraniteSpeech device-override CI failure are unrelated." + "reason": "Tokenizer special-token registration and a GraniteSpeech CI device override test failure are unrelated." }, { "left": "issue:43450", "right": "issue:43493", "accept": false, - "reason": "Batched video shape bug and SigLIP2 implementation discrepancy do not share the same underlying code path." + "reason": "Video batching shape bug and SigLIP2 implementation discrepancy affect different models and code paths." }, { "left": "issue:44360", "right": "issue:44561", "accept": false, - "reason": "A DSA indexer activation issue and the removal of is_torch_fx_available breaking remote-code models are unrelated." + "reason": "DSA indexer design discussion and removal of is_torch_fx_available are unrelated." }, { "left": "issue:43450", "right": "issue:43575", "accept": false, - "reason": "Video batch-shape handling and Qwen2-57B tensor-parallel OOM are different problems." + "reason": "Batched video processor shape bug and tensor-parallel OOM on Qwen2-57B are different issues." }, { "left": "issue:44792", "right": "issue:45563", "accept": false, - "reason": "A janus test failure and a stale paged-generate warning are not the same bug." + "reason": "Janus image generation test failure and paged generate warning are unrelated generation issues." }, { "left": "issue:45127", "right": "issue:45245", "accept": false, - "reason": "LoRA merge collapse with tied embeddings and a 2^24 category limit runtime error are unrelated." + "reason": "LoRA/tied-embeddings collapse and category-count RuntimeError are different model-training failures." }, { "left": "issue:43819", "right": "issue:45083", "accept": false, - "reason": "DAC from_latents/forward mismatch and a qwen3_omni_moe helper behavior issue are different model-specific bugs." + "reason": "DAC latent forward mismatch and qwen3_omni_moe feature-length helper behavior are unrelated." }, { "left": "issue:44206", "right": "issue:45584", "accept": false, - "reason": "Unsupported center arg in LasrFeatureExtractor and Whisper empty-transcription failure after token alignment are unrelated." + "reason": "Feature extractor center-arg crash and Whisper empty-transcription failure are different audio bugs." }, { - "left": "issue:43819", - "right": "issue:44112", + "left": "issue:43262", + "right": "issue:43450", "accept": false, - "reason": "DAC latent-path mismatch and GraniteSpeech CI device override issues have no common code path." + "reason": "Audio chat-template sampling-rate default and video batched-shape bug are different processor issues." }, { - "left": "issue:43262", - "right": "issue:43450", + "left": "issue:43819", + "right": "issue:44112", "accept": false, - "reason": "Audio chat-template sampling-rate defaulting and batched video shape handling are unrelated processor bugs." + "reason": "DAC STE mismatch and GraniteSpeech CI device override failure are unrelated." }, { "left": "issue:37428", "right": "issue:43425", "accept": false, - "reason": "Flash-attention import breakage and Torch 2.10 incompatibility are separate compatibility issues." + "reason": "Flash-attention import error and Torch 2.10 incompatibility are different compatibility issues." }, { "left": "issue:44610", "right": "issue:44857", "accept": false, - "reason": "OmDet-Turbo processor size mismatch and LwDetrImageLoss AMP crash are different vision bugs." - }, - { - "left": "issue:44857", - "right": "issue:45593", - "accept": false, - "reason": "AMP/CUDA loss crash and D-FINE auxiliary-loss omission are unrelated model-training issues." + "reason": "Processor output-size mismatch and CUDA AMP loss crash are unrelated detection bugs." }, { "left": "issue:43278", "right": "issue:43825", "accept": false, - "reason": "Embedding dtype regression and pipeline error-message/docs mismatch are unrelated." + "reason": "Training/eval dtype drift and pipeline docs/error-message wording are unrelated." }, { "left": "issue:45290", "right": "issue:45381", "accept": false, - "reason": "Chat-template crashes on tool-call messages and qwen2.5-vl vision_position_ids errors are different multimodal bugs." + "reason": "Chat-template tool-call crash and Qwen2.5-VL vision_position_ids bug are different failure modes." + }, + { + "left": "issue:44857", + "right": "issue:45593", + "accept": false, + "reason": "Float16/CUDA loss crash and missing auxiliary losses in denoising-off mode are different detection training bugs." }, { "left": "issue:43606", "right": "issue:45083", "accept": false, - "reason": "CPU offload device mismatch for bark-small and a qwen3_omni_moe helper function issue are unrelated." + "reason": "CPU offload device mismatch and qwen3_omni_moe feature-length helper behavior are unrelated." } ] }, - "evaluator_result": null, - "evaluator_used": false, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet and stays conservative: it treats most pairs as unrelated and only accepts the SigLIP2 pair, which is plausible from the titles. The canonical choice of issue:43493 is also reasonable as the broader SigLIP2 discrepancy report. No obvious overstatement beyond what the packet supports." + }, + "evaluator_used": true, "retried": false, - "accepted_nontrivial_soft_edge": false, + "accepted_nontrivial_soft_edge": true, "error_kind": null, "error_message": null }, @@ -7174,8 +7070,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -7246,40 +7142,40 @@ "issue:44855|issue:45083", "issue:44292|issue:45520", "issue:39692|issue:44493", - "issue:43335|issue:43504", - "issue:43550|issue:44112", "issue:43335|issue:45083", + "issue:43550|issue:44112", + "issue:43335|issue:43504", "issue:44448|issue:44779", "issue:38175|issue:44493", "issue:43335|issue:45072", "issue:44977|issue:45229", "issue:43720|issue:44112", - "issue:45459|issue:45712", "issue:43335|issue:45593", + "issue:45459|issue:45712", "issue:44568|issue:45491", "issue:44448|issue:44977", "issue:29127|issue:32090", "issue:43606|issue:44857", "issue:40444|issue:43526", - "issue:44857|issue:45072", "issue:45372|issue:45588", + "issue:44857|issue:45072", "issue:34689|issue:35141", "issue:44977|issue:45563" ], - "prepared_review_unit_hash": "46f1592943c14f0202b34c0f0dace142d75a82c29ab3854622a93b1c5a12e0cd", + "prepared_review_unit_hash": "d9666ce86a147524e89affa9712db5eaa65089a2006b28968307198ed3d2cf94", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 36, - "serialized_chars": 38872, + "serialized_chars": 38870, "estimated_input_tokens": 9718, "estimated_eval_tokens": 19692 }, @@ -7288,198 +7184,234 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "The set is mostly a mix of unrelated issues across different models and subsystems. Only one soft pair looks like a true duplicate: the repeated Qwen3.5 MoE from_pretrained error report.", - "confidence": 0.97, - "canonical_issue_reason": "Issue 45310 is the earlier, clearer report of the same Qwen3.5 MoE from_pretrained regression as 45478, and it already has inbound reference activity.", + "summary": "This cluster is mostly a grab-bag of unrelated Transformers bug reports. The only clear duplicate pair is the two identical Qwen3.5 MoE from_pretrained errors; the rest share broad themes like model loading, tokenizers, or flash-attention but point to different code paths and should stay separate.", + "confidence": 0.94, + "canonical_issue_reason": "issue:45310 is the earlier report of the same Qwen3.5 MoE from_pretrained failure and is the cleaner canonical target.", "canonical_pr_reason": null, - "best_issue_reason": "45310 is the best representative issue because it is the earliest duplicate in the only true duplicate pair and has slightly stronger discussion/reference signals.", + "best_issue_reason": "issue:45310 is the best issue candidate because it is the earliest and directly matches the duplicate report in issue:45478.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:44448", "right": "issue:44743", "accept": false, - "reason": "Different models and different failure modes: Pegasus output drift vs Qwen3.5 recurrent-state reset." + "reason": "Different problems: Pegasus output regression vs Qwen3.5 recurrent-state reset in modular code." }, { "left": "issue:45200", "right": "issue:45588", "accept": false, - "reason": "Both mention attention-related code paths, but the bugs are unrelated: Gemma4 token-type handling vs flash_attention.py handling of s_aux=None." + "reason": "Unrelated: Gemma 4 token-type IDs defaulting vs flash-attention crash on sink-less models." }, { "left": "issue:43335", "right": "issue:43550", "accept": false, - "reason": "Different architectures and bugs: SwitchTransformers config creates sparse layers incorrectly, while Bamba fails under torch.compile with SDPA." + "reason": "Different models and failures: SwitchTransformers sparse-layer config vs Bamba torch.compile/SDPA." }, { "left": "issue:44462", "right": "issue:44561", "accept": false, - "reason": "Tokenizer file resolution and removal of is_torch_fx_available are unrelated issues." + "reason": "Tokenizer.json loading bug is unrelated to removal of is_torch_fx_available breaking trust_remote_code models." }, { "left": "issue:44561", "right": "issue:44779", "accept": false, - "reason": "One is a trust_remote_code break from a removed helper; the other is DeepSeek tokenizer behavior in v5." + "reason": "Different tokenizer/runtime regressions with different symptoms and likely different fixes." }, { "left": "issue:37428", "right": "issue:44351", "accept": false, - "reason": "Both are import errors, but for different missing symbols and different code paths; not the same bug." + "reason": "Both are import errors, but for different missing symbols and different affected features." }, { "left": "issue:43611", "right": "issue:43931", "accept": false, - "reason": "Both are loading failures, but one concerns base_model_prefix compatibility and the other a concrete Qwen3-VL weight-shape mismatch." + "reason": "Model loading breaks for different reasons: base_model_prefix handling vs weight-shape mismatch." }, { "left": "issue:43388", "right": "issue:43526", "accept": false, - "reason": "Mask2Former metric gathering and BEiT label reduction are different components with different defects." + "reason": "Both concern labels, but one is gather_for_metrics truncation and the other is BeitImageProcessorFast reduce_labels behavior." }, { "left": "issue:39692", "right": "issue:43526", "accept": false, - "reason": "Documentation example errors in SigLIP2 are unrelated to BEiTImageProcessorFast label reduction." + "reason": "Documentation/example errors are unrelated to the Beit label reduction bug." }, { "left": "issue:43819", "right": "issue:44493", "accept": false, - "reason": "DAC latent decoding mismatch and unexpected position-id keys are different model behaviors." + "reason": "DAC latent conversion mismatch is unrelated to unexpected position-id keys." }, { "left": "issue:45310", "right": "issue:45478", "accept": true, - "reason": "Same Qwen3.5 MoE from_pretrained error, same version range, same underlying regression." + "reason": "Same title and same concrete failure: transformers>=5.4.0 Qwen3.5 MoE from_pretrained error." }, { "left": "issue:43335", "right": "issue:43653", "accept": false, - "reason": "SwitchTransformers sparse-layer config bug is unrelated to BigBirdTokenizer special-token registration." + "reason": "Different tokenizer/model areas: SwitchTransformers config vs BigBirdTokenizer special-token registration." }, { "left": "issue:29942", "right": "issue:32090", "accept": false, - "reason": "Flash Attention 2 test failures and a Trainer GPU broadcast NoneType error are unrelated." + "reason": "Flash Attention 2 test failures are unrelated to Trainer _gpu_broadcast_one NoneType errors." }, { "left": "issue:43404", "right": "issue:45200", "accept": false, - "reason": "Mistral3 lm_head tying and Gemma4 mm_token_type_ids are different model-specific issues." + "reason": "Different bugs: untied lm_head weights vs missing mm_token_type_ids default." }, { "left": "issue:44855", "right": "issue:45083", "accept": false, - "reason": "Python 3.13 decorator parsing failure is unrelated to qwen3_omni_moe feature-length helper behavior." + "reason": "Python 3.13 JIT parsing issue is unrelated to qwen3_omni_moe feature-length computation." }, { "left": "issue:44292", "right": "issue:45520", "accept": false, - "reason": "Qwen-3-8B-NVFP4 runtime failure and flash_attn import lookup on Python 3.13 are different problems." + "reason": "Qwen-3-8B-NVFP4 runtime issue is not the same as a flash_attn import-utils KeyError on Python 3.13." }, { "left": "issue:39692", "right": "issue:44493", "accept": false, - "reason": "SigLIP2 doc example errors are unrelated to the widespread unexpected position-id key issue." + "reason": "SigLIP2 doc/example mistakes are unrelated to the position-id key regression." }, { "left": "issue:43335", - "right": "issue:43504", + "right": "issue:45083", "accept": false, - "reason": "SwitchTransformers config and BEiT preset loading fail for different reasons and in different code paths." + "reason": "Different subsystems and failure modes; no shared underlying bug." }, { "left": "issue:43550", "right": "issue:44112", "accept": false, - "reason": "Bamba compile/SDPA failure and GraniteSpeech CI device-override flakiness are unrelated." + "reason": "Bamba SDPA compile failure is unrelated to a GraniteSpeech CI device-override test." }, { "left": "issue:43335", - "right": "issue:45083", + "right": "issue:43504", "accept": false, - "reason": "Sparse-layer config bug is unrelated to qwen3_omni_moe output-length helper behavior." + "reason": "SwitchTransformers sparse-layer config bug is unrelated to BEiT legacy-field loading failure." }, { "left": "issue:44448", "right": "issue:44779", "accept": false, - "reason": "Different models and different regressions: Pegasus v4/v5 output drift vs DeepSeek tokenizer correctness in v5." + "reason": "Different model regressions: Pegasus output mismatch vs DeepSeek tokenizer behavior change." }, { "left": "issue:38175", "right": "issue:44493", "accept": false, - "reason": "Unexpected zero probabilities in SigLIP2 is unrelated to the position-id key warning across many models." + "reason": "SigLIP2 zero probabilities and unexpected position-id keys are different bugs." }, { "left": "issue:43335", "right": "issue:45072", "accept": false, - "reason": "SwitchTransformers config bug is unrelated to dtype mismatches in bfloat16 inference." + "reason": "SwitchTransformers config bug and bfloat16 dtype mismatch in inference are unrelated." }, { "left": "issue:44977", "right": "issue:45229", "accept": false, - "reason": "Flash-attention generation bug and Gemma4 multi-GPU OOM are different failure modes." + "reason": "Qwen3.5 flash-attention generation bug is unrelated to Gemma4 multi-GPU CUDA OOM." }, { "left": "issue:43720", "right": "issue:44112", "accept": false, - "reason": "Packed-weight loading in BitNet and GraniteSpeech CI device override are unrelated." + "reason": "Packed-weight loading failure in BitNet is unrelated to a stale GraniteSpeech CI test." }, { - "left": "issue:45459", - "right": "issue:45712", + "left": "issue:43335", + "right": "issue:45593", "accept": false, - "reason": "Protobuf error masking in tokenizers and dummy_pt_objects cleanup are different maintenance bugs." + "reason": "SwitchTransformers sparse-layer creation and D-FINE auxiliary-loss behavior are different issues." }, { - "left": "issue:43335", - "right": "issue:45593", + "left": "issue:45459", + "right": "issue:45712", "accept": false, - "reason": "SwitchTransformers sparse-layer config and D-FINE auxiliary-loss disabling are unrelated." + "reason": "Tokenizer error masking due to missing protobuf is unrelated to dummy-class cleanup/check_repo failures." }, { "left": "issue:44568", "right": "issue:45491", "accept": false, - "reason": "Tokenizer special-token insertion and Gemma3 NaN embeddings are unrelated." + "reason": "Tokenizer special-token behavior is unrelated to Gemma3 NaN embeddings from mixed-length batching." + }, + { + "left": "issue:44448", + "right": "issue:44977", + "accept": false, + "reason": "Different model families and symptoms: Pegasus output drift vs Qwen3.5 flash-attention generation issues." + }, + { + "left": "issue:29127", + "right": "issue:32090", + "accept": false, + "reason": "LayoutLMV3 error-message clarity is unrelated to Trainer broadcast NoneType errors." + }, + { + "left": "issue:43606", + "right": "issue:44857", + "accept": false, + "reason": "CPU-offload device mismatch and float16 AMP CUDA crash are distinct bugs." + }, + { + "left": "issue:40444", + "right": "issue:43526", + "accept": false, + "reason": "Qwen2.5-VL iterable-dataset finetuning failure is unrelated to BeitImageProcessorFast label reduction." + }, + { + "left": "issue:45372", + "right": "issue:45588", + "accept": false, + "reason": "Gemma 4 processor loading ImportError is unrelated to flash-attention sink-less-model crash." + }, + { + "left": "issue:44857", + "right": "issue:45072", + "accept": false, + "reason": "Different precision-related failures in different models and code paths." }, { "left": "issue:34689", "right": "issue:35141", "accept": false, - "reason": "Llama 3.2 vision loading breakage and embedding reinitialization after resize are different bugs." + "reason": "Model-loading regression for Llama 3.2 Vision is unrelated to token-embedding reinit after resize." }, { "left": "issue:44977", "right": "issue:45563", "accept": false, - "reason": "Qwen3.5 flash-attention generation failure and a stale warning for num_return_sequences are not the same issue." + "reason": "Qwen3.5 flash-attention generation bug is unrelated to a stale paging warning for num_return_sequences." } ] }, "evaluator_result": { "accept": true, - "feedback": "Grounded overall. The selected duplicate pair (issue:45310, issue:45478) is well supported by the packet: identical titles, same model area, close timing, and asymmetric inbound reference activity. The rejected soft pairs are phrased conservatively and do not overclaim duplicate status. No PR-pair mergeability issue applies here." + "feedback": "The summary is grounded in the packet and stays conservative. The duplicate call on issue:45310 / issue:45478 is supported by the identical title and matching failure description, and the other soft-edge rejections are framed narrowly enough to avoid overstating similarity." }, "evaluator_used": true, "retried": false, @@ -7495,8 +7427,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -7572,209 +7504,203 @@ "issue:43994|issue:45412", "issue:44265|issue:45198", "issue:44361|issue:45198", - "issue:43421|issue:45491", "issue:42915|issue:42994", + "issue:43421|issue:45491", "issue:44871|issue:44977", - "issue:43065|issue:43994", "issue:29942|issue:30064", + "issue:43065|issue:43994", "issue:44206|issue:44479" ], - "prepared_review_unit_hash": "90aeed0360823db087350b83d82d56c5663db1dc8f0c8b724e78ee75e8c50778", + "prepared_review_unit_hash": "9aa90e8e356efdb5534fa79bd40b129e10b955e842351496bc07b85b6bb90992", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 28, - "serialized_chars": 36055, - "estimated_input_tokens": 9014, - "estimated_eval_tokens": 18284 + "serialized_chars": 36052, + "estimated_input_tokens": 9013, + "estimated_eval_tokens": 18282 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "These issues are mostly unrelated singleton bugs spanning tokenizers, generation, quantization, vision, and import regressions. No soft pair looks like a true duplicate or the same fixable code-path change.", - "confidence": 0.22, - "canonical_issue_reason": "No clear duplicate cluster emerges, so the broadest and most central open report here is `issue:43873` (offloading/quantization), which has the most discussion and inbound references among the set.", - "canonical_pr_reason": null, - "best_issue_reason": "`issue:43873` is the best global representative only by breadth and activity; it is still not a duplicate sink for the other issues.", - "best_pr_reason": null, + "summary": "The cluster is very heterogeneous: most items are unrelated bugs/regressions across tokenization, export, quantization, flash-attention, and model-specific behavior. I found no true duplicate pair among the soft edges.", + "confidence": 0.94, + "canonical_issue_reason": "No single canonical issue stands out because the cluster does not cohere around one underlying bug; the items span many distinct code paths and symptoms.", + "canonical_pr_reason": "No PRs are present in this cluster.", + "best_issue_reason": "If one issue must be used as a representative, issue 44462 is the cleanest self-contained report with a concrete, actionable loading symptom around AutoTokenizer/tokenizer.json.", + "best_pr_reason": "No PRs are present, so there is no best PR candidate.", "soft_edge_verdicts": [ { "left": "issue:43701", "right": "issue:45083", "accept": false, - "reason": "Resume-from-checkpoint key mismatch and Qwen3 Omni feature-length helper behavior are different failures in unrelated code paths." + "reason": "Checkpoint-resume key mismatch and qwen3_omni_moe feature-length behavior are different subsystems and different bugs." }, { "left": "issue:43927", "right": "issue:45698", "accept": false, - "reason": "Both involve save/load, but one is custom token IDs being lost in config while the other is loading the wrong custom module; different root causes." + "reason": "Both involve save/load, but one is about custom token IDs in config and the other about loading the wrong custom module." }, { "left": "issue:43550", "right": "issue:45083", "accept": false, - "reason": "Torch.compile SDPA failure in Bamba-9B-v2 is unrelated to the Qwen3 Omni helper length bug." + "reason": "torch.compile/SDPA failure in Bamba-9B-v2 is unrelated to the qwen3_omni_moe helper-function issue." }, { "left": "issue:43066", "right": "issue:44462", "accept": false, - "reason": "Tokenizer decoder-type regression and ignoring tokenizer.json are both tokenizer-loading issues, but not the same concrete bug." + "reason": "Both touch tokenizers, but one is a decoder-type mismatch in v5 while the other is ignoring tokenizer.json from the repo." }, { "left": "issue:43412", "right": "issue:45588", "accept": false, - "reason": "Executorch export with missing training guard/erfinv support is unrelated to flash-attention crashing on s_aux=None." + "reason": "Executorch export/erfinv/training-guard failure and flash_attention s_aux=None crash are different code paths." }, { "left": "issue:45245", "right": "issue:45563", "accept": false, - "reason": "Category-cardinality runtime error and a stale warning for num_return_sequences are distinct behaviors." + "reason": "Category-count overflow and a stale num_return_sequences warning are unrelated." }, { "left": "issue:43873", "right": "issue:44493", "accept": false, - "reason": "Quantization/offloading behavior and unexpected position-id keys are different bugs in different parts of model handling." + "reason": "Quantization offloading behavior and unexpected position-id key warnings are not the same bug." }, { "left": "issue:29942", "right": "issue:35141", "accept": false, - "reason": "Failing Flash Attention 2 tests and embedding reinitialization after resizing token embeddings are unrelated." + "reason": "Flash Attention 2 test failures and embedding reinit in post_init are distinct regressions." }, { "left": "issue:45372", "right": "issue:45459", "accept": false, - "reason": "Mistral_common import error in Gemma 4 loading and hidden tokenizer protobuf errors are separate import/error-handling issues." + "reason": "ReasoningEffort import breakage and protobuf error masking affect different tokenizer/processor paths." }, { "left": "issue:36010", "right": "issue:37428", "accept": false, - "reason": "Both are import errors, but they concern different missing symbols and different compatibility breakages." + "reason": "These are both import errors, but for different symbols in different modules." }, { "left": "issue:44448", "right": "issue:44466", "accept": false, - "reason": "Different output from a model and inconsistent lm_head.weight serialization are not the same bug." + "reason": "Pegasus output differences and lm_head.weight serialization inconsistency are separate model-generation vs save/load issues." }, { "left": "issue:43065", "right": "issue:45412", "accept": false, - "reason": "Dummy Conv2d in Sam3PixelDecoder and RT-DETR memory not being released are unrelated." + "reason": "Sam3PixelDecoder dummy Conv2d and RT-DETR memory release are unrelated." }, { "left": "issue:25251", "right": "issue:32090", "accept": false, - "reason": "Pipeline top_k nested-list output and Trainer NoneType broadcast errors do not share a concrete code path." + "reason": "Pipeline top_k nesting behavior and Trainer _gpu_broadcast_one NoneType error are unrelated failures." }, { "left": "issue:43278", "right": "issue:43381", "accept": false, - "reason": "Eval-mode dtype drift and gradient checkpointing being disallowed in eval are distinct eval-time behaviors." + "reason": "Embedding dtype drift in eval and gradient checkpointing being disallowed in eval are different eval-mode problems." }, { "left": "issue:44857", "right": "issue:45588", "accept": false, - "reason": "AMP/CUDA loss crash in LwDetrImageLoss is unrelated to flash-attention s_aux=None handling." + "reason": "LwDetrImageLoss AMP/CUDA crash and flash_attention s_aux=None crash do not share the same underlying defect." }, { "left": "issue:44451", "right": "issue:45520", "accept": false, - "reason": "Model load failure for ScandiBERT and a Python 3.13 flash_attn import key error are different loading/import issues." + "reason": "Model load failure and Python 3.13 flash_attn KeyError are different import/runtime problems." }, { "left": "issue:43653", "right": "issue:44743", "accept": false, - "reason": "Tokenizer special-token registration and Qwen recurrent state reset are unrelated." + "reason": "BigBirdTokenizer special-token decoding and Qwen recurrent-state reset are unrelated." }, { "left": "issue:44186", "right": "issue:45198", "accept": false, - "reason": "Both are tokenizer-related, but they affect different models and failure modes; not the same underlying bug." + "reason": "LayoutLMv2Tokenizer padding/truncation crash and Wav2Vec2 save_pretrained/tokenization failure are different tokenizer bugs." }, { "left": "issue:43976", "right": "issue:45440", "accept": false, - "reason": "Python-version compatibility and DeepseekV3MoE divergence are unrelated." + "reason": "Python-version incompatibility and DeepseekV3MoE divergence are not the same issue." }, { "left": "issue:43994", "right": "issue:45412", "accept": false, - "reason": "Siglip2 nonsensical outputs and RT-DETR memory leaks are different classes of problems." + "reason": "Siglip2 nonsensical outputs and RT-DETR memory leakage are unrelated." }, { "left": "issue:44265", "right": "issue:45198", "accept": false, - "reason": "torch.export failures with torch_compilable_check are unrelated to Wav2Vec2 save_pretrained/tokenization failures." + "reason": "torch.export.export with torch_compilable_check and Wav2Vec2 save/tokenization failure are different defects." }, { "left": "issue:44361", "right": "issue:45198", "accept": false, - "reason": "MLukeTokenizer AttributeError and Wav2Vec2 save/tokenization failure are different tokenizer bugs." - }, - { - "left": "issue:43421", - "right": "issue:45491", - "accept": false, - "reason": "Runtime special-token post-processor updates and Gemma3 NaN embeddings on mixed-length batches are not the same issue." + "reason": "MLukeTokenizer AttributeError and Wav2Vec2 save/tokenization failure are unrelated tokenizer issues." }, { "left": "issue:42915", "right": "issue:42994", "accept": false, - "reason": "Both mention quantization, but one is FP8 inference failure and the other is quantized model saving; separate code paths." + "reason": "Both mention quantization, but one is a FineGrainedFP8Config runtime failure and the other is quantized model saving failure." }, { - "left": "issue:44871", - "right": "issue:44977", + "left": "issue:43421", + "right": "issue:45491", "accept": false, - "reason": "Gemma3 eos_token_id inconsistency and Qwen3.5 flash-attention generation issues are unrelated." + "reason": "Runtime special-token post-processor updates and Gemma3 NaN embeddings are different bugs." }, { - "left": "issue:43065", - "right": "issue:43994", + "left": "issue:44871", + "right": "issue:44977", "accept": false, - "reason": "Sam3PixelDecoder dummy layer and Siglip2 output regression are unrelated model bugs." + "reason": "Gemma eos_token_id mismatch and Qwen3.5 flash-attention generation regression are unrelated." }, { "left": "issue:29942", "right": "issue:30064", "accept": false, - "reason": "Flash Attention 2 test failures and void segmentation-map preprocessing are different subsystems." + "reason": "Flash Attention 2 test failures and void segmentation map processing failure are unrelated." }, { "left": "issue:44206", "right": "issue:44479", "accept": false, - "reason": "Unsupported center argument in a feature extractor and the Qwen2.5/3.5 VL video-input regression are separate input-processing bugs." + "reason": "Unsupported center argument in LasrFeatureExtractor and Qwen VL video regression are different issues." } ] }, @@ -7793,8 +7719,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -7861,8 +7787,8 @@ "issue:43278|issue:43329", "issue:43381|issue:43701", "issue:43653|issue:44442", - "issue:41720|issue:45229", "issue:45440|issue:45593", + "issue:41720|issue:45229", "issue:29942|issue:36010", "issue:43824|issue:44279", "issue:32090|issue:33357", @@ -7872,214 +7798,214 @@ "issue:43526|issue:44857", "issue:42915|issue:43931", "issue:43441|issue:45588", - "issue:44464|issue:45491", "issue:43262|issue:44206", - "issue:44743|issue:44811", + "issue:44464|issue:45491", "issue:43526|issue:44610", + "issue:44743|issue:44811", "issue:44442|issue:45072" ], - "prepared_review_unit_hash": "dd8bf60ae46b839da652151f4049d4f56f0205e93fa3911e49404b61079e531f", + "prepared_review_unit_hash": "9100251c14a0e352eff00225799b65ae5670905cd88fdd9b5532dabc3cdaf994", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 29, - "serialized_chars": 36482, - "estimated_input_tokens": 9121, - "estimated_eval_tokens": 18498 + "serialized_chars": 36476, + "estimated_input_tokens": 9119, + "estimated_eval_tokens": 18494 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster does not contain any strong duplicate pairs: the issues span unrelated models, processors, attention paths, tokenizer/import problems, and memory/loading regressions. All soft edges were rejected. There are no PRs in the set.", - "confidence": 0.97, - "canonical_issue_reason": "No real duplicate canonical stands out because the issues are unrelated; if one issue must be chosen as the anchor, issue #43873 is the broadest and most active item, but it is not actually representative of the others as a duplicate target.", + "summary": "This cluster is a broad false-positive mix of unrelated issues across tokenizers, model loading, training, attention, quantization, and docs. None of the soft pairs describe the same concrete bug or change, so no duplicates should be merged here.", + "confidence": 0.98, + "canonical_issue_reason": "No canonical issue: the items are heterogeneous and do not share one underlying bug, code path, or fix surface.", "canonical_pr_reason": null, - "best_issue_reason": "Issue #43873 is the best single representative only because it is open, discussion-heavy, and broadly framed; however, it is not a true cluster canonical since the rest of the items do not describe the same bug.", + "best_issue_reason": "No single best issue for deduplication; none stands out as the representative of a real duplicate group.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:44734", "right": "issue:45406", "accept": false, - "reason": "Different subsystems and failures: KV-cache continuation indexing in serve vs Gemma4 processor initialization missing _tokenizer." + "reason": "Different failures: KV-cache continuation tensor indexing vs Gemma4Processor missing _tokenizer." }, { "left": "issue:33357", "right": "issue:39692", "accept": false, - "reason": "Both are CLIP/SigLIP-adjacent only at a high level; one is a Mac bus error, the other is a doc example/model-processor mismatch and quantization failure." + "reason": "Unrelated: MacOS CLIP bus error vs SigLIP2 doc example/model-processor mismatch and quantization issue." }, { "left": "issue:44610", "right": "issue:45412", "accept": false, - "reason": "Processor output size mismatch versus a model memory leak are unrelated bugs." + "reason": "Different subsystems: processor image-size mismatch vs RT-DETR memory not released on deletion." }, { "left": "issue:29127", "right": "issue:34567", "accept": false, - "reason": "LayoutLMv3 box-content error messaging and TrainerState token counting are unrelated issues." + "reason": "Different problems: LayoutLMv3 error clarity vs TrainerState token-count tracking not updating." }, { "left": "issue:45103", "right": "issue:45372", "accept": false, - "reason": "Auto-docstring future-annotations crash is unrelated to Gemma4 processor loading/import failure." + "reason": "Unrelated: auto_docstring crash with future annotations vs mistral_common import breakage for Gemma 4 loading." }, { "left": "issue:43873", "right": "issue:45229", "accept": false, - "reason": "Both mention memory/placement, but one is quantization offloading behavior and the other is multi-GPU OOM for a specific model; not the same bug." + "reason": "Different issues: quantization offloading behavior vs Gemma4 multi-GPU CUDA OOM." }, { "left": "issue:43756", "right": "issue:45440", "accept": false, - "reason": "Different models and different correctness problems: RoPE layer selection versus DeepSeekV3 implementation divergence." + "reason": "Different model-specific bugs: Smollm3 RoPE layer handling vs DeepseekV3MoE divergence from remote implementation." }, { "left": "issue:45372", "right": "issue:45406", "accept": false, - "reason": "Both involve Gemma4, but one is a mistral_common import dependency issue and the other is a serve-time processor attribute error." + "reason": "ImportError in Gemma4 processor loading is unrelated to serve-time _tokenizer attribute crash." }, { "left": "issue:43723", "right": "issue:44554", "accept": false, - "reason": "Tokenizer loading regression versus an MPS attention correctness issue; no shared code-path." + "reason": "Different areas: AutoTokenizer loading regression vs MPS attention correctness bug." }, { "left": "issue:43278", "right": "issue:43575", "accept": false, - "reason": "Embedding dtype drift in eval mode is unrelated to tensor-parallel OOM on Qwen2-57B." + "reason": "Embedding dtype mismatch in eval vs tensor-parallel OOM on model load are unrelated." }, { "left": "issue:43278", "right": "issue:43329", "accept": false, - "reason": "Different bugs: dtype change in evaluate versus undefined video-token helper/merge_size in multimodal token counting." + "reason": "No shared code path: dtype change in evaluate vs multimodal token counting bug." }, { "left": "issue:43381", "right": "issue:43701", "accept": false, - "reason": "Eval-mode gradient checkpointing restriction is unrelated to resume_from_checkpoint key mismatch." + "reason": "Gradient checkpointing eval-mode restriction is unrelated to checkpoint resume key mismatch." }, { "left": "issue:43653", "right": "issue:44442", "accept": false, - "reason": "Both are tokenizer-related, but one is a BigBird special-token registration issue and the other is AutoTokenizer failing for a specific tokenizer class." + "reason": "Different tokenizer bugs: special-token registration vs AutoTokenizer failing to load a specific tokenizer class." }, { - "left": "issue:41720", - "right": "issue:45229", + "left": "issue:45440", + "right": "issue:45593", "accept": false, - "reason": "Different model families and symptoms: Qwen3 auto device mapping asserts on A800, while Gemma4 hits multi-GPU OOM." + "reason": "Unrelated model correctness issue vs D-FINE auxiliary-loss behavior." }, { - "left": "issue:45440", - "right": "issue:45593", + "left": "issue:41720", + "right": "issue:45229", "accept": false, - "reason": "DeepSeekV3 implementation divergence is unrelated to D-FINE auxiliary-loss behavior." + "reason": "Different GPU inference failures: Qwen3 auto device mapping assert vs Gemma4 multi-GPU OOM." }, { "left": "issue:29942", "right": "issue:36010", "accept": false, - "reason": "FlashAttention test failures and GenerationMixin import errors are unrelated regressions." + "reason": "Flash Attention test failures are unrelated to a GenerationMixin import error." }, { "left": "issue:43824", "right": "issue:44279", "accept": false, - "reason": "Specific missing symbol import versus a vague dependency issue; not enough evidence of the same underlying bug." + "reason": "Specific missing symbol import vs generic dependency issue; not the same bug." }, { "left": "issue:32090", "right": "issue:33357", "accept": false, - "reason": "Trainer NoneType broadcast error and MacOS CLIP bus error are unrelated failures." + "reason": "Trainer broadcast NoneType error and MacOS CLIP bus error are unrelated." }, { "left": "issue:43334", "right": "issue:43531", "accept": false, - "reason": "Qwen3-VL pad_token_id attribute error and Qwen3-MoE sliding_window behavior are different code paths." + "reason": "Different Qwen3 variants and different failures: missing pad_token_id vs sliding_window behavior." }, { "left": "issue:43526", "right": "issue:44186", "accept": false, - "reason": "BEiT label reduction bug and LayoutLMv2 NER padding/truncation crash are different processor/tokenizer issues." + "reason": "BEiT label reduction bug vs LayoutLMv2 tokenizer crash on NER/padding are unrelated." }, { "left": "issue:44977", "right": "issue:45245", "accept": false, - "reason": "Flash-attention generation bug and category-count runtime limit are unrelated." + "reason": "Flash-attention generation issue is unrelated to the 2^24 category limit runtime error." }, { "left": "issue:43526", "right": "issue:44857", "accept": false, - "reason": "BEiT label handling and LwDetrImageLoss AMP/CUDA crash are not the same defect." + "reason": "Different modules and symptoms: image-label reduction vs LwDetrImageLoss AMP crash." }, { "left": "issue:42915", "right": "issue:43931", "accept": false, - "reason": "Qwen3Moe FineGrainedFP8Config failure and Qwen3-VL weight-shape mismatch are different loading/configuration problems." + "reason": "Qwen3Moe FP8 config failure is not the same as Qwen3-VL weight-shape mismatch." }, { "left": "issue:43441", "right": "issue:45588", "accept": false, - "reason": "Both touch flash attention, but they describe different failures, models, and triggers; too broad to merge as one bug." - }, - { - "left": "issue:44464", - "right": "issue:45491", - "accept": false, - "reason": "Compiled-forward chunked generation inconsistency is unrelated to Gemma3 sliding-window NaN embeddings." + "reason": "Both mention flash attention, but they affect different code paths and concrete crashes." }, { "left": "issue:43262", "right": "issue:44206", "accept": false, - "reason": "Audio chat-template sample-rate default and a feature extractor unsupported argument are different audio-processing bugs." + "reason": "Audio chat-template sample-rate defaulting is unrelated to an unsupported feature-extractor argument." }, { - "left": "issue:44743", - "right": "issue:44811", + "left": "issue:44464", + "right": "issue:45491", "accept": false, - "reason": "Qwen3.5 recurrent-state reset bug and Whisper batch_decode skip_special_tokens bug are unrelated." + "reason": "Compiled-forward chunked generation inconsistency is unrelated to Gemma3 NaN embeddings from sliding-window attention." }, { "left": "issue:43526", "right": "issue:44610", "accept": false, - "reason": "Different image-processing/model-loading bugs; no shared underlying defect." + "reason": "Image label reduction bug and OmDet-Turbo input-size mismatch are distinct processor issues." + }, + { + "left": "issue:44743", + "right": "issue:44811", + "accept": false, + "reason": "Recurrent-state cache reset bug is unrelated to Whisper batch_decode skip_special_tokens handling." }, { "left": "issue:44442", "right": "issue:45072", "accept": false, - "reason": "AutoTokenizer loading failure for one tokenizer class is unrelated to dtype mismatches in SwitchTransformers/TimmWrapperModel inference." + "reason": "Tokenizer-loading failure and dtype mismatches in bfloat16 inference are unrelated." } ] }, @@ -8098,8 +8024,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -8168,9 +8094,9 @@ "issue:43761|issue:44079", "issue:43931|issue:45216", "issue:42907|issue:45698", - "issue:39692|issue:45491", "issue:33290|issue:34567", "issue:34689|issue:43611", + "issue:39692|issue:45491", "issue:44442|issue:44898", "issue:43761|issue:44188", "issue:43493|issue:43827", @@ -8180,14 +8106,14 @@ "issue:43295|issue:45584", "issue:43994|issue:44206" ], - "prepared_review_unit_hash": "2ca1a459290b1ddaa6fab8ea4ea9168243f600346c9d47553a0b2db5dea573d6", + "prepared_review_unit_hash": "2aa24714c0edb08e816c0448714ca132825e9b25012224bbdd5cf84104deffcb", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, @@ -8202,168 +8128,168 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "The cluster is highly heterogeneous: it mixes unrelated regressions across models, loading, quantization, tokenization, and trainer behavior. No soft pair looks like the same underlying bug or change, so there is no clear duplicate set.", - "confidence": 0.95, - "canonical_issue_reason": "issue:43873 is the most central/open item with inbound references, but the cluster is too mixed to treat it as a true canonical duplicate target.", + "summary": "The cluster is very heterogeneous: mostly unrelated bugs/regressions spanning loading, quantization, attention, tokenization, docs, and model-specific failures. None of the soft pairs look like the same underlying bug, so all are rejected.", + "confidence": 0.27, + "canonical_issue_reason": "No single canonical duplicate stands out; the issues do not converge on one concrete bug and instead cover many distinct failure modes.", "canonical_pr_reason": null, - "best_issue_reason": "issue:43873 is the best available anchor because it is open and actively referenced, though it only represents one subtopic in an otherwise unrelated set.", + "best_issue_reason": "If forced to anchor the cluster, issue 45698 is the broadest representative of the recurring save/load/custom-module theme, but it is still not a true duplicate of the rest.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43452", "right": "issue:45588", "accept": false, - "reason": "Different bugs: gguf_file loading failures vs flash_attention.py crashing on s_aux=None." + "reason": "Different code paths: tokenizer/gguf loading vs flash-attention sink handling." }, { "left": "issue:43575", "right": "issue:43701", "accept": false, - "reason": "Different failure modes: tensor-parallel OOM for Qwen2-57B vs resume_from_checkpoint key mismatch." + "reason": "Both are regressions, but one is tensor-parallel OOM and the other is a checkpoint key mismatch." }, { "left": "issue:29127", "right": "issue:35141", "accept": false, - "reason": "Both are about confusing initialization behavior, but one is LayoutLMv3 box validation and the other is token-embedding reinit on resize; not the same bug." + "reason": "Unrelated bugs: LayoutLMv3 error messaging vs embedding reinitialization after resize." }, { "left": "issue:43646", "right": "issue:44077", "accept": false, - "reason": "Custom model initialization breakage is distinct from patchtsmixer optional post_init validation." + "reason": "Both mention initialization, but they describe different issues and model-specific constraints." }, { "left": "issue:45072", "right": "issue:45229", "accept": false, - "reason": "Different issues: dtype mismatch in CI vs multi-GPU Gemma4 CUDA OOM." + "reason": "Bfloat16 dtype mismatch in CI vs multi-GPU CUDA OOM are different failure modes." }, { "left": "issue:43575", "right": "issue:44387", "accept": false, - "reason": "Both mention OOM, but one is TP loading of Qwen2-57B and the other is int4 quantization memory growth in Transformers 5.x." + "reason": "Both involve memory pressure, but the causes are different: tp load OOM vs int4 reserved-memory growth." }, { "left": "issue:44977", "right": "issue:45584", "accept": false, - "reason": "Qwen3.5 flash-attention generation regression is unrelated to Whisper empty transcription after align_special_tokens." + "reason": "Different subsystems and symptoms: flash-attention generation vs empty-transcription chat-template failure." }, { "left": "issue:45588", "right": "issue:45600", "accept": false, - "reason": "flash_attention.py s_aux=None crash is unrelated to stale auto_mappings.py references to removed Sam3LiteText configs." + "reason": "Flash-attention sink handling and auto_mappings stale config references are unrelated." }, { "left": "issue:33357", "right": "issue:38175", "accept": false, - "reason": "Both involve SigLIP-like models, but one is a MacOS bus error on CLIP and the other is unexpected zero probabilities on siglip2-base; different symptoms and code paths." + "reason": "Different models and behaviors: MacOS bus error vs zero-probability output." }, { "left": "issue:43316", "right": "issue:43525", "accept": false, - "reason": "Gemma3TextConfig API mismatch is not the same as Llama4Config missing pad_token_id." + "reason": "Both are API discrepancies, but for different configs and different missing fields." }, { "left": "issue:44479", "right": "issue:45290", "accept": false, - "reason": "Video input regression in Qwen VL models is unrelated to chat-template crashes on assistant tool-call messages." + "reason": "Video-input regression in Qwen VL models is unrelated to assistant tool-call handling in chat templates." }, { "left": "issue:44964", "right": "issue:45440", "accept": false, - "reason": "Phi-4 multimodal loading failure and DeepseekV3MoE divergence are different model families and different bugs." + "reason": "Model loading failure for Phi-4 multimodal vs DeepSeekV3 implementation divergence are distinct bugs." }, { "left": "issue:43761", "right": "issue:44079", "accept": false, - "reason": "CLIPVisionModel hidden_states regression is unrelated to ModelOutput key assignment when a key was previously None." + "reason": "Hidden-states omission in CLIPVisionModel is not the same as ModelOutput key assignment logic." }, { "left": "issue:43931", "right": "issue:45216", "accept": false, - "reason": "Qwen3-VL weight shape mismatch is a loading issue, while Qwen3.5 save_pretrained corruption is a checkpoint-saving bug." + "reason": "Both are Qwen3-family load/save issues, but one is a shape-mismatch load error and the other is a bad saved checkpoint." }, { "left": "issue:42907", "right": "issue:45698", "accept": false, - "reason": "Saving dequantized Ministrals/Devstrals is not the same as loading the wrong custom module after save_pretrained." - }, - { - "left": "issue:39692", - "right": "issue:45491", - "accept": false, - "reason": "SigLIP2 docs/example errors are unrelated to Gemma3 NaN embeddings from mixed-length sliding-window batches." + "reason": "Save failure for dequantized models vs wrong custom-module loading after save_pretrained are different save/load bugs." }, { "left": "issue:33290", "right": "issue:34567", "accept": false, - "reason": "Adafactor+deepspeed OOM and TrainerState.num_input_tokens_seen not updating are unrelated trainer/optimizer issues." + "reason": "OOM in deepspeed/adafactor is unrelated to TrainerState token counting." }, { "left": "issue:34689", "right": "issue:43611", "accept": false, - "reason": "Llama 3.2 Vision loading regression is different from base_model_prefix handling in Transformers 5.0.0." + "reason": "Both are loading regressions, but they involve different root causes and model-loading paths." + }, + { + "left": "issue:39692", + "right": "issue:45491", + "accept": false, + "reason": "Documentation example errors are unrelated to Gemma3 NaN embeddings in mixed-length batching." }, { "left": "issue:44442", "right": "issue:44898", "accept": false, - "reason": "FastSpeech2ConformerTokenizer loading failure is unrelated to Perceiver non-default resolution/interpolation failure." + "reason": "Tokenizer loading failure and Perceiver resolution failure are separate model-specific bugs." }, { "left": "issue:43761", "right": "issue:44188", "accept": false, - "reason": "CLIPVisionModel hidden_states regression is unrelated to divergent attention kernels from allow_is_bidirectional_skip under torch.compile." + "reason": "CLIP hidden_states regression is unrelated to attention-kernel branching under torch.compile." }, { "left": "issue:43493", "right": "issue:43827", "accept": false, - "reason": "SigLIP2 implementation discrepancy is unrelated to docs still referencing pipeline() after removals." + "reason": "SigLIP2 implementation mismatch and stale docs referencing pipeline() are unrelated." }, { "left": "issue:43381", "right": "issue:43873", "accept": false, - "reason": "Gradient checkpointing in eval mode and quantization offloading behavior are different code paths and symptoms." + "reason": "Gradient checkpointing in eval mode and quantization offloading are different behaviors." }, { "left": "issue:44568", "right": "issue:45479", "accept": false, - "reason": "Tokenizer special-token addition regression is unrelated to zero-loss classification behavior." + "reason": "Tokenizer special-token behavior is unrelated to the zero-loss classification bug." }, { "left": "issue:43550", "right": "issue:44743", "accept": false, - "reason": "torch.compile+SDPA failure in Bamba is unrelated to recurrent-state reset in modular_qwen3_5.py." + "reason": "torch.compile/SDPA failure in Bamba is not the same as Qwen recurrent-state reset." }, { "left": "issue:43295", "right": "issue:45584", "accept": false, - "reason": "processor.tokenizer regression is unrelated to Whisper empty transcription after align_special_tokens." + "reason": "Processor/tokenizer regression in save/load is unrelated to Whisper empty-transcription generation failure." }, { "left": "issue:43994", "right": "issue:44206", "accept": false, - "reason": "SigLIP2 nonsensical outputs are unrelated to LasrFeatureExtractor passing an unsupported center arg." + "reason": "SigLIP2 inference quality issue and LasrFeatureExtractor crash are different defects." } ] }, @@ -8382,8 +8308,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -8439,8 +8365,8 @@ "soft_pairs": [ "issue:43316|issue:45759", "issue:41628|issue:45372", - "issue:43720|issue:44442", "issue:43643|issue:44913", + "issue:43720|issue:44442", "issue:32090|issue:33290", "issue:44190|issue:44496", "issue:43873|issue:44871", @@ -8456,8 +8382,8 @@ "issue:42175|issue:43316", "issue:43404|issue:43441", "issue:33290|issue:34689", - "issue:44162|issue:44792", "issue:25251|issue:28282", + "issue:44162|issue:44792", "issue:41950|issue:43475", "issue:43232|issue:43388", "issue:43295|issue:43825", @@ -8467,14 +8393,14 @@ "issue:43606|issue:43701", "issue:43295|issue:44871" ], - "prepared_review_unit_hash": "5f37eb78e498b445b52d0d1600773fa0ed878f7d7bf8c4d3a8181fde4ab45350", + "prepared_review_unit_hash": "7c89439a7dd20679f763df775b7ab3684c00bc9edd50dbfeea8717b8992341d9", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, @@ -8489,186 +8415,186 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "The cluster is heterogeneous: it mixes unrelated regressions in config loading, multimodal processors, generation, and distributed training. None of the soft edges look like true duplicates or mergeable PR-equivalents.", - "confidence": 0.92, - "canonical_issue_reason": "If one issue must anchor the cluster, issue 43316 is the best fit because it is a concrete config/API mismatch and is the most central item among the few config-loading similarities in the set.", + "summary": "This set is heterogeneous: it spans unrelated bugs in config loading, generation, multimodal processors, deepspeed, offloading, and training scripts. I don\u2019t see evidence that any soft pair is the same underlying issue, so all candidate duplicate links should be rejected.", + "confidence": 0.96, + "canonical_issue_reason": "No single canonical issue stands out because the items are not one bug cluster; if a representative issue must be chosen, issue:41950 is the clearest open, concrete, and actionable report.", "canonical_pr_reason": null, - "best_issue_reason": "Issue 43316 is the strongest representative of the repeated config/model-loading theme and appears in multiple soft similarities, even though the overall cluster is not a true duplicate set.", + "best_issue_reason": "issue:41950 is the best representative issue only in the sense that it is open, specific, and has a clear failure mode; however, it is not a duplicate of the rest of the set.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43316", "right": "issue:45759", "accept": false, - "reason": "Related to multimodal config handling, but one is a Gemma3TextConfig API discrepancy and the other is a Qwen composite-config unwrap issue; different models and failure modes." + "reason": "Both involve config-related multimodal/model behavior, but they describe different bugs: one is a Gemma3TextConfig API mismatch, the other is from_config not unwrapping text_config for Qwen composite configs." }, { "left": "issue:41628", "right": "issue:45372", "accept": false, - "reason": "Both are import-related, but AutoImageProcessor import failure and missing ReasoningEffort in Gemma 4 processor loading are separate dependency/path problems." + "reason": "Both are import errors, but they stem from different missing symbols and different code paths (AutoImageProcessor vs ReasoningEffort/Gemma 4 processor loading)." }, { - "left": "issue:43720", - "right": "issue:44442", + "left": "issue:43643", + "right": "issue:44913", "accept": false, - "reason": "Packed-weight loading in accelerate versus AutoTokenizer failing to load a specific tokenizer are unrelated code paths." + "reason": "Both concern config fields surviving load/reload, but the reported failures are different: missing fields with trust_remote_code vs rotary_pct resetting on GPTNeoXConfig reload." }, { - "left": "issue:43643", - "right": "issue:44913", + "left": "issue:43720", + "right": "issue:44442", "accept": false, - "reason": "Both concern config objects, but missing fields with trust_remote_code and rotary_pct resetting on reload are different bugs." + "reason": "Unrelated failures: packed BitNet weights during accelerate loading vs AutoTokenizer failing to load a specific tokenizer." }, { "left": "issue:32090", "right": "issue:33290", "accept": false, - "reason": "NoneType GPU broadcast in Trainer and OOM with Adafactor/DeepSpeed are distinct training failures." + "reason": "Both mention Trainer/DeepSpeed, but one is a NoneType broadcast error and the other is an optimizer OOM; these are not the same bug." }, { "left": "issue:44190", "right": "issue:44496", "accept": false, - "reason": "A local dataset loading script issue and an unrecognized model/config.json issue are not the same underlying bug." + "reason": "Different loading failures: local dataset handling in a no-trainer script vs an unrecognized model config missing model_type." }, { "left": "issue:43873", "right": "issue:44871", "accept": false, - "reason": "Quantization/offloading behavior and Gemma eos_token_id config inconsistency are unrelated." + "reason": "One is a quantization/offloading memory behavior issue; the other is an eos_token_id configuration mismatch in Gemma-3." }, { "left": "issue:34689", "right": "issue:36010", "accept": false, - "reason": "One is a Llama vision model-loading regression; the other is a GenerationMixin import error. Same library area, different problem." + "reason": "Both are model-loading/import regressions, but they are distinct problems affecting different APIs and code paths." }, { "left": "issue:43381", "right": "issue:43856", "accept": false, - "reason": "Gradient checkpointing in eval mode and Qwen3 MoE memory inefficiency are different training/runtime issues." + "reason": "Gradient checkpointing in eval mode and Qwen3 MoE memory usage are separate training behaviors, not the same defect." }, { "left": "issue:44077", "right": "issue:44479", "accept": false, - "reason": "patchtsmixer post_init restrictions and Qwen video-input regression are unrelated." + "reason": "Different model/runtime failures: optional post_init handling for patchtsmixer vs video-input regression for Qwen VL models." }, { "left": "issue:44855", "right": "issue:45137", "accept": false, - "reason": "Python 3.13 torch.jit parsing failure and DeepSpeed ZeRO3 deque pop error are separate defects." + "reason": "Python 3.13 decorator/comment parsing failure in DebertaV2Model is unrelated to the DeepSpeed ZeRO3 deque IndexError." }, { "left": "issue:33290", "right": "issue:45656", "accept": false, - "reason": "Both involve DeepSpeed/optimizers, but OOM with Adafactor and optimizer.step being called twice are different bugs." + "reason": "Both involve DeepSpeed optimizer behavior, but one reports OOM with Adafactor while the other reports a duplicated optimizer step call." }, { "left": "issue:43540", "right": "issue:44514", "accept": false, - "reason": "Both hit multimodal/video processing, but Qwen3OmniMoe video-input ValueError and Qwen2_5_VL batched chat-template crash are different code paths." + "reason": "Both are multimodal/video-related, but one is a Qwen3OmniMoe video processing ValueError and the other is a batched apply_chat_template crash in Qwen2_5_VLProcessor." }, { "left": "issue:43316", "right": "issue:44877", "accept": false, - "reason": "Both are config-loading issues, but Gemma3TextConfig API mismatch and strict loading of granite_speech config are not the same bug." + "reason": "Both involve config objects, but they are different bugs: API inconsistency in Gemma3TextConfig vs strict config blocking granite_speech loading." }, { "left": "issue:29127", "right": "issue:33357", "accept": false, - "reason": "LayoutLMv3 box-content error messaging and a MacOS CLIP bus error are unrelated." + "reason": "Completely unrelated issues: LayoutLMv3 error messaging vs a MacOS bus error on a CLIP model." }, { "left": "issue:43381", "right": "issue:43742", "accept": false, - "reason": "Gradient checkpointing eval-mode behavior and MobileLLM loading key error are separate issues." + "reason": "No overlap beyond generic training/loading context; gradient checkpointing eval behavior is unrelated to MobileLLM key errors." }, { "left": "issue:42175", "right": "issue:43316", "accept": false, - "reason": "Pip extras/backend packaging for TensorFlow versus Gemma3TextConfig API discrepancy are unrelated." + "reason": "Install/backend packaging for TensorFlow is unrelated to a Gemma3TextConfig API discrepancy." }, { "left": "issue:43404", "right": "issue:43441", "accept": false, - "reason": "Both are Mistral-family, but weight tying in Mistral3ForConditionalGeneration and FlashAttention failure in Ministral-3 are different concrete problems." + "reason": "Both touch Mistral-family models, but one is tied weights in model conversion/loading and the other is a FlashAttention runtime failure; not the same code-path bug." }, { "left": "issue:33290", "right": "issue:34689", "accept": false, - "reason": "DeepSpeed OOM with Adafactor and Llama 3.2 Vision model loading regression are unrelated." + "reason": "Different failures in different subsystems: DeepSpeed Adafactor OOM vs Llama 3.2 Vision model loading regression." }, { - "left": "issue:44162", - "right": "issue:44792", + "left": "issue:25251", + "right": "issue:28282", "accept": false, - "reason": "ESM2 breakage and janus test failure for image generation do not share the same underlying bug." + "reason": "Pipeline top_k nesting behavior is unrelated to an AutoModel PyTorch import error." }, { - "left": "issue:25251", - "right": "issue:28282", + "left": "issue:44162", + "right": "issue:44792", "accept": false, - "reason": "Pipeline top_k nesting bug and AutoModel missing PyTorch import are completely unrelated." + "reason": "ESM2 breakage and a janus test failure are different model-specific issues." }, { "left": "issue:41950", "right": "issue:43475", "accept": false, - "reason": "Both concern video/image processor behavior, but video-classification using image processors and missing fpn_position_embeddings in SAM3Video are different defects." + "reason": "Both are vision/video model reports, but one is a pipeline processor-selection bug and the other is a missing attribute in Sam3VisionEncoderOutput." }, { "left": "issue:43232", "right": "issue:43388", "accept": false, - "reason": "update_model_kwargs_for_generation after sync_gpus and gather_for_metrics dropping labels are separate distributed-training bugs." + "reason": "Generation kwargs after sync_gpus and metric gathering label truncation are unrelated problems." }, { "left": "issue:43295", "right": "issue:43825", "accept": false, - "reason": "Processor.tokenizer/images regression and pipeline translation-task error messaging are unrelated." + "reason": "Processor/tokenizer regression and pipeline error-message wording are different bugs with different symptoms and fixes." }, { "left": "issue:44871", "right": "issue:45584", "accept": false, - "reason": "Gemma eos_token_id inconsistency and Whisper empty-transcription failure after align_special_tokens are different tokenization/generation issues." + "reason": "Both involve special tokens, but one is a config inconsistency in Gemma-3 and the other is Whisper failing on empty transcription after align_special_tokens." }, { "left": "issue:45127", "right": "issue:45200", "accept": false, - "reason": "LoRA merge collapse with extended vocab and missing mm_token_type_ids default zeros are different multimodal/model-update bugs." + "reason": "LoRA merge with extended vocab causing collapse is unrelated to Gemma 4 requiring mm_token_type_ids defaults." }, { "left": "issue:43550", "right": "issue:43701", "accept": false, - "reason": "torch.compile SDPA failure in Bamba and resume_from_checkpoint key mismatch are unrelated." + "reason": "A torch.compile/SDPA model failure is unrelated to a resume_from_checkpoint key mismatch." }, { "left": "issue:43606", "right": "issue:43701", "accept": false, - "reason": "CPU offload device mismatch and resume_from_checkpoint key mismatch are different issues." + "reason": "CPU offload device mismatch for bark-small is unrelated to the checkpoint key mismatch issue." }, { "left": "issue:43295", "right": "issue:44871", "accept": false, - "reason": "A processor/tokenizer regression and a Gemma eos_token_id config inconsistency do not describe the same bug." + "reason": "These are separate regressions: processor.tokenizer access in custom code vs Gemma-3 eos_token_id inconsistency." } ] }, @@ -8687,8 +8613,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -8755,9 +8681,9 @@ "issue:43575|issue:43873", "issue:44077|issue:44534", "issue:43931|issue:44060", - "issue:44464|issue:44871", "issue:44479|issue:44734", "issue:44617|issue:45584", + "issue:44464|issue:44871", "issue:43065|issue:44610", "issue:43827|issue:43994", "issue:43906|issue:45290", @@ -8771,208 +8697,208 @@ "issue:45127|issue:45357", "issue:43335|issue:44855" ], - "prepared_review_unit_hash": "04003b66e72897fde8018c127330958b5debe6cbb1ffcd0b732f71877076ef7b", + "prepared_review_unit_hash": "8424c1f4cb75b643a62837b83d6e81d61b5e4e1177be761e74dafa34ad37bdb4", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, "soft_pair_count": 29, - "serialized_chars": 35846, - "estimated_input_tokens": 8962, - "estimated_eval_tokens": 18180 + "serialized_chars": 35843, + "estimated_input_tokens": 8961, + "estimated_eval_tokens": 18178 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is highly heterogeneous: the paired issues are mostly different regressions or unrelated model-loading/training bugs. I found no soft-edge pair that clearly represents the same underlying defect or change, so all candidate duplicate links should be rejected.", - "confidence": 0.86, - "canonical_issue_reason": "If a single issue must serve as the anchor, issue #43646 is the broadest and most central regression description, but the cluster overall does not form a coherent duplicate set.", + "summary": "This cluster is a heterogeneous mix of unrelated bugs/regressions (OOMs, config/load failures, generation/cache issues, multimodal model regressions, and docs/CI problems). I do not see a meaningful duplicate set or a single canonical issue for the group.", + "confidence": 0.97, + "canonical_issue_reason": "No single canonical issue: the items span distinct code paths and failure modes, so the cluster looks like a false-positive mix rather than one underlying bug.", "canonical_pr_reason": null, - "best_issue_reason": "#43646 is the best representative because it is the most general and can subsume several initialization/loading regressions, even though most items here are not true duplicates of it.", + "best_issue_reason": "No global best issue is suitable because there is no coherent duplicate family to represent; the issues are too diverse in subsystem and root cause.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43673", "right": "issue:44743", "accept": false, - "reason": "Both involve cache-related behavior, but one is a missing GenerationMixin cache in chunked prefill while the other resets recurrent states in modular_qwen3_5; different code paths and symptoms." + "reason": "Both involve cache/state behavior, but one is a v5 chunked-prefill cache missing bug and the other is Qwen3.5 recurrent-state reset with seq_len>1; different model paths and causes." }, { "left": "issue:43761", "right": "issue:44479", "accept": false, - "reason": "CLIPVisionModel hidden_states regression and Qwen video-input regression are unrelated model families with different failure modes." + "reason": "Both are multimodal regressions, but CLIPVision hidden_states handling and Qwen video input position handling are different bugs in different models." }, { "left": "issue:33290", "right": "issue:35141", "accept": false, - "reason": "OOM with Adafactor/DeepSpeed is unrelated to embedding reinitialization in post_init." + "reason": "OOM with Adafactor/DeepSpeed is unrelated to token embedding reinitialization after resize and post_init." }, { "left": "issue:33290", "right": "issue:36331", "accept": false, - "reason": "Optimizer memory blowup and CustomTrainer compute_loss signature mismatch are distinct bugs." + "reason": "Memory blowup in training infrastructure is not the same as a CustomTrainer compute_loss signature mismatch." }, { "left": "issue:43065", "right": "issue:44617", "accept": false, - "reason": "Dummy Conv2d in Sam3PixelDecoder is a model-implementation issue; Sam3Video OOM is a runtime memory issue, not the same defect." + "reason": "Same model family, but one is a dummy Conv2d implementation issue and the other is a CUDA OOM during video inference; not the same fix." }, { "left": "issue:43452", "right": "issue:44496", "accept": false, - "reason": "Both concern loading, but gguf_file handling and missing model_type in config.json are different root causes." + "reason": "Both are model-loading failures, but one is gguf/AutoTokenizer-AutoModel handling and the other is missing model_type in config.json." }, { "left": "issue:43575", "right": "issue:43742", "accept": false, - "reason": "Qwen2 large-model OOM with tensor parallelism and a KeyError loading MobileLLM are separate loading failures." + "reason": "Both fail on loading, but one is TP-related OOM for Qwen2-57B and the other is a config/key error for MobileLLM." }, { "left": "issue:44936", "right": "issue:45698", "accept": false, - "reason": "Trainer evaluate-after-train behavior is unrelated to from_pretrained choosing the wrong custom module after save_pretrained." + "reason": "trainer.evaluate after train and wrong custom module reload after save_pretrained are separate lifecycle bugs." }, { "left": "issue:43646", "right": "issue:44913", "accept": false, - "reason": "Generic custom model initialization breakage is not the same as GPTNeoXConfig rotary_pct not persisting on reload." + "reason": "Transformers 5 custom model init breakage is unrelated to GPTNeoX rotary_pct not persisting on reload." }, { "left": "issue:42617", "right": "issue:43575", "accept": false, - "reason": "3d_parallel.py execution failure and Qwen2 model-load OOM are different problems." + "reason": "3d_parallel.py execution failure is unrelated to Qwen2 TP OOM during model loading." }, { "left": "issue:43761", "right": "issue:44242", "accept": false, - "reason": "Hidden_states regression in CLIPVisionModel is unrelated to load-balancing loss omission when output_router_logits=False." + "reason": "CLIPVision hidden_states regression and missing load-balancing loss are different model components and failure modes." }, { "left": "issue:43575", "right": "issue:43873", "accept": false, - "reason": "Both mention memory, but tensor-parallel OOM and offloading/quantization misbehavior are not the same underlying bug." + "reason": "Both mention loading/memory, but TP OOM and quantization offloading not working are distinct problems." }, { "left": "issue:44077", "right": "issue:44534", "accept": false, - "reason": "PatchTSMixer post_init policy and non-persistent buffer junk in v5 are unrelated regressions." + "reason": "Optional post_init restriction in patchtsmixer is unrelated to junk-filled non-persistent buffers in v5." }, { "left": "issue:43931", "right": "issue:44060", "accept": false, - "reason": "Shape mismatch when loading Qwen3-VL is different from an incorrect tied-weights warning in Qwen3-Next." - }, - { - "left": "issue:44464", - "right": "issue:44871", - "accept": false, - "reason": "Compiled-forward chunked generation inconsistency and Gemma eos_token_id config mismatch are unrelated issues." + "reason": "Weight shape mismatch in Qwen3-VL and incorrect tied-weights warning in Qwen3-Next are different bugs." }, { "left": "issue:44479", "right": "issue:44734", "accept": false, - "reason": "Video-input regression in Qwen multimodal models is different from a KV-cache continuation indexing bug in transformers serve." + "reason": "Video-input regression in Qwen multimodal models is unrelated to KV-cache continuation tensor indexing in transformers serve." }, { "left": "issue:44617", "right": "issue:45584", "accept": false, - "reason": "Sam3Video OOM and Whisper empty-transcription failure are unrelated model-specific bugs." + "reason": "Sam3Video CUDA OOM and Whisper empty-transcription failure after align_special_tokens are unrelated." + }, + { + "left": "issue:44464", + "right": "issue:44871", + "accept": false, + "reason": "Chunked generation inconsistency with compiled forward is unrelated to Gemma eos_token_id config mismatch." }, { "left": "issue:43065", "right": "issue:44610", "accept": false, - "reason": "Dummy Conv2d in Sam3PixelDecoder is not the same as OmDet-Turbo processor output-size mismatch." + "reason": "Sam3PixelDecoder dummy layer issue is unrelated to OmDet-Turbo processor/model input-size mismatch." }, { "left": "issue:43827", "right": "issue:43994", "accept": false, - "reason": "Docs still referencing pipeline() and a SigLIP2 model producing nonsensical outputs are different issues." + "reason": "Docs still referencing pipeline() and Siglip2 nonsensical outputs are different issues." }, { "left": "issue:43906", "right": "issue:45290", "accept": false, - "reason": "A reproduction of issue #38071 is not the same as apply_chat_template crashing on assistant tool-call messages." + "reason": "A reproduction of issue 38071 is not the same as apply_chat_template crashing on assistant tool-call messages." }, { "left": "issue:44792", "right": "issue:44977", "accept": false, - "reason": "Janus image-generation test failure and Qwen3.5 flash-attention generation issues do not match." + "reason": "Janus image-generation test failure and Qwen3.5 flash-attention generation bug are unrelated." }, { "left": "issue:44534", "right": "issue:44913", "accept": false, - "reason": "Non-persistent buffers filled with junk and GPTNeoX rotary_pct reloading behavior are distinct bugs." + "reason": "Non-persistent buffer junk in v5 and GPTNeoX rotary_pct reload default are distinct configuration/runtime bugs." }, { "left": "issue:41720", "right": "issue:45245", "accept": false, - "reason": "Auto device mapping CUDA assert and category-count limit runtime error are unrelated." + "reason": "Qwen3 auto device mapping cudaErrorAssert is unrelated to the 2^24 category limit RuntimeError." }, { "left": "issue:44190", "right": "issue:44964", "accept": false, - "reason": "Local dataset loading in a training script and loading Phi-4 multimodal with latest transformers are not duplicates." + "reason": "Local dataset loading in no-trainer image classification and Phi-4 multimodal model loading are different paths." }, { "left": "issue:44792", "right": "issue:45600", "accept": false, - "reason": "Janus generation test failure and stale auto_mappings references to removed Sam3LiteText configs are unrelated." + "reason": "Janus test failure is unrelated to stale auto_mappings references for removed Sam3LiteText configs." }, { "left": "issue:44855", "right": "issue:45588", "accept": false, - "reason": "Python 3.13 import/parsing failure in DebertaV2Model is unrelated to flash_attention.py crashing on s_aux=None." + "reason": "Python 3.13 IndentationError in DebertaV2 import is unrelated to flash_attention's s_aux=None crash." }, { "left": "issue:44521", "right": "issue:45381", "accept": false, - "reason": "Both involve multimodal chat/vision handling, but all-zero assistant masks and wrong vision_position_ids are different defects." + "reason": "Both are multimodal/chat-template issues, but all-zero assistant masks and wrong vision_position_ids are different concrete bugs." }, { "left": "issue:45127", "right": "issue:45357", "accept": false, - "reason": "LoRA merge collapse with extended vocab and incorrect visual encoder keys saved in Qwen3.5 are different save/load behaviors." + "reason": "LoRA merge collapse with extended vocab is unrelated to incorrect visual encoder keys saved by Qwen3.5." }, { "left": "issue:43335", "right": "issue:44855", "accept": false, - "reason": "SwitchTransformers sparse-layer config bug and DebertaV2Model Python 3.13 import error are unrelated." + "reason": "SwitchTransformers sparse-layer creation and a Python 3.13 DebertaV2 import parse failure are unrelated." } ] }, @@ -8991,8 +8917,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -9047,14 +8973,14 @@ ], "soft_pairs": [ "issue:44938|issue:45542", - "issue:44077|issue:44877", "issue:43976|issue:44855", + "issue:44077|issue:44877", "issue:43644|issue:44913", "issue:43550|issue:44190", "issue:43761|issue:44743", "issue:43540|issue:44734", - "issue:43650|issue:44220", "issue:45276|issue:45584", + "issue:43650|issue:44220", "issue:44112|issue:45600", "issue:43116|issue:44190", "issue:43295|issue:45276", @@ -9077,20 +9003,20 @@ "issue:43643|issue:44496", "issue:45563|issue:45584" ], - "prepared_review_unit_hash": "5176cd372fc090325331bafb5ee6bb835457f4edc6a7b9f7b7bc667e647c0201", + "prepared_review_unit_hash": "5b0a488241905aeea3224192b954b039470a9e9718dea880e65c69d4e9af6f38", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 30, - "serialized_chars": 36767, + "serialized_chars": 36765, "estimated_input_tokens": 9192, "estimated_eval_tokens": 18640 }, @@ -9099,186 +9025,168 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is highly heterogeneous: most items are unrelated Transformers bugs across different models, APIs, or runtime environments. No soft pair looks like the same underlying bug or change, so I would not merge any of them as duplicates.", + "summary": "This set is highly heterogeneous: the soft pairs mostly share only broad themes like imports, configs, or regressions, but they do not describe the same concrete bug/change. I would not merge any of the proposed pairs.", "confidence": 0.89, - "canonical_issue_reason": "issue:44246 is the broadest, highest-activity issue in the set and best represents the cluster if a single issue must be chosen, but the group is not actually duplicate-heavy.", + "canonical_issue_reason": null, "canonical_pr_reason": null, - "best_issue_reason": "issue:44246 is the best standalone issue candidate because it has the strongest discussion signal and concerns core import-time behavior, which is more central than the other narrowly scoped reports.", + "best_issue_reason": "No issue is a strong duplicate hub for this set. If a single representative is needed, issue 44246 is the broadest user-facing runtime/reporting bug with active discussion, but it is not a true stand-in for the others.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:44938", "right": "issue:45542", "accept": false, - "reason": "Both involve loading/runtime behavior, but one is a Python 3.14 import/load failure and the other is a TensorBoard-without-TF backend error; different code paths." + "reason": "Different failures: Python 3.14 import/load problem vs missing TensorFlow backend when only tensorboard is installed." }, { - "left": "issue:44077", - "right": "issue:44877", + "left": "issue:43976", + "right": "issue:44855", "accept": false, - "reason": "Both mention config validation, but they target different models and different misconfigurations; not the same bug." + "reason": "Both are Python-version-related import errors, but they involve different models, symptoms, and root causes." }, { - "left": "issue:43976", - "right": "issue:44855", + "left": "issue:44077", + "right": "issue:44877", "accept": false, - "reason": "Both are Python-version-related import failures, but one is a broad version incompatibility and the other is a DebertaV2Model parsing error in Python 3.13." + "reason": "Both touch config validation, but one is about optional post_init on patchtsmixer and the other about loading granite_speech config." }, { "left": "issue:43644", "right": "issue:44913", "accept": false, - "reason": "One is a serialization/buffer regression and the other is a config field silently reverting on reload; different failure modes." + "reason": "One is a non-persistent buffer initialization regression; the other is a config field not persisting on reload." }, { "left": "issue:43550", "right": "issue:44190", "accept": false, - "reason": "These affect completely different workflows: torch.compile/SDPA model inference versus loading a local dataset in an example script." + "reason": "Completely different areas: SDPA/torch.compile on a model vs example script dataset loading." }, { "left": "issue:43761", "right": "issue:44743", "accept": false, - "reason": "Both are model-generation correctness issues, but they involve different models and unrelated cache/state logic." + "reason": "Different models and different bugs: CLIP hidden_states omission vs recurrent state reset with cache." }, { "left": "issue:43540", "right": "issue:44734", "accept": false, - "reason": "One is a Qwen3OmniMoe video-input processing error; the other is a transformers serve KV-cache indexing crash." + "reason": "Video input validation in Qwen3OmniMoe is unrelated to serve KV-cache continuation indexing." }, { - "left": "issue:43650", - "right": "issue:44220", + "left": "issue:45276", + "right": "issue:45584", "accept": false, - "reason": "The titles point to unrelated problems: missing data versus an fbank feature extraction issue." + "reason": "Token embedding resize wiring vs Whisper empty-transcription generation failure are distinct issues." }, { - "left": "issue:45276", - "right": "issue:45584", + "left": "issue:43650", + "right": "issue:44220", "accept": false, - "reason": "One is embedding weight propagation after resizing token embeddings; the other is Whisper empty-transcription generation after token alignment." + "reason": "The titles are unrelated; one is not a real technical bug description and the other is a feature-specific audio feature extraction issue." }, { "left": "issue:44112", "right": "issue:45600", "accept": false, - "reason": "Both mention CI, but one is a stale device-override test and the other is a broken auto_mappings reference; different root causes." + "reason": "Both mention CI, but they are different failures in different code paths and different models." }, { "left": "issue:43116", "right": "issue:44190", "accept": false, - "reason": "Different example scripts and different symptoms; no shared concrete bug." + "reason": "Example-script multi-label classification output bug vs local dataset loading failure are not the same problem." }, { "left": "issue:43295", "right": "issue:45276", "accept": false, - "reason": "One is a processor/tokenizer regression; the other is a resize-token-embeddings model-weights issue." + "reason": "Processor/tokenizer regression is unrelated to token embedding resizing behavior." }, { "left": "issue:43335", "right": "issue:44190", "accept": false, - "reason": "SwitchTransformers sparse-layer construction is unrelated to local dataset loading in an example script." + "reason": "SwitchTransformers config layer construction bug is unrelated to the image-classification dataset loader issue." }, { "left": "issue:45245", "right": "issue:45584", "accept": false, - "reason": "A category-cardinality RuntimeError is unrelated to Whisper generation after align_special_tokens." + "reason": "Generic category-count runtime error is different from Whisper post-processing/empty transcription handling." }, { "left": "issue:36246", "right": "issue:41628", "accept": false, - "reason": "Both are import errors, but they concern different symbols and different package exports; too broad to be duplicates." + "reason": "Both are import errors, but they expose different missing symbols from different modules and likely different regressions." }, { "left": "issue:43404", "right": "issue:45216", "accept": false, - "reason": "Different model families and different symptoms: untied lm_head weights versus incorrect checkpoint saving." + "reason": "Weight tying in Mistral3 is a different bug from incorrect Qwen3.5 checkpoint saving." }, { "left": "issue:43479", "right": "issue:45375", "accept": false, - "reason": "Both are config-field issues, but they involve different models and different missing/defaulted fields." + "reason": "Both involve multimodal config fields, but they affect different models and different missing/defaulted fields." }, { "left": "issue:44933", "right": "issue:45362", "accept": false, - "reason": "A missing image_utils import is not the same as Qwen3.5-35B crashing in chat." + "reason": "A missing image_utils import is unrelated to Qwen3.5 chat crashes." }, { "left": "issue:44734", "right": "issue:45632", "accept": false, - "reason": "Serve/KV-cache indexing and trust_remote_code cache-path collisions are unrelated problems." + "reason": "Serve KV-cache indexing crash and trust_remote_code cache-path collision are unrelated." }, { "left": "issue:44464", "right": "issue:45563", "accept": false, - "reason": "One is a correctness issue in chunked generation with compiled forward; the other is a stale warning in paged generate." + "reason": "Compiled forward generation inconsistency is not the same as a stale warning about num_return_sequences." }, { "left": "issue:44186", "right": "issue:44734", "accept": false, - "reason": "A tokenizer crash on NER inputs is not the same underlying bug as a serve-side tensor indexing crash." + "reason": "Tokenizer padding/NER crash is unrelated to serve-side KV-cache tensor indexing." }, { "left": "issue:38175", "right": "issue:43065", "accept": false, - "reason": "Siglip2 zero probabilities and a dummy Conv2d in Sam3PixelDecoder are unrelated model bugs." + "reason": "Zero probabilities in SigLIP2 and dummy Conv2d in Sam3PixelDecoder are entirely different model bugs." }, { "left": "issue:44561", "right": "issue:45632", "accept": false, - "reason": "Both involve trust_remote_code, but one is a removed utility breaking models and the other is a local cache-path collision." + "reason": "Removal of is_torch_fx_available affects remote-code imports broadly; cache-path collisions are a separate trust_remote_code bug." }, { "left": "issue:44683", "right": "issue:45446", "accept": false, - "reason": "Both are in flex_attention, but one is a compiled-runtime failure on newer torch and the other is an incorrect version check for AuxRequest import; different concrete bugs." + "reason": "Torch >= 2.9 flex_attention failure and incorrect version check for AuxRequest import are different flex_attention issues." }, { "left": "issue:44466", "right": "issue:44913", "accept": false, - "reason": "Serialization of tied lm_head weights is unrelated to GPTNeoX rotary_pct reloading incorrectly." + "reason": "Serialization of tied lm_head weights is unrelated to GPTNeoX rotary_pct reload behavior." }, { "left": "issue:44246", "right": "issue:45003", "accept": false, - "reason": "Import slowness and unsafe sys.modules access are only vaguely related; the titles do not indicate the same bug or fix." - }, - { - "left": "issue:45406", - "right": "issue:45632", - "accept": false, - "reason": "Gemma4Processor missing _tokenizer in transformers serve is unrelated to trust_remote_code cache collisions." - }, - { - "left": "issue:43643", - "right": "issue:44496", - "accept": false, - "reason": "A trust_remote_code config object missing fields is not the same as an unrecognized model due to missing model_type." - }, - { - "left": "issue:45563", - "right": "issue:45584", - "accept": false, - "reason": "Both are generation-related, but one is a warning bug and the other is an actual Whisper failure path." + "reason": "Import slowdowns and unsafe sys.modules access may be related in spirit, but they are not the same concrete bug." } ] }, @@ -9297,8 +9205,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -9362,8 +9270,8 @@ "issue:44016|issue:44220", "issue:43673|issue:44336", "issue:43299|issue:45216", - "issue:43531|issue:45600", "issue:42175|issue:44279", + "issue:43531|issue:45600", "issue:43526|issue:43901", "issue:45335|issue:45584", "issue:44075|issue:44368", @@ -9372,8 +9280,8 @@ "issue:44496|issue:45759", "issue:41950|issue:45020", "issue:43901|issue:44496", - "issue:44016|issue:44062", "issue:44242|issue:45200", + "issue:44016|issue:44062", "issue:43577|issue:45216", "issue:43976|issue:45405", "issue:44351|issue:45278", @@ -9382,214 +9290,214 @@ "issue:43479|issue:43525", "issue:44683|issue:45542" ], - "prepared_review_unit_hash": "bbf5891558fe2f0cb875d59460942d80c0b8712b3cf0fe873fdfa799328dcf79", + "prepared_review_unit_hash": "cd2e966ad2fb706b9e7214bbaa23986451317590debb6c1e0c4cf1ed98d2c8b5", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, "soft_pair_count": 30, - "serialized_chars": 36067, - "estimated_input_tokens": 9017, - "estimated_eval_tokens": 18290 + "serialized_chars": 36062, + "estimated_input_tokens": 9016, + "estimated_eval_tokens": 18288 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "The cluster is mostly a set of unrelated Transformers regression reports that only share broad subsystem or model-family similarity. I found no soft-edge pair that clearly represents the same concrete bug or change, so I would not merge any of the candidates as duplicates.", - "confidence": 0.79, - "canonical_issue_reason": "Issue 45020 is the best umbrella candidate: it is open, broad, and captures the recurring theme of recent-version regressions affecting model loading and remote-code paths better than the narrower one-off reports.", + "summary": "The cluster is mostly a mix of unrelated Transformers regressions, import errors, and docs/test issues. None of the soft-pair candidates look like true duplicates.", + "confidence": 0.84, + "canonical_issue_reason": "issue:45020 is the broadest representative report here: a general regression affecting remote_code model loading across recent Transformers versions.", "canonical_pr_reason": null, - "best_issue_reason": "45020 is the most representative issue for the cluster, but it is still too broad to serve as a true duplicate target for many of the more specific model/config regressions.", + "best_issue_reason": "issue:45020 is the best umbrella issue in this set because it captures the widest upstream regression surface.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:41950", "right": "issue:44016", "accept": false, - "reason": "Different bugs: one is a pipeline processor lookup problem, the other is a notebook syntax error." + "reason": "Different problems: video pipeline processor lookup vs notebook syntax error." }, { "left": "issue:44964", "right": "issue:45600", "accept": false, - "reason": "Unrelated model-load failure vs stale auto-mappings reference; same repo area, different concrete issue." + "reason": "Different model-loading bug vs stale auto_mappings CI references." }, { "left": "issue:42907", "right": "issue:43441", "accept": false, - "reason": "Both involve Ministral, but one is save/export of dequantized models and the other is FlashAttention loading; not the same bug." + "reason": "Different code paths: save/dequantization vs FlashAttention model loading." }, { "left": "issue:43388", "right": "issue:43688", "accept": false, - "reason": "Distinct training/evaluation issues affecting different code paths and model families." + "reason": "Metrics batching bug vs MoE auxiliary-loss normalization; unrelated." }, { "left": "issue:43295", "right": "issue:45335", "accept": false, - "reason": "Both are regressions, but one is a processor/tokenizer API break and the other is a resize_token_embeddings bug in t5gemma." + "reason": "Processor/tokenizer regression vs t5gemma embedding resize bug." }, { "left": "issue:44112", "right": "issue:44964", "accept": false, - "reason": "CI device-override test failure is not the same as a Phi-4 multimodal loading regression." + "reason": "CI device override test issue vs Phi-4 multimodal loading failure." }, { "left": "issue:44496", "right": "issue:45042", "accept": false, - "reason": "Different failures: generic model-type resolution vs PIL image processor torchvision dependency regression." + "reason": "Unrecognized model/config issue vs PIL image processor torchvision dependency bug." }, { "left": "issue:43854", "right": "issue:44964", "accept": false, - "reason": "Both are model-loading issues, but they involve different models and different failure modes." + "reason": "Both are load failures, but for different models and different underlying causes." }, { "left": "issue:44016", "right": "issue:44220", "accept": false, - "reason": "Notebook syntax error and fbank feature extraction bug are unrelated." + "reason": "Notebook syntax error is unrelated to audio feature extraction failure." }, { "left": "issue:43673", "right": "issue:44336", "accept": false, - "reason": "Generation cache regression and ANSI output formatting are unrelated." + "reason": "Cache/memory issue vs ANSI terminal reporting bug; unrelated." }, { "left": "issue:43299", "right": "issue:45216", "accept": false, - "reason": "Both touch Qwen models, but one is MoE loading and the other is save_pretrained serialization." + "reason": "Qwen loading regression vs Qwen save_pretrained regression; different concrete bugs." }, { - "left": "issue:43531", - "right": "issue:45600", + "left": "issue:42175", + "right": "issue:44279", "accept": false, - "reason": "Sliding-window behavior in Qwen3-MoE is unrelated to stale auto_mappings references." + "reason": "Backend dependency packaging issue vs generic dependency problem; too broad to merge." }, { - "left": "issue:42175", - "right": "issue:44279", + "left": "issue:43531", + "right": "issue:45600", "accept": false, - "reason": "Package/backend dependency complaints, but not the same dependency or code path." + "reason": "Qwen3-MoE sliding window bug vs removed Sam3LiteText config references." }, { "left": "issue:43526", "right": "issue:43901", "accept": false, - "reason": "Label reduction in BeitImageProcessorFast is unrelated to a pipeline doc mismatch." + "reason": "ImageProcessorFast label reduction bug vs documentation mismatch." }, { "left": "issue:45335", "right": "issue:45584", "accept": false, - "reason": "Token embedding resize bug and Whisper empty-transcription generation failure are different code paths." + "reason": "Embedding resize bug vs Whisper empty-transcription generation failure." }, { "left": "issue:44075", "right": "issue:44368", "accept": false, - "reason": "Optimizer argument handling and a config warning about tied embeddings are not the same defect." + "reason": "Optimizer argument handling vs LoRA warning about tie_word_embeddings." }, { "left": "issue:43381", "right": "issue:45593", "accept": false, - "reason": "Gradient checkpointing eval-mode behavior and D-FINE auxiliary-loss handling are unrelated." + "reason": "Gradient checkpointing eval-mode bug vs D-FINE auxiliary-loss behavior." }, { "left": "issue:43299", "right": "issue:45357", "accept": false, - "reason": "Same model family, but the first is a load regression and the second is a save_pretrained key serialization bug; not clearly the same fix." + "reason": "Loading failure vs save_pretrained checkpoint-key regression; not same bug." }, { "left": "issue:44496", "right": "issue:45759", "accept": false, - "reason": "Both involve model/config resolution, but one is an unrecognized model error and the other is unwrapping text_config for multimodal configs." + "reason": "Unrecognized model registration vs from_config text_config unwrapping; related area but different defect." }, { "left": "issue:41950", "right": "issue:45020", "accept": false, - "reason": "Broad Transformers regressions, but one is a specific pipeline processor lookup bug and the other is a generic remote_code complaint." + "reason": "General version-regression theme only; different subsystems and concrete failures." }, { "left": "issue:43901", "right": "issue:44496", "accept": false, - "reason": "Pipeline docs mismatch and model type recognition failure are unrelated." + "reason": "Docs behavior note vs model-loading config error; unrelated." }, { - "left": "issue:44016", - "right": "issue:44062", + "left": "issue:44242", + "right": "issue:45200", "accept": false, - "reason": "Notebook syntax error and AddedToken keyword conflict are unrelated." + "reason": "MoE load-balancing loss bug vs Gemma 4 mm_token_type_ids defaulting issue." }, { - "left": "issue:44242", - "right": "issue:45200", + "left": "issue:44016", + "right": "issue:44062", "accept": false, - "reason": "Different model/training features: router-logit load-balancing loss vs Gemma 4 multimodal token type IDs." + "reason": "Notebook syntax error vs AddedToken keyword-argument conflict." }, { "left": "issue:43577", "right": "issue:45216", "accept": false, - "reason": "BLIP-2 dtype propagation and Qwen3.5 save_pretrained serialization are unrelated." + "reason": "BLIP2 dtype propagation vs Qwen save_pretrained checkpoint regression." }, { "left": "issue:43976", "right": "issue:45405", "accept": false, - "reason": "Python version compatibility and an unreleased dependency pin are different release-management problems." + "reason": "Python support breakage vs unreleased PEFT version bump." }, { "left": "issue:44351", "right": "issue:45278", "accept": false, - "reason": "HybridCache missing import is a specific symbol regression; 45278 is a broad import-errors report, not the same concrete bug." + "reason": "Specific HybridCache import failure vs broad many-import-errors report; not the same bug." }, { "left": "issue:43824", "right": "issue:44933", "accept": false, - "reason": "Specific Qwen2.5-VL import failure vs a different missing image_utils import; too different to merge." + "reason": "Different missing imports from different modules and causes." }, { "left": "issue:43388", "right": "issue:43827", "accept": false, - "reason": "Metric gathering logic bug and docs referencing removed pipeline APIs are unrelated." + "reason": "Training metric batching bug vs docs still referencing removed pipeline API." }, { "left": "issue:43479", "right": "issue:43525", "accept": false, - "reason": "Phi-4 multimodal config defaulting and Llama4 missing pad_token_id are different config bugs." + "reason": "Different config initialization issue vs missing pad_token_id attribute." }, { "left": "issue:44683", "right": "issue:45542", "accept": false, - "reason": "Compiled flex_attention on torch 2.9 and TensorBoard/TensorFlow backend packaging are unrelated." + "reason": "Compiled flex_attention torch-version bug vs TensorBoard/TensorFlow packaging issue." } ] }, @@ -9608,8 +9516,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -9667,10 +9575,10 @@ "issue:44279|issue:45405", "issue:43299|issue:43854", "issue:39692|issue:43901", - "issue:41950|issue:43650", "issue:32090|issue:35141", - "issue:44297|issue:44462", + "issue:41950|issue:43650", "issue:43295|issue:44779", + "issue:44297|issue:44462", "issue:38175|issue:44617", "issue:43976|issue:44279", "issue:44462|issue:44779", @@ -9684,8 +9592,8 @@ "issue:44261|issue:44263", "issue:43441|issue:44683", "issue:43650|issue:45290", - "issue:43723|issue:44162", "issue:44623|issue:45397", + "issue:43723|issue:44162", "issue:43784|issue:45542", "issue:44623|issue:44749", "issue:44671|issue:44779", @@ -9695,220 +9603,220 @@ "issue:43122|issue:43650", "issue:44361|issue:45593" ], - "prepared_review_unit_hash": "e3939c2da6e72935c928f81185ecef00ad12ad1047fea4b2f080bff2b2f5e883", + "prepared_review_unit_hash": "50c5b93e896db6ce1eb9065d8b60b2ebe2ca3b4b652291773692f750d61fa29d", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 31, - "serialized_chars": 36959, - "estimated_input_tokens": 9240, - "estimated_eval_tokens": 18736 + "serialized_chars": 36962, + "estimated_input_tokens": 9241, + "estimated_eval_tokens": 18738 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "Mostly a heterogeneous set of tokenizer-loading/regression, multimodal, quantization, and model-specific bugs. The only common thread is loose v5/tokenizer similarity, but the concrete failure modes differ, so no soft edges are accepted.", - "confidence": 0.34, - "canonical_issue_reason": "issue:44462 is the clearest concrete tokenizer-loading regression in this set and the best umbrella representative for the loose tokenizer/v5 cluster, though the overall group is still quite heterogeneous.", + "summary": "This looks like a loose collection of Transformers regressions, with several tokenizer/model-loading and multimodal-processing complaints, but the proposed soft edges mostly share only broad subsystem overlap rather than the same concrete bug. No pair here is a safe duplicate merge.", + "confidence": 0.63, + "canonical_issue_reason": "issue:43723 is the broadest and most central tokenizer-loading regression in the set; it can anchor the nearby tokenizer/processor complaints better than the model-specific or unrelated issues.", "canonical_pr_reason": null, - "best_issue_reason": "issue:44462 has the most representative and actionable failure mode: AutoTokenizer ignoring repository tokenizer.json. It is specific enough to anchor the tokenizer-related subset better than the more generic or highly model-specific reports.", + "best_issue_reason": "issue:43723 is the clearest representative issue: generic, regression-oriented, and closer in scope to several tokenizer-related reports than the others.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:44462", "right": "issue:45081", "accept": false, - "reason": "Both involve tokenizers, but one is AutoTokenizer repository loading while the other is a Mistral regex patch crash; different code paths and fixes." + "reason": "Both concern tokenizer loading, but one is AutoTokenizer ignoring repository tokenizer.json while the other is a Mistral regex patch crash; different failure points and fixes." }, { "left": "issue:44279", "right": "issue:45405", "accept": false, - "reason": "Unrelated: generic dependency trouble vs a specific PEFT minimum-version bump." + "reason": "Unrelated topics: a dependency problem vs a bumped PEFT minimum version. No shared bug." }, { "left": "issue:43299", "right": "issue:43854", "accept": false, - "reason": "Different models and failures: Qwen3VL loading regression vs GLM-4.7-Flash unit-test load issue." + "reason": "Both are model-loading failures, but they target different models and likely different load-time issues." }, { "left": "issue:39692", "right": "issue:43901", "accept": false, - "reason": "Both mention docs/examples, but one is a SigLIP2 example failure and the other is a pipeline docs mismatch; not the same bug." - }, - { - "left": "issue:41950", - "right": "issue:43650", - "accept": false, - "reason": "Video pipeline processor lookup vs an unrelated placeholder issue; no shared underlying defect." + "reason": "One is a SigLIP2 example/model mismatch and quantization issue; the other is a docs mismatch about return_all_scores. Not the same bug." }, { "left": "issue:32090", "right": "issue:35141", "accept": false, - "reason": "Trainer GPU broadcast NoneType error and embedding reinitialization are unrelated." + "reason": "Trainer gpu broadcast TypeError vs output embedding reinit after resizing embeddings. Different code paths and symptoms." }, { - "left": "issue:44297", - "right": "issue:44462", + "left": "issue:41950", + "right": "issue:43650", "accept": false, - "reason": "Both are tokenizer-related, but one is save_pretrained metadata mismatch and the other is AutoTokenizer ignoring tokenizer.json; distinct root causes." + "reason": "Video-classification pipeline processor lookup vs a placeholder issue title; no meaningful overlap." }, { "left": "issue:43295", "right": "issue:44779", "accept": false, - "reason": "A processor.tokenizer regression with image inputs is not the same as Deepseek tokenization drift." + "reason": "Both mention tokenizer behavior, but one is a processor/tokenizer API regression and the other is Deepseek tokenization correctness; too model-specific to merge." + }, + { + "left": "issue:44297", + "right": "issue:44462", + "accept": false, + "reason": "Related to tokenizer save/load metadata, but one is tokenizer_class mismatch on save and the other is AutoTokenizer ignoring tokenizer.json on load." }, { "left": "issue:38175", "right": "issue:44617", "accept": false, - "reason": "SigLIP2 zero probabilities and Sam3Video OOM are unrelated model/runtime issues." + "reason": "SigLIP2 zero probabilities and Sam3Video CUDA OOM are completely different issues." }, { "left": "issue:43976", "right": "issue:44279", "accept": false, - "reason": "Python version compatibility break vs generic dependency issue; not the same concrete bug." + "reason": "Python-version support breakage vs a generic dependency issue; not the same underlying bug." }, { "left": "issue:44462", "right": "issue:44779", "accept": false, - "reason": "Both concern tokenizers, but one is repository tokenizer selection/loading and the other is Deepseek tokenization correctness; too different to merge." + "reason": "Both are tokenizer-related, but one is repository tokenizer.json being ignored and the other is incorrect Deepseek tokenization results; different regressions." }, { "left": "issue:44521", "right": "issue:45325", "accept": false, - "reason": "Both are multimodal-related, but assistant mask generation and rope index scaling are different code paths." + "reason": "Both touch multimodal chat/VL behavior, but assistant mask generation and temporal rope position IDs are separate problems." }, { "left": "issue:43976", "right": "issue:44987", "accept": false, - "reason": "Different compatibility failures; one is Python support, the other is loading a specific model on transformers>=5.1.0." + "reason": "Environment/version incompatibility vs a specific model load failure; not duplicates." }, { "left": "issue:45600", "right": "issue:45712", "accept": false, - "reason": "CI break from removed config references vs leftover dummy classes failing repo checks; separate maintenance issues." + "reason": "Both are CI/repo hygiene problems, but one is stale auto_mappings references and the other is leftover dummy classes. Different fixes." }, { "left": "issue:43450", "right": "issue:44279", "accept": false, - "reason": "Video processor batch-shape bug is unrelated to a dependency issue." + "reason": "Video processor shape bug vs dependency issue; unrelated." }, { "left": "issue:43065", "right": "issue:44898", "accept": false, - "reason": "Sam3PixelDecoder dummy Conv2d and Perceiver non-default-resolution failure are different model bugs." + "reason": "Sam3PixelDecoder dummy layer vs Perceiver non-default-resolution failure. Different models and failures." }, { "left": "issue:42994", "right": "issue:43317", "accept": false, - "reason": "Quantized model saving and dequantized load/offload failure are both quantization-adjacent but different failure modes and fixes." + "reason": "Both are quantization/load related, but saving a quantized model and loading a dequantized model with device_map=auto are different paths." }, { "left": "issue:44466", "right": "issue:45276", "accept": false, - "reason": "Both touch weights/embeddings, but serialization inconsistency and resize_token_embeddings behavior in gemma4 are not the same defect." + "reason": "Both involve embeddings/tied weights, but one is device-dependent serialization and the other is resize_token_embeddings not updating Gemma4 embeddings/output heads." }, { "left": "issue:44261", "right": "issue:44263", "accept": false, - "reason": "MLA RMS norm precision and GlmMoeDsaIndexer split behavior are unrelated." + "reason": "MLA rms_norm_eps precision concern vs torch.split return values in GlmMoeDsaIndexer. Different components." }, { "left": "issue:43441", "right": "issue:44683", "accept": false, - "reason": "Both are attention-related, but one is Ministral FlashAttention and the other is compiled flex_attention on newer torch; different code paths." + "reason": "Attention backend failures, but one is Ministral-3 with FlashAttention in v5 RC and the other is compiled flex_attention on torch >= 2.9. Not the same bug." }, { "left": "issue:43650", "right": "issue:45290", "accept": false, - "reason": "Placeholder/odd issue vs a real chat-template crash; no overlap." + "reason": "Placeholder title vs a concrete chat-template crash with tool calls. No duplicate signal." }, { - "left": "issue:43723", - "right": "issue:44162", + "left": "issue:44623", + "right": "issue:45397", "accept": false, - "reason": "Generic tokenizer loading in v5 vs an ESM2 regression; not the same underlying bug." + "reason": "Processor save_pretrained missing files vs Gemma-4 zero3 from_pretrained loading. Related area, different issue." }, { - "left": "issue:44623", - "right": "issue:45397", + "left": "issue:43723", + "right": "issue:44162", "accept": false, - "reason": "processor.save_pretrained missing files and gemma-4 zero3 from_pretrained are unrelated." + "reason": "Tokenizer-loading regression vs ESM2 model breakage; different model families and failure modes." }, { "left": "issue:43784", "right": "issue:45542", "accept": false, - "reason": "Import NameError for sentence-transformers vs tf backend installation detection; different subsystems." + "reason": "Import-time NameError in sentence-transformers integration vs backend selection error when only tensorboard is installed. Different cause." }, { "left": "issue:44623", "right": "issue:44749", "accept": false, - "reason": "Missing processor save files and slow filtering after upgrade are unrelated." + "reason": "Processor save_pretrained files missing vs a tokenizer performance regression after upgrade. Not duplicates." }, { "left": "issue:44671", "right": "issue:44779", "accept": false, - "reason": "CamemBERT masked LM regression and Deepseek tokenizer output regression are different model-specific bugs." + "reason": "Both are v5 correctness regressions in NLP models, but CamemBERT masked LM output and Deepseek tokenization are separate concrete bugs." }, { "left": "issue:43653", "right": "issue:45593", "accept": false, - "reason": "BigBirdTokenizer special-token registration and D-FINE auxiliary loss behavior have no common root cause." + "reason": "BigBirdTokenizer special-token registration bug vs D-FINE auxiliary-loss behavior. Unrelated." }, { "left": "issue:44987", "right": "issue:45542", "accept": false, - "reason": "Model loading failure and TensorFlow/tensorboard backend detection are unrelated." + "reason": "Model loading failure vs TensorFlow backend/installation issue. No overlap." }, { "left": "issue:43901", "right": "issue:43994", "accept": false, - "reason": "Docs mismatch for text-classification pipeline vs actual SigLIP2 inference malfunction; not duplicates." + "reason": "Documentation/API mismatch vs SigLIP2 nonsensical outputs. Different scope." }, { "left": "issue:43122", "right": "issue:43650", "accept": false, - "reason": "Tokenizer version drift and an unrelated placeholder issue." + "reason": "Tokenizer behavior changed across versions vs placeholder issue title. No duplicate." }, { "left": "issue:44361", "right": "issue:45593", "accept": false, - "reason": "MLukeTokenizer AttributeError and D-FINE auxiliary-loss behavior are unrelated." + "reason": "MLukeTokenizer AttributeError vs D-FINE denoising/aux-loss behavior. Completely different bugs." } ] }, @@ -9927,8 +9835,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -9987,16 +9895,16 @@ "issue:43475|issue:44016", "issue:44297|issue:45397", "issue:43673|issue:43761", - "issue:43525|issue:45759", "issue:44188|issue:44336", + "issue:43525|issue:45759", "issue:43525|issue:43881", "issue:44589|issue:45362", "issue:43295|issue:44617", "issue:42491|issue:43299", "issue:41628|issue:43976", "issue:43901|issue:44246", - "issue:43352|issue:45250", "issue:44336|issue:44373", + "issue:43352|issue:45250", "issue:16998|issue:22355", "issue:44242|issue:45593", "issue:44295|issue:45397", @@ -10005,41 +9913,41 @@ "issue:44246|issue:44496", "issue:44462|issue:45356", "issue:36246|issue:42757", - "issue:36246|issue:44351", "issue:44336|issue:44485", + "issue:36246|issue:44351", "issue:43673|issue:44561", "issue:43824|issue:43976", "issue:43976|issue:44964", "issue:44246|issue:44871", "issue:44242|issue:44857" ], - "prepared_review_unit_hash": "d64743beb55d771774a776947a65ec6d7311044b1b20bce32a34604673e725da", + "prepared_review_unit_hash": "ce782dec63463b2cbd8429aa5adcf7090407f73c3f25f89d868382405a98a69e", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 30, - "serialized_chars": 36530, - "estimated_input_tokens": 9133, - "estimated_eval_tokens": 18522 + "serialized_chars": 36525, + "estimated_input_tokens": 9132, + "estimated_eval_tokens": 18520 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "The cluster is mostly heterogeneous import/config/regression reports. None of the soft pairs look like the same underlying bug or change, so I would not merge any of them.", - "confidence": 0.91, - "canonical_issue_reason": "Issue 43824 is the strongest anchor in the set: it has the highest inbound reference count and clearly describes a concrete, widely hit import regression for Qwen2.5-VL.", + "summary": "These soft-similarity pairs are mostly superficial; they share broad themes like imports, loading, tokenizers, or multimodal processing, but they do not appear to be the same underlying bug. I would reject all soft edges.", + "confidence": 0.96, + "canonical_issue_reason": "No single canonical issue: the set spans unrelated bugs (imports, model loading, tokenizer/pipeline regressions, docs), so no item cleanly represents the whole cluster.", "canonical_pr_reason": null, - "best_issue_reason": "43824 is the best representative issue because it is specific, high-signal, and has the most external references/discussion among the issues shown.", + "best_issue_reason": "Issue 44521 is the strongest standalone representative because it is a current, well-documented multimodal regression with inbound references, but it still does not unify the rest of the set.", "best_pr_reason": null, "soft_edge_verdicts": [ { @@ -10052,181 +9960,175 @@ "left": "issue:43262", "right": "issue:44521", "accept": false, - "reason": "Both involve chat templates, but one is sampling-rate defaulting and the other is assistant mask generation; different bugs." + "reason": "Both touch apply_chat_template, but one is audio sampling-rate defaulting and the other is multimodal assistant masks; different bugs." }, { "left": "issue:43475", "right": "issue:44016", "accept": false, - "reason": "SAM 3 video attribute error vs. notebook syntax error; no shared code-path." + "reason": "SAM 3 vision encoder attribute error is unrelated to a notebook syntax error." }, { "left": "issue:44297", "right": "issue:45397", "accept": false, - "reason": "Tokenizer_class save metadata mismatch vs. zero3 from_pretrained issue; distinct problems." + "reason": "Tokenizer save_pretrained metadata mismatch is unrelated to gemma-4 zero3 loading." }, { "left": "issue:43673", "right": "issue:43761", "accept": false, - "reason": "Cache/chunked-prefill regression vs. CLIP vision hidden_states return bug; unrelated." + "reason": "Chunked prefill cache issue and CLIP hidden_states regression are different code paths." }, { - "left": "issue:43525", - "right": "issue:45759", + "left": "issue:44188", + "right": "issue:44336", "accept": false, - "reason": "Missing pad_token_id in Llama4Config vs. composite Qwen config unwrapping; similar area, different concrete defects." + "reason": "torch.compile attention-kernel divergence vs ANSI codes in loading_report; no overlap beyond general utilities." }, { - "left": "issue:44188", - "right": "issue:44336", + "left": "issue:43525", + "right": "issue:45759", "accept": false, - "reason": "Attention-kernel divergence vs. ANSI escape output in loading_report; unrelated." + "reason": "Both are model-loading/config issues, but Llama4 pad_token_id and Qwen multimodal text_config unwrapping are distinct failures." }, { "left": "issue:43525", "right": "issue:43881", "accept": false, - "reason": "Llama4 config attribute error vs. GLM-4V loading failure; different root causes." + "reason": "Different loading errors for different models; no clear shared concrete bug." }, { "left": "issue:44589", "right": "issue:45362", "accept": false, - "reason": "Float8 storage lookup TypeError vs. Qwen3.5 chat crash; not the same bug." + "reason": "Float8 storage deserialization and Qwen3.5 chat crash are unrelated." }, { "left": "issue:43295", "right": "issue:44617", "accept": false, - "reason": "Processor/tokenizer regression vs. CUDA OOM; unrelated." + "reason": "processor.tokenizer regression vs CUDA OOM; not the same underlying problem." }, { "left": "issue:42491", "right": "issue:43299", "accept": false, - "reason": "Both are Qwen3 MoE-related, but one is LoRA incompatibility and the other is model loading regression; too different to merge." + "reason": "Both involve Qwen/MoE loading regressions, but they affect different model families and failure modes." }, { "left": "issue:41628", "right": "issue:43976", "accept": false, - "reason": "Missing AutoImageProcessor import vs. Python version compatibility problem; unrelated." + "reason": "ImportError for AutoImageProcessor is unrelated to Python version support incompatibility." }, { "left": "issue:43901", "right": "issue:44246", "accept": false, - "reason": "Docs/behavior mismatch vs. slow import; different areas entirely." + "reason": "Docs mismatch for return_all_scores vs intermittent import slowness; different concerns." }, { - "left": "issue:43352", - "right": "issue:45250", + "left": "issue:44336", + "right": "issue:44373", "accept": false, - "reason": "Model-specific Flash Attention unsupported error vs. generic Flash Attention topic; not the same concrete issue." + "reason": "ANSI color handling in loading_report is unrelated to a docstring typo." }, { - "left": "issue:44336", - "right": "issue:44373", + "left": "issue:43352", + "right": "issue:45250", "accept": false, - "reason": "Terminal color output bug vs. docstring issue; unrelated." + "reason": "Specific model lacks Flash Attention 2 support; the other is just a generic Flash Attention topic." }, { "left": "issue:16998", "right": "issue:22355", "accept": false, - "reason": "model_max_length question vs. missing transformers.onnx module; no overlap." + "reason": "model_max_length question and missing transformers.onnx module are unrelated." }, { "left": "issue:44242", "right": "issue:45593", "accept": false, - "reason": "Mixture-of-experts load-balancing loss bug vs. D-FINE auxiliary-loss omission; different models and code paths." + "reason": "Both are loss-related when an option is disabled, but they affect different models and mechanisms." }, { "left": "issue:44295", "right": "issue:45397", "accept": false, - "reason": "position_ids buffer access bug vs. zero3 model loading issue; unrelated." + "reason": "position_ids buffer access error vs zero3 from_pretrained issue; no shared code path." }, { "left": "issue:44661", "right": "issue:45542", "accept": false, - "reason": "Tokenizer-mapping failure in add-new-model-like vs. TensorFlow backend detection problem; different subsystems." + "reason": "add-new-model-like tokenizer mapping bug is unrelated to the TensorFlow backend import error." }, { "left": "issue:41950", "right": "issue:44220", "accept": false, - "reason": "Video-classification pipeline processor selection bug vs. fbank extraction issue; different failures." + "reason": "Video-classification processor lookup and _torch_extract_fbank_features() are different pipeline components." }, { "left": "issue:44246", "right": "issue:44496", "accept": false, - "reason": "Import slowness vs. unrecognized model/config error; not a duplicate." + "reason": "Slow import is unrelated to unrecognized model config loading." }, { "left": "issue:44462", "right": "issue:45356", "accept": false, - "reason": "AutoTokenizer ignoring tokenizer.json vs. codec handling regression in Kimi-K2.5; distinct tokenizer bugs." + "reason": "AutoTokenizer ignoring tokenizer.json and Kimi-K2.5 codec regression are different tokenizer bugs." }, { "left": "issue:36246", "right": "issue:42757", "accept": false, - "reason": "Qwen2.5-VL import error vs. huggingface_hub is_offline_mode import error; unrelated." + "reason": "Different missing imports from different packages; same symptom class, not the same bug." }, { - "left": "issue:36246", - "right": "issue:44351", + "left": "issue:44336", + "right": "issue:44485", "accept": false, - "reason": "Different missing symbols from different modules; not the same issue." + "reason": "ANSI-output bug and GLM-5 RoPE discussion are unrelated." }, { - "left": "issue:44336", - "right": "issue:44485", + "left": "issue:36246", + "right": "issue:44351", "accept": false, - "reason": "ANSI code emission bug vs. GLM-5 RoPE implementation discussion; unrelated." + "reason": "Both are import errors, but for different missing symbols and modules." }, { "left": "issue:43673", "right": "issue:44561", "accept": false, - "reason": "Generation cache regression vs. trust_remote_code breakage from removed helper; different root cause." + "reason": "cache-related chunked prefill regression and removed helper breaking trust_remote_code are unrelated." }, { "left": "issue:43824", "right": "issue:43976", "accept": false, - "reason": "Qwen2.5-VL import failure vs. Python version support issue; unrelated." + "reason": "Missing Qwen2_5_VL export and Python version support issue are unrelated." }, { "left": "issue:43976", "right": "issue:44964", "accept": false, - "reason": "Python compatibility vs. Phi-4 multimodal loading failure; different bugs." + "reason": "Python compatibility regression and Phi-4 multimodal loading failure are different issues." }, { "left": "issue:44246", "right": "issue:44871", "accept": false, - "reason": "Import slowness vs. Gemma-3 EOS token config mismatch; unrelated." + "reason": "Import timing problem vs Gemma-3 eos_token_id mismatch; no shared bug." }, { "left": "issue:44242", "right": "issue:44857", "accept": false, - "reason": "Aux-loss gating bug vs. float16 AMP crash in LwDetr; different symptoms and code paths." - }, - { - "left": "issue:44246", - "right": "issue:44496", - "accept": false, - "reason": "Import performance issue vs. model-type/config validation error; not a duplicate." + "reason": "Different loss/AMP failures in different models; too broad to merge." } ] }, @@ -10245,8 +10147,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -10302,17 +10204,17 @@ "soft_pairs": [ "issue:45362|issue:45375", "issue:44162|issue:45200", - "issue:43441|issue:43540", "issue:44295|issue:44749", "issue:44779|issue:45292", + "issue:43441|issue:43540", "issue:45292|issue:45584", "issue:44877|issue:45375", "issue:36246|issue:43596", "issue:44987|issue:45459", "issue:45292|issue:45701", "issue:41628|issue:41950", - "issue:45561|issue:45563", "issue:44779|issue:45701", + "issue:45561|issue:45563", "issue:43901|issue:45003", "issue:45375|issue:45759", "issue:43065|issue:43408", @@ -10328,24 +10230,24 @@ "issue:44995|issue:45588", "issue:43653|issue:44162", "issue:44368|issue:45563", - "issue:44368|issue:44977", "issue:45397|issue:45656", + "issue:44368|issue:44977", "issue:43673|issue:44734" ], - "prepared_review_unit_hash": "d04f88793b4ce9d93d6fd7a7aa50dcbd7089ae4b0d144aa169be609844911769", + "prepared_review_unit_hash": "d035c946b33a38e7aefba61d37ac04b970a6d8e17c2a11c151e9e80669c52969", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 31, - "serialized_chars": 37084, + "serialized_chars": 37083, "estimated_input_tokens": 9271, "estimated_eval_tokens": 18798 }, @@ -10354,198 +10256,198 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "No soft edge looks like a true duplicate; the set is mostly unrelated bug reports across different subsystems, with only loose thematic overlap in a few tokenization/configuration issues.", - "confidence": 0.24, - "canonical_issue_reason": "issue:45701 is the broadest issue here and the closest thing to an umbrella tokenization-regression report, but the cluster does not contain a strong true duplicate set.", + "summary": "This cluster is heterogeneous: most items are distinct Transformers regressions across tokenization, config loading, generation, DeepSpeed, and SAM3/Qwen/Gemma code paths. None of the soft pairs look like the same concrete bug/change, so I would not merge any of them as duplicates.", + "confidence": 0.68, + "canonical_issue_reason": "issue:45292 is the best representative of the small embedding/weight-resizing regression subset: it is concrete, broadly applicable, and closer to other nearby resize/tied-weight reports than the more model-specific tickets.", "canonical_pr_reason": null, - "best_issue_reason": "issue:45701 is the most general and reusable representative among these issues; the others are mostly narrower, model-specific, or unrelated failures.", + "best_issue_reason": "issue:45292 is the clearest standalone bug in the set and the most reusable triage anchor for this cluster; it has a direct API symptom and broad model impact.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:45362", "right": "issue:45375", "accept": false, - "reason": "Same Qwen family, but one is a chat crash and the other is a missing config field; different failure modes and code paths." + "reason": "Both mention Qwen 3.5, but one is a chat/runtime crash and the other is a strict-config field drop; different code paths." }, { "left": "issue:44162", "right": "issue:45200", "accept": false, - "reason": "Both are model/config bugs, but they affect different models and different inputs; not the same underlying defect." - }, - { - "left": "issue:43441", - "right": "issue:43540", - "accept": false, - "reason": "Different models and different symptoms: FlashAttention failure vs video-input validation error." + "reason": "ESM2 loading is unrelated to Gemma 4 mm_token_type_ids defaulting; different models and failures." }, { "left": "issue:44295", "right": "issue:44749", "accept": false, - "reason": "Both mention version-related regressions, but one is a buffer access issue and the other is a performance slowdown; not the same bug." + "reason": "One is a position_ids buffer bug, the other is a post-upgrade tokenization slowdown; not the same issue." }, { "left": "issue:44779", "right": "issue:45292", "accept": false, - "reason": "Both touch tokenization/embeddings, but the concrete problems and affected paths differ." + "reason": "DeepSeek tokenizer output regression and token embedding resize propagation are different bugs." + }, + { + "left": "issue:43441", + "right": "issue:43540", + "accept": false, + "reason": "FlashAttention crash in Ministral-3 and Qwen3OmniMoe video-input validation are unrelated." }, { "left": "issue:45292", "right": "issue:45584", "accept": false, - "reason": "Resize-embedding behavior and empty-transcription handling are unrelated failures." + "reason": "Resize-token-embeddings affects embedding weights, while Whisper empty-transcription failure is about special-token/chat-template handling." }, { "left": "issue:44877", "right": "issue:45375", "accept": false, - "reason": "Both involve config handling, but they target different models and different missing/strict fields." + "reason": "Both involve config handling, but one is strict loading of granite_speech and the other is a missing Qwen3_5MoeVisionConfig field." }, { "left": "issue:36246", "right": "issue:43596", "accept": false, - "reason": "An import/export issue and a DeepSpeed ZeRO3 indexing error are not the same bug." + "reason": "Importing a missing Qwen2.5 VL processor and a ZeRO3/BertModel index error are clearly different bugs." }, { "left": "issue:44987", "right": "issue:45459", "accept": false, - "reason": "Both are load-time failures, but one is a model loading regression and the other is tokenizer error handling without protobuf." + "reason": "Model loading failure on transformers>=5.1.0 and protobuf-related tokenizer error masking are different failure points." }, { "left": "issue:45292", "right": "issue:45701", "accept": false, - "reason": "Tokenization-version regression and resize_token_embeddings mismatch are different code paths." + "reason": "Resize-token-embeddings behavior and version-dependent tokenization changes are not the same concrete regression." }, { "left": "issue:41628", "right": "issue:41950", "accept": false, - "reason": "One is an import error for AutoImageProcessor; the other is a pipeline selecting image processors for video." + "reason": "AutoImageProcessor import failure and video-classification pipeline choosing image processors are related only by naming, not the same bug." }, { - "left": "issue:45561", - "right": "issue:45563", + "left": "issue:44779", + "right": "issue:45701", "accept": false, - "reason": "Testing-utils xdist file race and a stale generation warning are unrelated." + "reason": "Both concern tokenization, but one is DeepSeek-specific incorrect output and the other is a broader version-dependent tokenization change." }, { - "left": "issue:44779", - "right": "issue:45701", + "left": "issue:45561", + "right": "issue:45563", "accept": false, - "reason": "Both concern tokenization, but they describe different models and different symptoms; not a single concrete bug." + "reason": "A pytest-xdist file race and a stale generation warning are unrelated test vs generation issues." }, { "left": "issue:43901", "right": "issue:45003", "accept": false, - "reason": "Docs mismatch versus unsafe sys.modules access are unrelated issues." + "reason": "Docs drift for return_all_scores is unrelated to the sys.modules access bug in modeling_utils." }, { "left": "issue:45375", "right": "issue:45759", "accept": false, - "reason": "Both involve Qwen multimodal configs, but one is a missing config field and the other is AutoModelForCausalLM config unwrapping." + "reason": "One is a missing config field being dropped, the other is AutoModelForCausalLM.from_config not unwrapping text_config." }, { "left": "issue:43065", "right": "issue:43408", "accept": false, - "reason": "Same SAM3 area, but one is a decoder implementation detail and the other is a model-type warning." + "reason": "SAM3 pixel-decoder dummy conv and sam3_video/sam3_tracker warning are different SAM3 issues with different symptoms." }, { "left": "issue:43232", "right": "issue:44336", "accept": false, - "reason": "Generation-kwargs handling and ANSI logging output are unrelated." + "reason": "Generation kwargs after sync_gpus and ANSI codes in loading_report are unrelated." }, { "left": "issue:43257", "right": "issue:45137", "accept": false, - "reason": "Both involve distributed training, but one is weight conversion and the other is a ZeRO3 deque underflow." + "reason": "Both mention DeepSpeed, but Qwen3 MOE weight conversion and ZeRO3 deque underflow are different code paths." }, { "left": "issue:43526", "right": "issue:45003", "accept": false, - "reason": "A label-reduction bug and a sys.modules access bug are unrelated." + "reason": "BeitImageProcessorFast label reduction and unsafe sys.modules access are unrelated." }, { "left": "issue:44995", "right": "issue:45200", "accept": false, - "reason": "A stale indexer-cache crash and missing default token-type IDs do not share the same root cause." + "reason": "GlmMoeDsa stale indexer cache and Gemma 4 token-type defaults are unrelated." }, { "left": "issue:44496", "right": "issue:45003", "accept": false, - "reason": "Unrecognized model config and unsafe sys.modules access are different loading-path problems." + "reason": "Unrecognized model config and sys.modules access are separate core-loading failures." }, { "left": "issue:43262", "right": "issue:44246", "accept": false, - "reason": "Audio chat-template sampling-rate defaulting and slow imports are unrelated." + "reason": "Audio processor sampling-rate defaults and occasional slow imports are unrelated." }, { "left": "issue:42886", "right": "issue:45229", "accept": false, - "reason": "Tokenizer offline-cache loading and Gemma4 multi-GPU OOM are different classes of failure." + "reason": "Offline tokenizer cache loading and Gemma 4 multi-GPU OOM are different problems." }, { "left": "issue:43475", "right": "issue:43540", "accept": false, - "reason": "Both are video/multimodal issues, but they involve different models, objects, and errors." + "reason": "SAM3 vision encoder output attribute error and Qwen3OmniMoe video input ValueError are distinct model bugs." }, { "left": "issue:44466", "right": "issue:45335", "accept": false, - "reason": "Both mention tied embeddings/resize behavior, but they affect different model structures and serialization paths." + "reason": "Tied-weight serialization inconsistency is not the same as resize_token_embeddings failing to update decoder.embed_tokens." }, { "left": "issue:44995", "right": "issue:45588", "accept": false, - "reason": "Stale cache indexing and a FlashAttention s_aux=None crash are unrelated." + "reason": "A stale cache crash in GlmMoeDsa and flash_attention.py s_aux=None handling are different issues." }, { "left": "issue:43653", "right": "issue:44162", "accept": false, - "reason": "Tokenizer special-token registration and an ESM2 model breakage are not the same issue." + "reason": "BigBirdTokenizer special-token registration and ESM2 broken loading are unrelated." }, { "left": "issue:44368", "right": "issue:45563", "accept": false, - "reason": "Different warnings in different subsystems; not a shared defect." + "reason": "Both are warnings, but one is tie_word_embeddings config noise and the other is num_return_sequences warning drift." }, { - "left": "issue:44368", - "right": "issue:44977", + "left": "issue:45397", + "right": "issue:45656", "accept": false, - "reason": "A tie_word_embeddings warning and a flash-attention generation failure are unrelated." + "reason": "Gemma-4 ZeRO3 loading and duplicate optimizer.step() under DeepSpeed are different failures." }, { - "left": "issue:45397", - "right": "issue:45656", + "left": "issue:44368", + "right": "issue:44977", "accept": false, - "reason": "Both mention DeepSpeed, but one is a from_pretrained loading bug and the other is a double optimizer-step issue." + "reason": "Tie-word-embeddings warning noise is unrelated to Qwen3.5 flash-attention generation failure." }, { "left": "issue:43673", "right": "issue:44734", "accept": false, - "reason": "Both involve generation cache behavior, but one is chunked_prefill cache availability and the other is KV-cache continuation indexing." + "reason": "GenerationMixin cache missing in chunked_prefill and KV-cache continuation indexing are both cache-related but not the same concrete bug." } ] }, @@ -10564,8 +10466,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -10628,8 +10530,8 @@ "issue:43646|issue:44162", "issue:44361|issue:44749", "issue:43232|issue:43673", - "issue:28282|issue:30990", "issue:44623|issue:45656", + "issue:28282|issue:30990", "issue:43673|issue:43827", "issue:43452|issue:45003", "issue:42994|issue:43482", @@ -10637,8 +10539,8 @@ "issue:43475|issue:43643", "issue:44749|issue:45656", "issue:45230|issue:45397", - "issue:42371|issue:45092", "issue:44038|issue:44623", + "issue:42371|issue:45092", "issue:43475|issue:45381", "issue:45081|issue:45103", "issue:41628|issue:44933", @@ -10649,20 +10551,20 @@ "issue:45310|issue:45375", "issue:42222|issue:45200" ], - "prepared_review_unit_hash": "583a8e17b277b3dcbbce6013f1a48066c6a24e5789f325054339cd8882940e99", + "prepared_review_unit_hash": "8a45655a08e6d8a93d8fc4e2698ffe6d9edd405b96aa877f87921227b76fef7e", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, "soft_pair_count": 30, - "serialized_chars": 36050, + "serialized_chars": 36049, "estimated_input_tokens": 9013, "estimated_eval_tokens": 18282 }, @@ -10671,198 +10573,198 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "Most pairs are unrelated regressions or model-specific bugs; only two pairs look like true duplicates based on identical titles and the same concrete failure mode.", - "confidence": 0.9, - "canonical_issue_reason": "issue:45310 is the clearest representative of a concrete, version-specific loading regression, and it has an exact duplicate in issue:45478 with the same title and failure context.", + "summary": "The set is mostly a heterogeneous grab bag of Transformers/model-loading regressions, not one tight duplicate cluster. The only clear mini-cluster is the Qwen3.5-MoE loading/config issue, where one report names the likely root cause and the others look like symptom duplicates.", + "confidence": 0.37, + "canonical_issue_reason": "issue:45375 is the best anchor because it identifies the likely root cause (missing `deepstack_visual_indexes` being dropped by `@strict`) behind the Qwen3.5-MoE `from_pretrained` failures; the other Qwen3.5-MoE reports are downstream symptom duplicates.", "canonical_pr_reason": null, - "best_issue_reason": "issue:45310 is a strong canonical issue candidate because it is specific, actionable, and appears to have an exact duplicate; the rest of the set is too heterogeneous to serve as a better global representative.", + "best_issue_reason": "issue:45375 is the most actionable and root-cause-oriented report in the only clear duplicate subcluster, and it best explains the related loading errors in 45310/45478.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:44060", "right": "issue:44077", "accept": false, - "reason": "Different bugs: a tied-weights warning in Qwen3-Next versus a patchtsmixer post_init policy issue." + "reason": "Different model families and different warnings; no shared concrete bug path." }, { "left": "issue:45003", "right": "issue:45561", "accept": false, - "reason": "Different code paths: unsafe sys.modules access versus an xdist race in captured_info.txt." + "reason": "`sys.modules` access bug vs xdist race on `captured_info.txt`; unrelated failures." }, { "left": "issue:45137", "right": "issue:45600", "accept": false, - "reason": "Unrelated failures: DeepSpeed ZeRO3 deque error versus stale Sam3LiteText references in auto_mappings.py." + "reason": "DeepSpeed ZeRO3 deque error vs removed Sam3LiteText config references; different code paths." }, { "left": "issue:43931", "right": "issue:44368", "accept": false, - "reason": "Different model problems: Qwen3-VL shape mismatch versus a tie_word_embeddings warning during Qwen3.5 fine-tuning." + "reason": "Both mention Qwen, but one is a weight-shape load error and the other is a tied-weights warning; not the same bug." }, { "left": "issue:44933", "right": "issue:45459", "accept": false, - "reason": "Different tokenizer/import issues: missing image_utils import versus protobuf decode error masking." + "reason": "Import missing from `image_utils` vs protobuf exception masking tokenizer errors; different root causes." }, { "left": "issue:44869", "right": "issue:45356", "accept": false, - "reason": "Different tokenizer/runtime bugs: Whisper timestamp decode crash versus Kimi-K2.5 codec handling regression." + "reason": "Whisper timestamp decode crash vs Kimi tokenizer codec/regression warning; unrelated." }, { "left": "issue:44297", "right": "issue:44623", "accept": false, - "reason": "Both are save_pretrained issues, but one is tokenizer_class mismatch and the other is processor file emission; not the same bug." + "reason": "Tokenizer save metadata mismatch vs processor save missing files; same area, but not the same concrete bug." }, { "left": "issue:43646", "right": "issue:44162", "accept": false, - "reason": "Different breakages: custom model initialization under v5 versus ESM2 being broken." + "reason": "Broad v5 custom-model init breakage vs an ESM2-specific breakage; too different to merge." }, { "left": "issue:44361", "right": "issue:44749", "accept": false, - "reason": "Different problems: MLukeTokenizer AttributeError versus a performance regression during filtering." + "reason": "Tokenizer AttributeError vs a data-filtering slowdown regression; not the same failure." }, { "left": "issue:43232", "right": "issue:43673", "accept": false, - "reason": "Different generation regressions: sync_gpus update bug versus missing GenerationMixin cache in chunked_prefill." + "reason": "Different generation internals (`sync_gpus` kwargs update vs chunked prefill cache missing)." }, { - "left": "issue:28282", - "right": "issue:30990", + "left": "issue:44623", + "right": "issue:45656", "accept": false, - "reason": "Different startup issues: missing PyTorch ImportError versus Sentence Transformers hanging while loading." + "reason": "Processor serialization bug vs DeepSpeed optimizer stepping twice; unrelated." }, { - "left": "issue:44623", - "right": "issue:45656", + "left": "issue:28282", + "right": "issue:30990", "accept": false, - "reason": "Different subsystems: processor.save_pretrained output bug versus DeepSpeed optimizer step being called twice." + "reason": "ImportError for PyTorch vs a loading hang in Sentence Transformers; different symptoms and likely causes." }, { "left": "issue:43673", "right": "issue:43827", "accept": false, - "reason": "Different scope: generation cache regression versus documentation still referencing removed pipeline() APIs." + "reason": "Generation cache regression vs docs still referencing removed `pipeline()` APIs; not a duplicate bug." }, { "left": "issue:43452", "right": "issue:45003", "accept": false, - "reason": "Different failures: gguf_file/from_pretrained loading breakage versus unsafe sys.modules access in modeling_utils." + "reason": "GGUF loading breakage vs unsafe `sys.modules` access; different code paths." }, { "left": "issue:42994", "right": "issue:43482", "accept": false, - "reason": "Related to GGUF/quantization broadly, but one is saving quantized models and the other is loading Qwen2.5-GGUF; not the same concrete bug." + "reason": "Quantized saving failure vs Qwen2.5-GGUF loading failure; opposite sides of the workflow and different bugs." }, { "left": "issue:44749", "right": "issue:45310", "accept": false, - "reason": "Different issues: data filtering slowdown versus Qwen3.5-Moe from_pretrained error." + "reason": "Performance regression on upgraded Transformers vs Qwen3.5-MoE loading error; unrelated." }, { "left": "issue:43475", "right": "issue:43643", "accept": false, - "reason": "Different failures: SAM3 vision output attribute error versus trust_remote_code causing missing fields in AutoConfig." + "reason": "SAM3 vision encoder attribute error vs missing fields returned from `AutoConfig`; different models and failures." }, { "left": "issue:45230", "right": "issue:45397", "accept": false, - "reason": "Too generic on the left and specific on the right; no evidence they describe the same Zero3 from_pretrained bug." + "reason": "Generic bug report vs Gemma-4 DeepSpeed ZeRO3 loading issue; no evidence they match." }, { - "left": "issue:42371", - "right": "issue:45092", + "left": "issue:44038", + "right": "issue:44623", "accept": false, - "reason": "Unrelated: TF32 settings guidance versus old InternVL2 remote-code/meta-init incompatibility." + "reason": "Qwen3-VL-Moe bug vs processor save missing files; not the same underlying problem." }, { - "left": "issue:44038", - "right": "issue:44623", + "left": "issue:42371", + "right": "issue:45092", "accept": false, - "reason": "Different problems: Qwen3-VL-Moe/v5 compatibility versus missing files in processor.save_pretrained." + "reason": "TF32 settings docs issue vs InternVL2 meta-init incompatibility; documentation vs runtime bug." }, { "left": "issue:43475", "right": "issue:45381", "accept": false, - "reason": "Different multimodal bugs: SAM3 output attribute error versus qwen2.5-vl video position_ids issue." + "reason": "Different video-model failures (`Sam3VisionEncoderOutput` attribute vs Qwen2.5-VL position IDs)." }, { "left": "issue:45081", "right": "issue:45103", "accept": false, - "reason": "Different crashes: Mistral regex patching versus auto_docstring handling of future annotations." + "reason": "Tokenizer regex crash vs docstring/annotations crash; unrelated subsystems." }, { "left": "issue:41628", "right": "issue:44933", "accept": false, - "reason": "Both are import-related, but they concern different missing symbols and different import paths." + "reason": "Different import failures in different modules; not the same missing symbol bug." }, { "left": "issue:45375", "right": "issue:45478", - "accept": false, - "reason": "Different Qwen3.5/Moe issues: missing config field versus from_pretrained regression." + "accept": true, + "reason": "Same Qwen3.5-MoE loading regression; 45375 explains the likely root cause that would surface as 45478's `from_pretrained` error." }, { "left": "issue:43475", "right": "issue:44661", "accept": false, - "reason": "Unrelated: SAM3 vision encoder output bug versus add-new-model-like failure in TOKENIZER_MAPPING_NAMES." + "reason": "SAM3 output attribute issue vs tokenizer mapping failure when adding a model; unrelated." }, { "left": "issue:42673", "right": "issue:43704", - "accept": true, - "reason": "Same bug and same title: Qwen3ForCausalLM VRAM leak with multiple dataloader threads." + "accept": false, + "reason": "Same title, but they are separate reports with no evidence of a shared code-path bug from the provided text." }, { "left": "issue:42994", "right": "issue:43525", "accept": false, - "reason": "Different model/load/save issues: quantized model saving failure versus Llama4Config missing pad_token_id." + "reason": "Quantized save failure vs missing `pad_token_id` on `Llama4Config`; different bugs." }, { "left": "issue:45310", "right": "issue:45375", - "accept": false, - "reason": "Different bugs: Qwen3.5-Moe loading error versus missing deepstack_visual_indexes in Qwen3_5MoeVisionConfig." + "accept": true, + "reason": "Same Qwen3.5-MoE failure family; 45375 pinpoints the missing config field that plausibly causes 45310's `from_pretrained` error." }, { "left": "issue:42222", "right": "issue:45200", "accept": false, - "reason": "Different model families and failures: vitpose breakage versus Gemma 4 token-type defaults." + "reason": "VitPose model breakage vs Gemma-4 token-type-id defaulting; different model-specific issues." }, { - "left": "issue:45310", - "right": "issue:45478", - "accept": true, - "reason": "Exact duplicate report of the same Qwen3.5-Moe from_pretrained error, with the same title and regression context." + "left": "issue:44623", + "right": "issue:45656", + "accept": false, + "reason": "Processor save regression vs DeepSpeed optimizer-step duplication; not the same bug." } ] }, "evaluator_result": { "accept": true, - "feedback": "The analysis is grounded in the packet metadata, and the soft-edge verdicts stay conservative overall. The two accepted duplicate pairs both have identical titles, and the rejected pairs are reasonably distinguished by different titles/subsystems." + "feedback": "Grounded overall and conservative in the soft-edge calls. The summary appropriately treats the packet as mostly heterogeneous and limits the duplicate claim to the Qwen3.5-MoE loading/config trio, which is the only clearly supported mini-cluster from the titles provided." }, "evaluator_used": true, "retried": false, @@ -10878,8 +10780,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -10950,242 +10852,248 @@ "issue:43232|issue:44792", "issue:43232|issue:44188", "issue:43784|issue:43824", - "issue:43278|issue:45137", "issue:44188|issue:44373", + "issue:43278|issue:45137", "issue:44623|issue:45310", "issue:44734|issue:45103", "issue:30333|issue:33290", "issue:42898|issue:45356", "issue:43723|issue:44568", "issue:44060|issue:45125", - "issue:45405|issue:45712", "issue:42898|issue:44779", + "issue:45405|issue:45712", "issue:43874|issue:45092", "issue:42898|issue:44448", "issue:43479|issue:44079", "issue:44188|issue:44964", "issue:44295|issue:45656" ], - "prepared_review_unit_hash": "f1eafe935c77f67c730dad1c977ccbc01fa5dc81718b23986af672b47ec1fc94", + "prepared_review_unit_hash": "45df2ddf32ced97ef49075c2cb6c421075e9ed25609c76dd9efc1b94bbe9cb4e", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 32, - "serialized_chars": 37273, - "estimated_input_tokens": 9319, - "estimated_eval_tokens": 18894 + "serialized_chars": 37267, + "estimated_input_tokens": 9317, + "estimated_eval_tokens": 18890 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is heterogeneous: most items are separate Transformers regressions around imports, tokenizers, serialization, multimodal models, or DeepSpeed/FP8 behavior. I do not see a single underlying duplicate bug/change tying them together.", - "confidence": 0.93, - "canonical_issue_reason": "No single canonical issue: the items span unrelated failure modes (imports, tokenizer behavior, save/load, model-specific regressions, and training/runtime issues) rather than one duplicate thread.", + "summary": "This cluster is highly heterogeneous: mostly unrelated bug reports spanning tokenizer regressions, import errors, save/load issues, multimodal model failures, and DeepSpeed/compile edge cases. None of the soft-edge pairs look like the same underlying defect, so all are rejected.", + "confidence": 0.92, + "canonical_issue_reason": "issue:42898 is the broadest, most central regression in the set: a v5 behavior change affecting tokenization semantics rather than a model-specific one-off. It best represents the general \u2018transformers v5 breaking behavior\u2019 theme visible across several nearby issues.", "canonical_pr_reason": null, - "best_issue_reason": "No issue is a strong global representative. If forced, issue 42898 is the broadest v5 tokenizer regression, but it still does not explain the rest of the cluster.", + "best_issue_reason": "issue:42898 is the strongest representative issue because it is concrete, broadly relevant, and not tied to a single model family or integration path.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:42831", "right": "issue:44222", "accept": false, - "reason": "Both mention FP8, but one is an accuracy regression and the other is a save_pretrained failure for MoE; different bugs and code paths." + "reason": "Both mention FP8, but one is an accuracy issue and the other is a save_pretrained failure for MoE; different code paths." }, { "left": "issue:41628", "right": "issue:44351", "accept": false, - "reason": "Both are import errors, but for different symbols (AutoImageProcessor vs HybridCache) with no clear shared root cause." + "reason": "Both are import errors, but they concern different missing symbols and different modules." }, { "left": "issue:44222", "right": "issue:45310", "accept": false, - "reason": "Both involve Qwen/MoE-ish loading or saving problems, but one is FP8 save_pretrained and the other is a from_pretrained error for a different model/version path." + "reason": "One is a FP8 save_pretrained bug and the other is a Qwen3.5 from_pretrained error; not the same defect." }, { "left": "issue:42994", "right": "issue:44164", "accept": false, - "reason": "Both are save/load issues, but 42994 is about quantized model saving while 44164 is about extra_state handling; too different to merge." + "reason": "Both involve serialization, but one is quantized model saving and the other is extra_state handling in save/from_pretrained." }, { "left": "issue:42175", "right": "issue:45712", "accept": false, - "reason": "Both touch packaging/backend availability, but one is torch extra installation missing TensorFlow backend and the other is dummy_pt_objects cleanup; different underlying problems." + "reason": "Packaging/backend availability versus dummy class leakage into torch-less environments are different problems." }, { "left": "issue:43377", "right": "issue:44242", "accept": false, - "reason": "One is a batching/padding-mask correctness bug in MIMI Encoder; the other is an MoE load-balancing-loss condition. Not the same change." + "reason": "Missing padding-mask support in an encoder is unrelated to load-balancing loss gating on router logits." }, { "left": "issue:36010", "right": "issue:41628", "accept": false, - "reason": "Different missing imports for different classes; these are separate export/regression issues." + "reason": "Different missing imports (`GenerationMixin` vs `AutoImageProcessor`) and different failure surfaces." }, { "left": "issue:44038", "right": "issue:45310", "accept": false, - "reason": "Both concern Qwen models on v5, but one is a Qwen3-VL-Moe bug and the other is a Qwen3.5 MoE from_pretrained error; not the same concrete defect." + "reason": "Both touch Qwen/MoE and v5, but the reported failures are distinct and no shared code path is shown." }, { "left": "issue:44496", "right": "issue:45561", "accept": false, - "reason": "An unrecognized model/config issue is unrelated to a pytest-xdist race on a captured_info.txt test artifact." + "reason": "A model config loading error and a pytest-xdist file race are unrelated." }, { "left": "issue:45561", "right": "issue:45600", "accept": false, - "reason": "One is a parallel test-file race; the other is stale auto_mappings references breaking CI. Different failures." + "reason": "Concurrency on captured_info.txt is unrelated to stale auto_mappings references." }, { "left": "issue:44964", "right": "issue:45137", "accept": false, - "reason": "Model loading failure for Phi-4 multimodal is unrelated to the DeepSpeed ZeRO3 deque IndexError." + "reason": "Multimodal model loading and a DeepSpeed ZeRO3 deque error are separate bugs." }, { "left": "issue:42898", "right": "issue:44568", "accept": false, - "reason": "Both are tokenizer behavior regressions in v5, but they concern different APIs and symptoms: cleanup spaces vs add_special_tokens/BOS-EOS." + "reason": "Both are tokenization regressions in v5, but they affect different tokenizer behaviors and models." }, { "left": "issue:30990", "right": "issue:35141", "accept": false, - "reason": "Sentence-Transformers loading hangs and token embedding reinitialization are unrelated bugs." + "reason": "A loading hang and an embedding reinitialization bug are not the same underlying issue." }, { "left": "issue:43901", "right": "issue:45561", "accept": false, - "reason": "Pipeline doc mismatch is unrelated to the xdist captured_info.txt race." + "reason": "A documentation mismatch is unrelated to a test utility race condition." }, { "left": "issue:43232", "right": "issue:44792", "accept": false, - "reason": "Generation kwargs update after sync_gpus and a janus image-generation test failure are not the same bug." + "reason": "Generation kwargs handling and a Janus image-generation test failure do not share the same concrete bug." }, { "left": "issue:43232", "right": "issue:44188", "accept": false, - "reason": "Different generation/attention code paths: sync_gpus kwargs handling vs torch.compile branching in attention kernels." + "reason": "Generation bookkeeping after sync_gpus is unrelated to torch.compile-dependent attention kernel divergence." }, { "left": "issue:43784", "right": "issue:43824", "accept": false, - "reason": "Both are import errors from latest transformers, but they break different downstream symbols and do not share a specific fix path." + "reason": "Both are import failures, but for different names in different packages/models." }, { - "left": "issue:43278", - "right": "issue:45137", + "left": "issue:44188", + "right": "issue:44373", "accept": false, - "reason": "Embedding dtype drift in eval is unrelated to the ZeRO3 deque IndexError." + "reason": "A runtime kernel divergence bug is unrelated to a wrong docstring." }, { - "left": "issue:44188", - "right": "issue:44373", + "left": "issue:43278", + "right": "issue:45137", "accept": false, - "reason": "A kernel divergence bug and a docstring issue are not the same underlying problem." + "reason": "BF16/FP32 dtype drift during eval is unrelated to a DeepSpeed deque underflow." }, { "left": "issue:44623", "right": "issue:45310", "accept": false, - "reason": "Missing files in processor.save_pretrained is unrelated to Qwen3.5 MoE from_pretrained failures." + "reason": "Missing files in processor.save_pretrained and a Qwen3.5 from_pretrained error are different operations and bugs." + }, + { + "left": "issue:44734", + "right": "issue:45103", + "accept": false, + "reason": "A serving KV-cache indexing crash and an auto-docstring annotation crash are unrelated." }, { - "left": "issue:44734", - "right": "issue:45103", + "left": "issue:44779", + "right": "issue:44568", "accept": false, - "reason": "Serving KV-cache indexing crash and an auto_docstring annotations crash are completely different subsystems." + "reason": "Both are tokenizer regressions, but they involve different models and different tokenization behaviors." }, { - "left": "issue:30333", - "right": "issue:33290", + "left": "issue:44792", + "right": "issue:44964", "accept": false, - "reason": "MLflow reporting on failed jobs and DeepSpeed OOM with Adafactor are unrelated training issues." + "reason": "Different multimodal models and different failure modes; not mergeable into one fix." }, { - "left": "issue:42898", - "right": "issue:45356", + "left": "issue:45092", + "right": "issue:45137", "accept": false, - "reason": "Both are tokenizer regressions, but one is about cleanup_spaces behavior while the other is codec handling and a warning mismatch for Kimi-K2.5." + "reason": "Remote-code/meta-init incompatibility is unrelated to a ZeRO3 deque error." }, { - "left": "issue:43723", - "right": "issue:44568", + "left": "issue:45103", + "right": "issue:44734", "accept": false, - "reason": "Tokenizer loading in v5 and add_special_tokens BOS/EOS behavior are separate problems." + "reason": "Reversed pair of the same two unrelated issues; still not a duplicate." }, { - "left": "issue:44060", - "right": "issue:45125", + "left": "issue:45125", + "right": "issue:45310", "accept": false, - "reason": "Incorrect tied-weights warning and missing _tp_plan for tensor parallelism are different Qwen3.5 MoE issues." + "reason": "Both mention Qwen3.5 MoE, but one is tensor-parallel metadata and the other is a from_pretrained failure; too different to treat as one bug." }, { - "left": "issue:45405", - "right": "issue:45712", + "left": "issue:45356", + "right": "issue:42898", "accept": false, - "reason": "A dependency version pin issue is unrelated to leftover dummy classes and dir(transformers) leakage." + "reason": "Two separate tokenizer regressions affecting different models and behaviors." }, { - "left": "issue:42898", - "right": "issue:44779", + "left": "issue:45405", + "right": "issue:45712", "accept": false, - "reason": "Both are tokenizer regressions, but they affect different models and behavior changes; no single shared bug is evident." + "reason": "A dependency version bump issue is unrelated to dummy class cleanup." }, { "left": "issue:43874", "right": "issue:45092", "accept": false, - "reason": "Missing get_number_of_image_patches in GLM46V and InternVL2 meta-initialization incompatibility are distinct multimodal loading issues." + "reason": "Different multimodal model families and different missing-method/meta-init problems." }, { "left": "issue:42898", "right": "issue:44448", "accept": false, - "reason": "Tokenizer cleanup behavior and Pegasus output differences across v4/v5 are related only at a broad version level, not the same concrete bug." + "reason": "A tokenizer behavior change and a model output mismatch across v4/v5 are not the same defect." }, { "left": "issue:43479", "right": "issue:44079", "accept": false, - "reason": "Phi4 multimodal config initialization and ModelOutput key assignment are unrelated object-construction bugs." + "reason": "Default config initialization and ModelOutput key assignment are unrelated internals." }, { "left": "issue:44188", "right": "issue:44964", "accept": false, - "reason": "Attention kernel divergence under torch.compile is unrelated to a Phi-4 multimodal model loading failure." + "reason": "Compile-dependent attention divergence does not match a model loading failure." }, { "left": "issue:44295", "right": "issue:45656", "accept": false, - "reason": "A position_ids buffer read error and an optimizer step being called twice under DeepSpeed are different training/runtime issues." + "reason": "A position_ids buffer access error and double optimizer stepping under DeepSpeed are different bugs." } ] }, @@ -11204,8 +11112,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -11272,15 +11180,15 @@ "issue:44222|issue:44623", "issue:30333|issue:36331", "issue:41628|issue:42757", - "issue:43723|issue:44977", "issue:30333|issue:34689", + "issue:43723|issue:44977", "issue:44297|issue:45310", "issue:28282|issue:30333", "issue:30333|issue:35141", - "issue:44483|issue:45406", - "issue:43873|issue:44945", "issue:44222|issue:44484", + "issue:44483|issue:45406", "issue:44263|issue:44360", + "issue:43873|issue:44945", "issue:44263|issue:45092", "issue:44514|issue:45103", "issue:43479|issue:43644", @@ -11294,232 +11202,232 @@ "issue:41628|issue:43824", "issue:42886|issue:42994" ], - "prepared_review_unit_hash": "1aed5c76ca6ebca6657ea0edd2e27ba5f6ae0d5a487ff4f0ea4bc41460aab8ca", + "prepared_review_unit_hash": "14f28dffa56d704361441560d585f26a54f614daa74bb28b78392701f258f765", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 34, - "serialized_chars": 37707, - "estimated_input_tokens": 9427, - "estimated_eval_tokens": 19110 + "serialized_chars": 37709, + "estimated_input_tokens": 9428, + "estimated_eval_tokens": 19112 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is highly heterogeneous: most items are unrelated Transformers bugs across imports, tokenization, saving/loading, quantization, and training. I do not see a single canonical duplicate target, and none of the soft pairs look like true duplicates.", - "confidence": 0.91, - "canonical_issue_reason": null, + "summary": "This cluster is heterogeneous: most items are unrelated Transformers regressions spanning imports, tokenizers, multimodal processors, training, quantization, and save/load behavior. The soft-similarity pairs share broad themes only, not the same concrete bug or change, so none should be merged.", + "confidence": 0.82, + "canonical_issue_reason": "issue:41628 is the most representative user-facing import regression in the set and sits near several other import/load-related reports, making it the best single issue to anchor the cluster.", "canonical_pr_reason": null, - "best_issue_reason": null, + "best_issue_reason": "issue:41628 is the broadest and clearest representative of the import/load subset; it is actionable and more general than the model-specific or training-specific items around it.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:42898", "right": "issue:43295", "accept": false, - "reason": "Different regressions: tokenizer cleanup behavior vs processor/tokenizer attribute access and batched image handling." + "reason": "Both are v5 regressions, but one is tokenization cleanup behavior and the other is processor/tokenizer attribute access for custom model code; different code paths." }, { "left": "issue:44933", "right": "issue:45712", "accept": false, - "reason": "Both involve imports, but one is a missing import from image_utils and the other is dummy PT objects leaking into dir(transformers); not the same bug." + "reason": "Both involve image_utils-related symbols, but one is a bad import target while the other is dummy PT classes leaking into dir() without torch; not the same bug." }, { "left": "issue:42175", "right": "issue:44933", "accept": false, - "reason": "Backend extra installation problem vs a missing image_utils import; unrelated code paths." + "reason": "TensorFlow backend packaging and a missing image_utils import are different dependency/import problems." }, { "left": "issue:41628", "right": "issue:45459", "accept": false, - "reason": "Importing AutoImageProcessor is a public API export issue; hidden tokenizer errors without protobuf is an error-handling bug." + "reason": "Different missing symbols in different modules: transformers import exposure vs tokenizer protobuf error handling." }, { "left": "issue:43976", "right": "issue:44933", "accept": false, - "reason": "Python version compatibility regression vs import exposure in image_utils; no shared underlying defect." + "reason": "Python version compatibility regression is unrelated to the image_utils import issue." }, { "left": "issue:44222", "right": "issue:44297", "accept": false, - "reason": "Both touch save_pretrained, but one is FP8 model saving and the other is tokenizer metadata serialization; different artifacts and fixes." + "reason": "FP8 save_pretrained failure and tokenizer save metadata mismatch affect different save paths." }, { "left": "issue:44297", "right": "issue:45656", "accept": false, - "reason": "Tokenizer config serialization bug vs optimizer stepping twice under deepspeed; unrelated." + "reason": "Tokenizer config serialization and deepspeed optimizer-step behavior are unrelated." }, { "left": "issue:43479", "right": "issue:44869", "accept": false, - "reason": "Multimodal config initialization vs Whisper timestamp decoding crash; different subsystems and failures." + "reason": "Multimodal config default initialization and Whisper timestamp decoding are different model-specific bugs." }, { "left": "issue:44206", "right": "issue:45103", "accept": false, - "reason": "Feature extractor argument mismatch vs auto_docstring annotation parsing crash; not the same code path." + "reason": "Feature extractor argument regression and auto_docstring annotation crash are separate code paths." }, { "left": "issue:42831", "right": "issue:44623", "accept": false, - "reason": "FP8 accuracy issue vs missing processor save files; different symptom and fix surface." + "reason": "FP8 accuracy issues and processor.save_pretrained file omissions are not the same underlying defect." }, { "left": "issue:44222", "right": "issue:44623", "accept": false, - "reason": "Both mention saving, but one is model quantization save logic and the other is processor serialization; not the same bug." + "reason": "Both affect saving, but FP8/moe serialization and processor artifact saving are distinct failures." }, { "left": "issue:30333", "right": "issue:36331", "accept": false, - "reason": "Training job/MLflow reporting failure vs trainer loss-signature TypeError; separate trainer concerns." + "reason": "MLflow reporting on failed jobs and a Trainer compute_loss signature break are unrelated." }, { "left": "issue:41628", "right": "issue:42757", "accept": false, - "reason": "Different import failures from different dependencies/modules; no shared root cause." + "reason": "Missing AutoImageProcessor import and missing is_offline_mode from huggingface_hub are different import regressions." }, { - "left": "issue:43723", - "right": "issue:44977", + "left": "issue:30333", + "right": "issue:34689", "accept": false, - "reason": "Tokenizer loading regression vs flash-attention generation bug; different inference stages." + "reason": "MLflow status reporting and Llama 3.2 model loading are unrelated bugs." }, { - "left": "issue:30333", - "right": "issue:34689", + "left": "issue:43723", + "right": "issue:44977", "accept": false, - "reason": "MLflow reporting on failed training jobs is unrelated to model loading regression for a vision model." + "reason": "Tokenizer loading from cache and flash-attention generation errors are different failure modes." }, { "left": "issue:44297", "right": "issue:45310", "accept": false, - "reason": "Tokenizer class metadata mismatch vs Qwen3.5 from_pretrained regression; similar area, but not the same concrete bug." + "reason": "Tokenizer save metadata mismatch and Qwen3.5 MoE from_pretrained regression are separate issues." }, { "left": "issue:28282", "right": "issue:30333", "accept": false, - "reason": "PyTorch import dependency error vs MLflow reporting behavior; unrelated." + "reason": "PyTorch missing ImportError and MLflow job reporting are not the same problem." }, { "left": "issue:30333", "right": "issue:35141", "accept": false, - "reason": "Training telemetry issue vs embedding resize/post_init reinitialization bug; different code paths." + "reason": "Failed-job reporting and embedding resize/post_init reinitialization are unrelated." }, { - "left": "issue:44483", - "right": "issue:45406", + "left": "issue:44222", + "right": "issue:44484", "accept": false, - "reason": "OpenAI-style chat-completions API rejection vs Gemma4Processor server crash due to missing _tokenizer; different server failures." + "reason": "FP8 save failure and a max_shard_size behavior question are not duplicates." }, { - "left": "issue:43873", - "right": "issue:44945", + "left": "issue:44483", + "right": "issue:45406", "accept": false, - "reason": "Quantization offloading bug vs pipeline-parallelism output correctness issue; unrelated inference behavior." + "reason": "Chat completions request handling and Gemma4Processor _tokenizer access crash are different API paths." }, { - "left": "issue:44222", - "right": "issue:44484", + "left": "issue:44263", + "right": "issue:44360", "accept": false, - "reason": "FP8 serialization issue vs max_shard_size default question; no shared defect." + "reason": "Torch.split indexing behavior and a missing ReLU in the DSA indexer are different bugs in different logic." }, { - "left": "issue:44263", - "right": "issue:44360", + "left": "issue:43873", + "right": "issue:44945", "accept": false, - "reason": "Same broad DSA indexer area, but one is torch.split return handling and the other is a missing ReLU; not the same underlying bug." + "reason": "Quantization offloading and pipeline parallelism output corruption are separate distributed-inference issues." }, { "left": "issue:44263", "right": "issue:45092", "accept": false, - "reason": "GLM DSA indexer bug vs remote-code/meta-init incompatibility for InternVL2; unrelated." + "reason": "torch.split return handling and remote-code meta initialization incompatibility do not share the same concrete defect." }, { "left": "issue:44514", "right": "issue:45103", "accept": false, - "reason": "Processor chat-template batching crash vs auto_docstring annotation crash; different components." + "reason": "Batched chat template padding crash and auto_docstring annotation crash are unrelated processor/docstring issues." }, { "left": "issue:43479", "right": "issue:43644", "accept": false, - "reason": "Config default initialization bug vs non-persistent buffer initialization corruption; different mechanisms." + "reason": "Default multimodal config initialization and non-persistent buffer junk are different initialization bugs." }, { "left": "issue:44373", "right": "issue:44485", "accept": false, - "reason": "Docstring typo vs RoPE implementation bug; not duplicates." + "reason": "Wrong docstring for position_ids and GLM-5 RoPE implementation are not the same bug." }, { "left": "issue:44295", "right": "issue:44623", "accept": false, - "reason": "Position_ids buffer access problem vs missing processor save files; unrelated serialization/runtime issues." + "reason": "Buffer readback after registration and processor.save_pretrained missing files are different serialization/storage issues." }, { "left": "issue:43723", "right": "issue:44871", "accept": false, - "reason": "Tokenizer cache/loading regression vs eos_token_id config inconsistency; distinct tokenizer/model config issues." + "reason": "Tokenizer loading failure and eos_token_id config inconsistency are distinct tokenizer-related bugs." }, { "left": "issue:45092", "right": "issue:45357", "accept": false, - "reason": "Different Qwen/InternVL model-family regressions and different failure modes; not mergeable as one fix." + "reason": "Transformers v5 meta-init incompatibility and incorrect visual encoder keys on save_pretrained are separate model lifecycle problems." }, { "left": "issue:44261", "right": "issue:45092", "accept": false, - "reason": "MLA layernorm epsilon precision issue vs meta-init incompatibility for old InternVL2 remote-code checkpoints; unrelated." + "reason": "MLA layernorm epsilon precision error and remote-code meta initialization incompatibility are unrelated." }, { "left": "issue:30990", "right": "issue:38175", "accept": false, - "reason": "Loading hang vs unexpected zero probabilities; different runtime symptoms and likely causes." + "reason": "Sentence Transformers loading hang and zero probabilities in siglip2-base-patch16-224 are different model behaviors." }, { "left": "issue:41628", "right": "issue:43824", "accept": false, - "reason": "Both are import errors, but they affect different symbols and modules; too broad to be the same bug." + "reason": "Both are import errors, but they concern different symbols and model families; not the same missing code path." }, { "left": "issue:42886", "right": "issue:42994", "accept": false, - "reason": "Offline tokenizer cache loading vs quantized model saving failure; separate features and fixes." + "reason": "Offline tokenizer cache loading and quantized model saving failures are unrelated." } ] }, @@ -11538,8 +11446,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -11595,8 +11503,8 @@ "soft_pairs": [ "issue:22355|issue:28282", "issue:42831|issue:45310", - "issue:44945|issue:45753", "issue:44263|issue:44373", + "issue:44945|issue:45753", "issue:42831|issue:44038", "issue:42757|issue:45278", "issue:43576|issue:43784", @@ -11618,220 +11526,227 @@ "issue:43784|issue:45600", "issue:42907|issue:43937", "issue:41628|issue:45278", - "issue:44360|issue:44485", "issue:44261|issue:44373", + "issue:44360|issue:44485", "issue:43232|issue:44743", "issue:43116|issue:45200", - "issue:44625|issue:44829" + "issue:44625|issue:44829", + "issue:43122|issue:44779" ], - "prepared_review_unit_hash": "4b26a9501599cc0e4f3aeafce773862cb48fb58200650947bfcf5b13dff6647e", + "prepared_review_unit_hash": "5c489127a4c42cad715d8a6e163ae4dda5e9d789bbe6f427a4da87f76567a1d2", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, - "soft_pair_count": 30, - "serialized_chars": 36295, - "estimated_input_tokens": 9074, - "estimated_eval_tokens": 18404 + "soft_pair_count": 31, + "serialized_chars": 36587, + "estimated_input_tokens": 9147, + "estimated_eval_tokens": 18550 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is highly heterogeneous: most items are distinct issue reports, and none of the soft pairs looks like a true duplicate or one concrete fixable bug/change. The closest shared theme is v5-era import/loading regressions, but the individual reports still target different symbols and code paths.", - "confidence": 0.72, - "canonical_issue_reason": "issue:45278 is the broadest umbrella report in the set, explicitly summarizing multiple import failures after a version upgrade, so it is the best representative issue for this mixed cluster.", + "summary": "This looks like a loose bundle of mostly unrelated Transformers issues, not a tight duplicate cluster. No soft edge is convincing enough to merge; the closest theme is the broad import-regression report, which is the best anchor if one must be chosen.", + "confidence": 0.63, + "canonical_issue_reason": "issue:45278 is the broadest umbrella-style report in the set (many import errors after upgrading), and it is the closest match to the other import-related regressions.", "canonical_pr_reason": null, - "best_issue_reason": "issue:45278 is the most general, high-signal issue here and the closest thing to a cluster representative, even though the surrounding reports are not true duplicates.", + "best_issue_reason": "issue:45278 is the best global representative because it is the most general and likely to subsume other import-related failures in this bundle.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:22355", "right": "issue:28282", "accept": false, - "reason": "Different import errors in different areas: ONNX module missing vs PyTorch dependency detection." + "reason": "Both are import/load failures, but one is about transformers.onnx and the other about missing PyTorch; different root causes." }, { "left": "issue:42831", "right": "issue:45310", "accept": false, - "reason": "Unrelated bugs: FineGrainedFP8 accuracy regression vs Qwen3.5 MoE from_pretrained failure." + "reason": "FineGrainedFP8 accuracy drift and a Qwen3.5 MOE from_pretrained error are unrelated problems." }, { - "left": "issue:44945", - "right": "issue:45753", + "left": "issue:44263", + "right": "issue:44373", "accept": false, - "reason": "Different problem types and code paths: pipeline-parallel output corruption vs image-specific infinite loop." + "reason": "A torch.split return-value issue and a docstring bug are not the same underlying defect." }, { - "left": "issue:44263", - "right": "issue:44373", + "left": "issue:44945", + "right": "issue:45753", "accept": false, - "reason": "Different fixes: DSA indexer torch.split behavior vs a docstring issue for position_ids." + "reason": "Pipeline-parallelism output corruption and a specific Qwen3_5 image infinite loop are distinct model/runtime bugs." }, { "left": "issue:42831", "right": "issue:44038", "accept": false, - "reason": "Distinct model-specific issues with no clear shared underlying bug." + "reason": "Different symptoms and code paths: FP8 accuracy vs Qwen3-VL-Moe loading/usage." }, { "left": "issue:42757", "right": "issue:45278", "accept": false, - "reason": "Both are import-related, but one is a specific missing hub symbol and the other is a broad migration report; not the same concrete bug." + "reason": "45278 is a broad import-regression report, but 42757 is a specific huggingface_hub symbol import error; not enough to call them the same bug." }, { "left": "issue:43576", "right": "issue:43784", "accept": false, - "reason": "Unrelated: broken env CLI command vs sentence-transformers import NameError." + "reason": "An env command regression and a sentence-transformers import NameError are different failures." }, { "left": "issue:42886", "right": "issue:43066", "accept": false, - "reason": "Different tokenizer bugs: offline cache loading vs wrong decoder type in v5." + "reason": "Offline cache loading and wrong tokenizer decoder type are different tokenizer issues." }, { "left": "issue:43232", "right": "issue:45593", "accept": false, - "reason": "Generation cache handling is unrelated to D-FINE denoising/auxiliary loss behavior." + "reason": "Generation state handling after sync_gpus is unrelated to D-FINE auxiliary-loss behavior." }, { "left": "issue:44117", "right": "issue:44704", "accept": false, - "reason": "TOKENIZER_MAPPING_NAMES fallback bug vs AutoProcessor kwargs passthrough bug are separate code paths." + "reason": "TOKENIZER_MAPPING_NAMES/ from_pretrained logic differs from AutoProcessor kwargs forwarding." }, { "left": "issue:30990", "right": "issue:33357", "accept": false, - "reason": "A loading hang is not the same as a MacOS bus error." + "reason": "Sentence Transformers loading hang and a MacOS bus error on CLIP are separate issues." }, { "left": "issue:42371", "right": "issue:44263", "accept": false, - "reason": "TF32 API guidance vs DSA indexer split semantics are unrelated." + "reason": "TF32 settings guidance and torch.split behavior have no direct overlap." }, { "left": "issue:44704", "right": "issue:45290", "accept": false, - "reason": "Processor kwargs forwarding is unrelated to chat template crashes with tool-call messages." + "reason": "Cached-file kwargs forwarding is unrelated to apply_chat_template crashing on tool-call assistant messages." }, { "left": "issue:43756", "right": "issue:44060", "accept": false, - "reason": "Different model architecture concerns: RoPE layer handling vs tied-weights warning in Qwen3-Next." + "reason": "Smollm3 RoPE layer handling and Qwen3-Next tied-weights warnings are different model-specific bugs." }, { "left": "issue:42907", "right": "issue:44945", "accept": false, - "reason": "Saving dequantized models is not the same bug as incorrect pipeline-parallel outputs." + "reason": "Saving dequantized Ministrals and incorrect pipeline-parallel output are unrelated." }, { "left": "issue:43122", "right": "issue:44514", "accept": false, - "reason": "Tokenization differences across versions vs batched apply_chat_template padding crash are distinct." + "reason": "Both touch tokenization/chat templating, but one is a version-to-version tokenization difference and the other is a batched multimodal crash." }, { "left": "issue:44295", "right": "issue:44297", "accept": false, - "reason": "Reading a registered buffer is unrelated to tokenizer_class metadata written by save_pretrained." + "reason": "Reading a registered position_ids buffer and saving tokenizer_class metadata are unrelated." }, { "left": "issue:43232", "right": "issue:45245", "accept": false, - "reason": "Generation kwargs update logic is unrelated to the categories limit runtime error." + "reason": "Generation cache behavior and the 2^24 category limit error are not the same bug." }, { "left": "issue:43723", "right": "issue:44779", "accept": false, - "reason": "Tokenizer loading failure vs incorrect Deepseek tokenization are related only broadly, not the same underlying bug." + "reason": "A generic tokenizer load issue and incorrect DeepSeek tokenization are different failure modes." }, { "left": "issue:43408", "right": "issue:43475", "accept": false, - "reason": "Model-type mismatch warning and missing attribute in SAM3 are different failures." + "reason": "Model-type mismatch warning and missing fpn_position_embeddings are related to the same family, but not the same concrete defect." }, { "left": "issue:43723", "right": "issue:44442", "accept": false, - "reason": "Both involve tokenizer loading, but they are different tokenizers and different failure modes." + "reason": "Generic AutoTokenizer load regression and FastSpeech2ConformerTokenizer load failure are distinct." }, { "left": "issue:43122", "right": "issue:44521", "accept": false, - "reason": "Different chat-template bugs: tokenization inconsistency vs all-zero assistant masks." + "reason": "Different tokenization across versions and zero assistant masks for multimodal inputs are not the same issue." }, { "left": "issue:43784", "right": "issue:45600", "accept": false, - "reason": "Sentence-transformers import NameError is unrelated to Sam3LiteText auto-mapping references." + "reason": "Sentence-transformers import NameError and broken auto_mappings CI are unrelated." }, { "left": "issue:42907", "right": "issue:43937", "accept": false, - "reason": "Dequantization save behavior is unrelated to invalid GenerationConfig." + "reason": "Dequantized save behavior and invalid GenerationConfig for GLM-5 are separate problems." }, { "left": "issue:41628", "right": "issue:45278", "accept": false, - "reason": "Although both mention import problems after upgrade, one is a single missing symbol and the other is a broad bundle of failures; not one concrete duplicate." + "reason": "Both are import-related, but 41628 is a single missing AutoImageProcessor import while 45278 is a broad upgrade regression; not enough evidence of one root cause." }, { - "left": "issue:44360", - "right": "issue:44485", + "left": "issue:44261", + "right": "issue:44373", "accept": false, - "reason": "DSA indexer missing ReLU is unrelated to GLM-5 RoPE implementation." + "reason": "A missing rms_norm_eps config value and a wrong docstring do not share the same bug." }, { - "left": "issue:44261", - "right": "issue:44373", + "left": "issue:44360", + "right": "issue:44485", "accept": false, - "reason": "A precision/config issue in MLA is unrelated to a docstring correction." + "reason": "DSA indexer activation choice and GLM-5 RoPE implementation are unrelated." }, { "left": "issue:43232", "right": "issue:44743", "accept": false, - "reason": "Different generation bugs: sync_gpus kwargs handling vs recurrent state reset with cache." + "reason": "Both involve generation/cache semantics, but the sync_gpus bug and recurrent-state reset bug are different code paths." }, { "left": "issue:43116", "right": "issue:45200", "accept": false, - "reason": "Example script multi-label output bug is unrelated to mm_token_type_ids defaults." + "reason": "Example-script multi-label output is unrelated to mm_token_type_ids defaults for Gemma 4 fine-tuning." }, { "left": "issue:44625", "right": "issue:44829", "accept": false, - "reason": "num_labels propagation and flash_attention_3 degenerate training are separate issues." + "reason": "num_labels propagation and flash_attention_3 degenerate training are different sequence-classification issues." + }, + { + "left": "issue:43122", + "right": "issue:44779", + "accept": false, + "reason": "Generic version-to-version tokenization changes and DeepSeek-specific incorrect tokenization are not clearly the same defect." } ] }, @@ -11850,8 +11765,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -11859,8 +11774,8 @@ "issue:30990", "issue:33453", "issue:36296", + "issue:38617", "issue:39401", - "issue:42175", "issue:42222", "issue:42371", "issue:42757", @@ -11893,11 +11808,11 @@ "issue:44704", "issue:44779", "issue:44936", - "issue:44938", "issue:44964", "issue:44977", "issue:45083", "issue:45216", + "issue:45278", "issue:45446", "issue:45478", "issue:45563", @@ -11905,7 +11820,6 @@ ], "soft_pairs": [ "issue:29127|issue:30990", - "issue:43122|issue:44779", "issue:42907|issue:44964", "issue:44038|issue:44297", "issue:42907|issue:43452", @@ -11935,249 +11849,243 @@ "issue:43116|issue:43493", "issue:43064|issue:43673", "issue:45083|issue:45446", - "issue:42175|issue:44938", "issue:43937|issue:45563", "issue:33453|issue:39401", - "issue:39401|issue:44779" + "issue:39401|issue:44779", + "issue:38617|issue:45278" ], - "prepared_review_unit_hash": "3029b93a2fd762a3667154c3d6b7f3e91890d9f56a3ebfe07c54225266ddd4ec", + "prepared_review_unit_hash": "47783cbf09764d6fe4b3847189916d7374639fb0d3e5c6d7deffa2ff6333faac", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 47, "item_count": 47, - "soft_pair_count": 35, - "serialized_chars": 37060, - "estimated_input_tokens": 9265, - "estimated_eval_tokens": 18786 + "soft_pair_count": 34, + "serialized_chars": 36776, + "estimated_input_tokens": 9194, + "estimated_eval_tokens": 18644 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This looks like a heterogeneous grab-bag of Transformers issues, with a few tokenization/loading/version-regression themes but no strong evidence that any pair is the same underlying bug. The soft links are mostly topic-level similarity, not true duplicates.", - "confidence": 0.86, - "canonical_issue_reason": "issue:43122 is the broadest tokenizer-regression report in the set and is the closest thing to a representative issue, but the overall collection is still too mixed to treat as one duplicate family.", + "summary": "This cluster is heterogeneous: import errors, tokenizer regressions, model-specific loading/saving issues, training bugs, and unrelated warnings. None of the soft edges look like the same concrete bug or change, so all candidate duplicate links should be rejected.", + "confidence": 0.87, + "canonical_issue_reason": "Issue 45278 is the broadest umbrella for the import-error subset, but the overall cluster is too mixed to justify a true canonical duplicate.", "canonical_pr_reason": null, - "best_issue_reason": "issue:43122 is the best single representative of the tokenization-regression subgroup because it is generic and not tied to one specific model family.", + "best_issue_reason": "45278 is the best representative only for the import-regression subset; it is broader than the others and most likely to absorb similar reports.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:29127", "right": "issue:30990", "accept": false, - "reason": "Different subsystems and symptoms: LayoutLMv3 error messaging vs Sentence Transformers loading hang." - }, - { - "left": "issue:43122", - "right": "issue:44779", - "accept": false, - "reason": "Both are tokenizer regressions in v5, but they concern different models and concrete failure modes; not enough to call the same bug." + "reason": "Different subsystems and symptoms: layoutlmv3 box validation vs sentence-transformers loading hang." }, { "left": "issue:42907", "right": "issue:44964", "accept": false, - "reason": "One is a save/dequantization problem, the other is a model load failure; unrelated code paths." + "reason": "Both are model-loading/saving related, but they are different model families and failure modes." }, { "left": "issue:44038", "right": "issue:44297", "accept": false, - "reason": "Qwen3-VL-Moe from_pretrained failure vs tokenizer save_pretrained metadata mismatch are different bugs." + "reason": "Qwen3-VL-Moe runtime bug vs tokenizer save metadata mismatch; not the same code path." }, { "left": "issue:42907", "right": "issue:43452", "accept": false, - "reason": "Dequantized model saving and gguf_file loading breakage are distinct loading/saving paths." + "reason": "Dequantized save failure vs gguf/AutoTokenizer loading breakage are different bugs." }, { "left": "issue:44038", "right": "issue:44295", "accept": false, - "reason": "Model loading failure and position_ids buffer read error are not the same underlying issue." + "reason": "Model-specific generation/load issue vs position_ids buffer read error; unrelated." }, { "left": "issue:39401", "right": "issue:44488", "accept": false, - "reason": "Tokenizer offset_mapping regression vs a specific model load failure; too different to merge." + "reason": "Tokenizer offset mapping bug vs inability to load a specific model checkpoint." }, { "left": "issue:36296", "right": "issue:43116", "accept": false, - "reason": "Tensor-parallel training optimizer bug vs example-script multi-label output issue are unrelated." + "reason": "Tensor-parallel training state bug vs example-script multi-label inference issue." }, { "left": "issue:43824", "right": "issue:44351", "accept": false, - "reason": "Importing a specific model class and importing HybridCache are separate API-export problems." + "reason": "Both are import errors, but for different missing symbols from different modules." }, { "left": "issue:44038", "right": "issue:44222", "accept": false, - "reason": "Qwen3-VL-Moe from_pretrained error and FP8 save_pretrained moe issue affect different paths." + "reason": "Qwen3-VL-Moe bug vs FP8 save_pretrained bug; same broad area, different concrete failure." }, { "left": "issue:43066", "right": "issue:44779", "accept": false, - "reason": "Both involve tokenizers in v5, but one is decoder type selection and the other is incorrect Deepseek tokenization; not clearly the same bug." + "reason": "Different tokenizer regressions affecting different models and code paths." }, { "left": "issue:39401", "right": "issue:44521", "accept": false, - "reason": "Wrong offset_mapping and all-zero assistant_masks are different tokenizer/template behaviors." + "reason": "Tokenizer offset mapping bug vs multimodal chat-template assistant mask bug." }, { "left": "issue:45216", "right": "issue:45478", "accept": false, - "reason": "Both mention Qwen3.5/Moe save/load regressions, but one is checkpoint saving and the other is from_pretrained loading; not the same concrete bug." + "reason": "Both mention Qwen3.5 and v5.4+, but one is save_pretrained corruption and the other is from_pretrained failure." }, { "left": "issue:42757", "right": "issue:44351", "accept": false, - "reason": "Missing huggingface_hub symbol vs missing transformers symbol are separate import regressions." + "reason": "Import errors from different packages: huggingface_hub vs transformers." }, { "left": "issue:43122", "right": "issue:44977", "accept": false, - "reason": "Both are tokenizer regressions, but they are model-specific reports with different symptoms and no evidence of a shared root cause." + "reason": "Tokenization regression vs flash-attention generation failure; not the same bug." }, { "left": "issue:43992", "right": "issue:44704", "accept": false, - "reason": "Missing embed_tokens.weight in UMT5Encoder vs AutoProcessor kwargs forwarding are unrelated loading behaviors." + "reason": "Encoder weight loading issue vs AutoProcessor kwargs propagation issue." }, { "left": "issue:42907", "right": "issue:44222", "accept": false, - "reason": "Saving dequantized models and FP8 save_pretrained are different serialization bugs." + "reason": "Different save/pretrained bugs in different quantization formats and model families." }, { "left": "issue:43784", "right": "issue:45446", "accept": false, - "reason": "Importing sentence-transformers fails because of nn, while AuxRequest has a PyTorch version check issue; unrelated." + "reason": "Sentence-transformers import NameError vs flex_attention version check bug are unrelated." }, { "left": "issue:43937", "right": "issue:44485", "accept": false, - "reason": "GenerationConfig validation error vs RoPE implementation discussion are not the same bug." + "reason": "GenerationConfig validation problem vs RoPE implementation discussion; different concrete issues." }, { "left": "issue:42222", "right": "issue:43441", "accept": false, - "reason": "Broken vitpose model loading and Ministral FlashAttention failure are distinct model-specific regressions." + "reason": "vitpose breakage vs Ministral FlashAttention failure; unrelated model paths." }, { "left": "issue:42371", "right": "issue:43874", "accept": false, - "reason": "TF32 API guidance issue and missing image-patch method in GLM46V are unrelated." + "reason": "TF32 API deprecation notice vs missing image-patch helper; not the same bug." }, { "left": "issue:43976", "right": "issue:45446", "accept": false, - "reason": "Python version compatibility problem vs a flex_attention version check bug are different compatibility issues." + "reason": "Python version compatibility problem vs a specific PyTorch import/version check." }, { "left": "issue:43493", "right": "issue:45753", "accept": false, - "reason": "SigLIP2 implementation discrepancy and Qwen3_5 infinite-loop behavior are unrelated model bugs." + "reason": "SigLIP2 implementation discrepancy vs Qwen3_5 infinite-loop-on-image issue." }, { "left": "issue:42831", "right": "issue:45753", "accept": false, - "reason": "FineGrainedFP8 accuracy regression and a Qwen3_5 generation loop are different failure modes." + "reason": "FP8 accuracy issue vs multimodal generation loop are different failures." }, { "left": "issue:44075", "right": "issue:44936", "accept": false, - "reason": "Optimizer argument handling and evaluate-after-train behavior are not the same training bug." + "reason": "Optimizer argument handling vs trainer.evaluate after train; both trainer-related but not the same bug." }, { "left": "issue:43116", "right": "issue:43825", "accept": false, - "reason": "Example-script multi-label output bug vs pipeline translation-task error messaging are unrelated." + "reason": "Empty multi-label results vs stale translation-task error message; unrelated example/pipeline issues." }, { "left": "issue:44222", "right": "issue:45478", "accept": false, - "reason": "Both mention FP8/moe or Qwen3.5-related failures, but one is save_pretrained and the other is from_pretrained loading; not enough for duplication." + "reason": "Both involve Qwen3.5/MoE, but one is FP8 save_pretrained and the other is from_pretrained loading." }, { "left": "issue:44038", "right": "issue:45478", "accept": false, - "reason": "Same broad model family, but different versions and failure surfaces; likely separate regressions." + "reason": "Different Qwen model variants and different failure modes; not mergeable as one fix." }, { "left": "issue:43116", "right": "issue:43493", "accept": false, - "reason": "Example script classification output vs SigLIP2 implementation discrepancy are unrelated." + "reason": "Classification example bug vs SigLIP2 model discrepancy are unrelated." }, { "left": "issue:43064", "right": "issue:43673", "accept": false, - "reason": "Optimizer-state mismatch under FSDP2/PEFT vs missing GenerationMixin cache in chunked_prefill are different training/generation paths." + "reason": "Trainer/FSDP optimizer-state bug vs GenerationMixin cache issue; different code paths." }, { "left": "issue:45083", "right": "issue:45446", "accept": false, - "reason": "Feature-extraction helper behavior in qwen3_omni_moe and a PyTorch version check issue are not the same problem." - }, - { - "left": "issue:42175", - "right": "issue:44938", - "accept": false, - "reason": "Missing tensorflow backend in pip install vs Python 3.14 load failure are separate environment/packaging issues." + "reason": "qwen3_omni_moe feature-length helper bug vs flex_attention import/version check bug." }, { "left": "issue:43937", "right": "issue:45563", "accept": false, - "reason": "GenerationConfig validation vs stale num_return_sequences warning are different generation issues." + "reason": "GLM-5 generation-config validation vs paged-generate warning text are unrelated." }, { "left": "issue:33453", "right": "issue:39401", "accept": false, - "reason": "Tokenizer loading regression and wrong offset_mapping are related at a high level, but the available evidence is too weak to merge them." + "reason": "Generic tokenizer loading regression vs Qwen3 offset_mapping bug; not the same defect." }, { "left": "issue:39401", "right": "issue:44779", "accept": false, - "reason": "Both are tokenizer regressions, but they affect different models and report different concrete wrong outputs." + "reason": "Both are tokenizer regressions, but they affect different models and outputs, so not one concrete bug." + }, + { + "left": "issue:38617", + "right": "issue:45278", + "accept": false, + "reason": "45278 is a broad import-error umbrella, but 38617 is a specific missing-symbol report and not clearly the same underlying bug." } ] }, @@ -12196,8 +12104,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -12251,7 +12159,6 @@ "issue:45656" ], "soft_pairs": [ - "issue:38617|issue:45278", "issue:44117|issue:45588", "issue:22355|issue:29127", "issue:43827|issue:45446", @@ -12265,8 +12172,8 @@ "issue:38617|issue:41628", "issue:44829|issue:45216", "issue:43408|issue:43742", - "issue:44410|issue:45310", "issue:42222|issue:45542", + "issue:44410|issue:45310", "issue:44291|issue:45071", "issue:42175|issue:42222", "issue:44484|issue:45310", @@ -12279,217 +12186,211 @@ "issue:38617|issue:44220", "issue:42371|issue:44261", "issue:42222|issue:45459", - "issue:44484|issue:45478", - "issue:44410|issue:45478" + "issue:44410|issue:45478", + "issue:44484|issue:45478" ], - "prepared_review_unit_hash": "37066426244f58d96b08b68e51c5c6c29abdd9497575f8445137f363547512a3", + "prepared_review_unit_hash": "1394739772ea01ba1b9c4f5cdbe55d98fc530467a74e2c5eea30d8e351a0e513", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, - "soft_pair_count": 30, - "serialized_chars": 36622, - "estimated_input_tokens": 9156, - "estimated_eval_tokens": 18568 + "soft_pair_count": 29, + "serialized_chars": 36336, + "estimated_input_tokens": 9084, + "estimated_eval_tokens": 18424 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "The set is mostly heterogeneous: the soft pairs share broad themes like import errors, tokenizer bugs, or model regressions, but they do not look like the same concrete underlying bug/change. I would not merge any of the provided soft edges.", - "confidence": 0.91, - "canonical_issue_reason": "issue:45310 is the clearest representative issue in this batch: it names a specific Qwen3.5 MoE from_pretrained regression with a concrete version boundary, and it appears to have a later near-identical duplicate (issue:45478).", + "summary": "This cluster is highly heterogeneous: most items are unrelated model, tokenizer, optimizer, or import-regression bugs. None of the soft pairs look like true duplicates, so all soft edges are rejected. The closest thing to a canonical issue is the broad upgrade/import-regression report, but overall the cluster is not a clean duplicate set.", + "confidence": 0.33, + "canonical_issue_reason": "Issue 45278 is the broadest umbrella report in the set, covering multiple import failures after upgrading to 5.5.0, but the cluster is too mixed for a strong canonical duplicate.", "canonical_pr_reason": null, - "best_issue_reason": "issue:45310 is the best single issue to anchor this set because it is specific, reproducible, and clearly tied to a distinct regression; it is also the most obviously duplicate-prone report here.", + "best_issue_reason": "45278 is the most general, user-facing issue and the best fit if one must anchor the cluster, though it still only matches a subset of the items.", "best_pr_reason": null, "soft_edge_verdicts": [ - { - "left": "issue:38617", - "right": "issue:45278", - "accept": false, - "reason": "Both are import-error reports after upgrades, but they fail on different symbols and different symptoms; not the same bug." - }, { "left": "issue:44117", "right": "issue:45588", "accept": false, - "reason": "Tokenization mapping fallback logic and a flash-attention crash are unrelated code paths." + "reason": "Different failure modes: tokenizer mapping fallback vs flash-attention crash on sink-less models." }, { "left": "issue:22355", "right": "issue:29127", "accept": false, - "reason": "A missing module import and a layoutlmv3 validation message issue are unrelated." + "reason": "Unrelated bugs: missing transformers.onnx module vs LayoutLMv3 error-message clarity." }, { "left": "issue:43827", "right": "issue:45446", "accept": false, - "reason": "Documentation references to pipeline() and a flex_attention version-check bug are different issues." + "reason": "Docs/pipeline removal mismatch is unrelated to the flex_attention PyTorch version check bug." }, { "left": "issue:43381", "right": "issue:44945", "accept": false, - "reason": "Gradient checkpointing in eval mode and incorrect output under pipeline parallelism are different behaviors." + "reason": "Gradient checkpointing in eval mode and incorrect pipeline-parallel output are different code paths." }, { "left": "issue:43937", "right": "issue:44977", "accept": false, - "reason": "GenerationConfig validation and flash-attention generation failure are not the same concrete defect." + "reason": "Both affect generation, but one is invalid GenerationConfig and the other is a Qwen3.5 flash-attention issue." }, { "left": "issue:43450", "right": "issue:43986", "accept": false, - "reason": "Batched video processor output shape and AutoProcessor without torchvision are separate problems." + "reason": "Batched video processor shape bug is unrelated to AutoProcessor crashing without torchvision." }, { "left": "issue:45405", "right": "issue:45446", "accept": false, - "reason": "A dependency version bump and a PyTorch import version check are unrelated." + "reason": "PEFT version pinning and a flex_attention import/version check are unrelated." }, { "left": "issue:39401", "right": "issue:43421", "accept": false, - "reason": "Wrong offset mappings and runtime post-processor updates are both tokenizer-related but not the same bug." + "reason": "Tokenizer offset_mapping regression is different from post-processor not updating after runtime special-token changes." }, { "left": "issue:44117", "right": "issue:45081", "accept": false, - "reason": "A None entry in TOKENIZER_MAPPING_NAMES and a Mistral regex patch crash are different failure modes." + "reason": "Tokenizer mapping None and Mistral regex patch backend_tokenizer errors are different bugs." }, { "left": "issue:42673", "right": "issue:45072", "accept": false, - "reason": "A VRAM leak in threaded inference and bfloat16 dtype mismatches are unrelated." + "reason": "VRAM leak across dataloader threads is unrelated to dtype mismatches in inference." }, { "left": "issue:38617", "right": "issue:41628", "accept": false, - "reason": "Both are import errors, but they concern different missing symbols and different affected APIs." + "reason": "Both are import errors, but they concern different missing symbols and different root causes." }, { "left": "issue:44829", "right": "issue:45216", "accept": false, - "reason": "Flash-attention training degeneration and a save_pretrained checkpoint regression are not the same issue." + "reason": "Flash-attention training degeneration is unrelated to a save_pretrained checkpoint regression." }, { "left": "issue:43408", "right": "issue:43742", "accept": false, - "reason": "A model-type mismatch warning and a key error while loading a model are different problems." + "reason": "Model-type warning for sam3_tracker is unrelated to a MobileLLM loading key error." }, { - "left": "issue:44410", - "right": "issue:45310", + "left": "issue:42222", + "right": "issue:45542", "accept": false, - "reason": "Missing projections in qwen3next and Qwen3.5 MoE from_pretrained loading failure are distinct defects." + "reason": "Broken vitpose models and missing TensorFlow backend from tensorboard-only install are unrelated." }, { - "left": "issue:42222", - "right": "issue:45542", + "left": "issue:44410", + "right": "issue:45310", "accept": false, - "reason": "Vitpose model breakage and a TensorFlow backend availability issue are unrelated." + "reason": "Missing Qwen3Next projections and Qwen3.5 MoE from_pretrained error are different model-specific loading bugs." }, { "left": "issue:44291", "right": "issue:45071", "accept": false, - "reason": "These are both v5 compatibility regressions, but they affect different APIs and code paths." + "reason": "init_empty_weights argument handling and PreTrainedConfig type checking are separate regressions." }, { "left": "issue:42175", "right": "issue:42222", "accept": false, - "reason": "Missing TensorFlow in a torch-only install and broken vitpose models are not the same bug." + "reason": "TensorFlow backend packaging issue and vitpose model breakage are unrelated." }, { "left": "issue:44484", "right": "issue:45310", "accept": false, - "reason": "save_pretrained shard-size behavior and Qwen3.5 MoE loading failure are unrelated." + "reason": "Chat/completions API failure is a different surface from Qwen3.5 MoE from_pretrained loading error." }, { "left": "issue:43937", "right": "issue:44496", "accept": false, - "reason": "GenerationConfig invalidity and an unrecognized model/config error are different loading paths." + "reason": "Invalid GenerationConfig and unrecognized model/config problems are different validation failures." }, { "left": "issue:44483", "right": "issue:45464", "accept": false, - "reason": "A /v1/chat/completions request-shape issue and a streaming inference failure on Qwen3.5 are not clearly the same defect." + "reason": "These both touch chat/completions, but one is a request-acceptance bug and the other is streaming inference failure." }, { "left": "issue:44075", "right": "issue:45656", "accept": false, - "reason": "SGD arguments being ignored and deepspeed calling optimizer.step() twice are different optimizer bugs." + "reason": "SGD args ignored and deepspeed double-step are distinct optimizer bugs." }, { "left": "issue:43066", "right": "issue:43122", "accept": false, - "reason": "Wrong decoder type and changed tokenization output are related to tokenizer regressions but not the same concrete issue." + "reason": "Both are tokenizer regressions, but one is wrong decoder type and the other is changed tokenization output." }, { "left": "issue:44964", "right": "issue:45278", "accept": false, - "reason": "A specific Phi-4 multimodal load failure and a broad set of import errors after upgrade are not the same bug." + "reason": "A single model-loading failure is not the same as the broad multi-import regression report." }, { "left": "issue:43012", "right": "issue:43577", "accept": false, - "reason": "A warning about bfloat16 compilation and a model remaining float32 on load are different dtype-related issues." + "reason": "Bfloat16 compilation warning and BLIP2 dtype remaining float32 are different dtype-related bugs." }, { "left": "issue:38617", "right": "issue:44220", "accept": false, - "reason": "An import error and an fbank feature-extraction issue are unrelated." + "reason": "Missing import symbol and _torch_extract_fbank_features failure are unrelated." }, { "left": "issue:42371", "right": "issue:44261", "accept": false, - "reason": "TF32 control API guidance and an rms_norm_eps precision problem are unrelated." + "reason": "TF32 behavior warning and MLA rms_norm_eps precision mismatch are unrelated." }, { "left": "issue:42222", "right": "issue:45459", "accept": false, - "reason": "Vitpose breakage and protobuf-related tokenizer error masking are different subsystems." + "reason": "Vitpose model breakage and protobuf-hidden tokenizer errors do not share the same underlying bug." }, { - "left": "issue:44484", + "left": "issue:44410", "right": "issue:45478", "accept": false, - "reason": "max_shard_size behavior and Qwen3.5 MoE from_pretrained loading failure are unrelated." + "reason": "Both mention Qwen models, but they are different loading failures with different model paths." }, { - "left": "issue:44410", + "left": "issue:44484", "right": "issue:45478", "accept": false, - "reason": "Missing qwen3next projections and Qwen3.5 MoE loading failure are different model bugs." + "reason": "Chat API request handling and Qwen3.5 MoE model loading are unrelated." } ] }, @@ -12508,8 +12409,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -12578,8 +12479,8 @@ "issue:43097|issue:43452", "issue:43931|issue:44038", "issue:43335|issue:45071", - "issue:43976|issue:44945", "issue:43408|issue:44368", + "issue:43976|issue:44945", "issue:43874|issue:44263", "issue:44829|issue:45245", "issue:38617|issue:42757", @@ -12599,160 +12500,160 @@ "issue:43502|issue:43992", "issue:43097|issue:43873" ], - "prepared_review_unit_hash": "a664ae640b2f15f54679694acf367432a9607c616c76425102cbca83f8780b0e", + "prepared_review_unit_hash": "05aff5baaaf86eb57dcafb34aa43f2b5ba3838cc3897b3275706e720027949a6", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 35, - "serialized_chars": 38164, - "estimated_input_tokens": 9541, - "estimated_eval_tokens": 19338 + "serialized_chars": 38166, + "estimated_input_tokens": 9542, + "estimated_eval_tokens": 19340 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is highly heterogeneous: it mixes import errors, model-loading regressions, flash-attention bugs, config warnings, and unrelated model-specific failures. I did not find any soft-edge pair that looks like the same underlying bug/change.", - "confidence": 0.96, - "canonical_issue_reason": "issue:43097 is the most cluster-central release-regression report by linkage/activity, so it\u2019s the best triage anchor among these unrelated issues.", + "summary": "Loose similarity cluster of many unrelated Transformers issues (import errors, model-specific load bugs, warnings, and training regressions). None of the soft pairs look like true duplicates.", + "confidence": 0.94, + "canonical_issue_reason": "No single canonical issue: the cluster spans multiple unrelated failure modes and model families, so one issue does not cleanly subsume the others.", "canonical_pr_reason": null, - "best_issue_reason": "issue:43097 is the best representative issue for this cluster because it is a concrete, user-facing v5 regression and has the strongest evidence of being a hub among the listed items.", + "best_issue_reason": "Issue 43097 is the closest representative anchor because it is a broad v5 regression and has inbound references, but it still does not explain the rest of the cluster.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43097", "right": "issue:44493", "accept": false, - "reason": "Both are transformer regressions, but one is about a removed config warning and the other is about unexpected position-id keys; different bug classes and code paths." + "reason": "Removed API warning vs unexpected position-id key warning; different regression and code path." }, { "left": "issue:43874", "right": "issue:44261", "accept": false, - "reason": "Different models and failures: missing image-patch helper in GLM46V vs a missing rms_norm_eps config field affecting MLA precision." + "reason": "Different models and failures: missing image-patch method vs missing rms_norm_eps in config." }, { "left": "issue:44263", "right": "issue:44485", "accept": false, - "reason": "One is a torch.split/indexing bug in GlmMoeDsaIndexer; the other is a GLM-5 RoPE implementation discussion. Not the same concrete defect." + "reason": "GlmMoeDsaIndexer split behavior and GLM-5 RoPE implementation are distinct bugs." }, { "left": "issue:43012", "right": "issue:43408", "accept": false, - "reason": "A PyTorch precision warning during bfloat16 compilation is unrelated to a model-type mismatch warning between sam3_video and sam3_tracker." + "reason": "Float32 precision warning during compile is unrelated to the sam3 tracker/video model-type warning." }, { "left": "issue:43441", "right": "issue:44829", "accept": false, - "reason": "Both mention flash attention, but they target different symptoms, models, and training/inference paths; too broad to merge." + "reason": "Both mention flash attention, but one is a load-time failure and the other is degenerate training; not the same bug." }, { "left": "issue:41628", "right": "issue:45071", "accept": false, - "reason": "Importing AutoImageProcessor and PretrainedConfig type checking are unrelated issues." + "reason": "Missing AutoImageProcessor import and PretrainedConfig type-checking breakage are unrelated issues." }, { "left": "issue:45071", "right": "issue:45542", "accept": false, - "reason": "Type-checking regression in PretrainedConfig is unrelated to the TensorFlow/tensorboard backend detection issue." + "reason": "Config type-checking regression vs undefined TF backend from tensorboard-only install; different causes." }, { "left": "issue:42673", "right": "issue:44977", "accept": false, - "reason": "VRAM leak in multi-threaded dataloaders is a different problem from Qwen3.5 flash-attention generation failures." + "reason": "VRAM leak in multi-threaded dataloading vs generation failure with flash-attention; distinct symptoms and paths." }, { "left": "issue:42371", "right": "issue:42831", "accept": false, - "reason": "TF32 API guidance and FineGrainedFP8 accuracy loss are distinct numeric/configuration issues." + "reason": "TF32 API-setting guidance and FineGrainedFP8 accuracy loss are different concerns." }, { "left": "issue:42673", "right": "issue:43906", "accept": false, - "reason": "The second is only an isolated reproduction of another issue, but the titles don\u2019t indicate the same concrete bug as the VRAM leak report." + "reason": "VRAM leak report vs reproduction of another issue; no concrete shared defect." }, { "left": "issue:43986", "right": "issue:44484", "accept": false, - "reason": "AutoProcessor crash without torchvision is unrelated to save_pretrained max_shard_size behavior." + "reason": "AutoProcessor/video-model crash without torchvision is unrelated to save_pretrained shard sizing." }, { "left": "issue:44117", "right": "issue:44987", "accept": false, - "reason": "A tokenizer mapping None-handling bug is not the same as loading failures for physical-intelligence/fast." + "reason": "TOKENIZER_MAPPING_NAMES None handling and a model load failure for physical-intelligence/fast are unrelated." }, { "left": "issue:43097", "right": "issue:43452", "accept": false, - "reason": "Removed tie_embeddings API behavior is unrelated to gguf_file loading failures." + "reason": "Config removal warning and gguf_file loading breakage are different failures." }, { "left": "issue:43931", "right": "issue:44038", "accept": false, - "reason": "Both involve Qwen3-VL family loading, but they describe different model variants and failure modes; not enough to treat as one bug." + "reason": "Qwen3-VL weight mismatch and Qwen3-VL-Moe loading bug are different model-specific issues." }, { "left": "issue:43335", "right": "issue:45071", "accept": false, - "reason": "SwitchTransformers sparse-layer creation bug is unrelated to a typing regression in PretrainedConfig." + "reason": "SwitchTransformers sparse-layer creation bug and config type-checking regression are unrelated." }, { - "left": "issue:43976", - "right": "issue:44945", + "left": "issue:43408", + "right": "issue:44368", "accept": false, - "reason": "Python version compatibility and incorrect output under pipeline parallelism are different issues." + "reason": "Different warnings on different models: sam3 tracker/type mismatch vs tie_word_embeddings guidance." }, { - "left": "issue:43408", - "right": "issue:44368", + "left": "issue:43976", + "right": "issue:44945", "accept": false, - "reason": "Model-type warning for SAM3 is unrelated to a tie_word_embeddings warning during Qwen3.5 LoRA fine-tuning." + "reason": "Python-version compatibility problem vs incorrect output under pipeline parallelism; not the same defect." }, { "left": "issue:43874", "right": "issue:44263", "accept": false, - "reason": "Different GLM subcomponents and different symptoms; one is an AttributeError in an image processor path, the other a torch.split/indexing issue." + "reason": "Missing image-patch method and torch.split return-value issue are separate GLM bugs." }, { "left": "issue:44829", "right": "issue:45245", "accept": false, - "reason": "Flash-attention training degeneration is unrelated to a categorical cardinality runtime limit." + "reason": "FlashAttention training degeneration and category-count runtime limit are unrelated." }, { "left": "issue:38617", "right": "issue:42757", "accept": false, - "reason": "Both are import errors, but they concern different symbols from different packages and different missing APIs." + "reason": "Both are import errors, but they involve different missing symbols from different packages." }, { "left": "issue:43064", "right": "issue:44315", "accept": false, - "reason": "Wrong optimizer states under FSDP2/PEFT is not the same underlying bug as Liger Kernel not being applied with model_init." + "reason": "FSDP2/PEFT optimizer-state mismatch and Liger Kernel not applied with model_init are separate training-stack issues." }, { "left": "issue:44464", @@ -12764,79 +12665,79 @@ "left": "issue:44410", "right": "issue:45375", "accept": false, - "reason": "Different Qwen configuration bugs affecting different fields and model families; not the same defect." + "reason": "Missing projections in qwen3next and missing deepstack_visual_indexes in config are distinct model/config bugs." }, { "left": "issue:44062", "right": "issue:44117", "accept": false, - "reason": "AddedToken argument duplication is unrelated to tokenizer mapping returning None." + "reason": "AddedToken keyword duplication and TOKENIZER_MAPPING_NAMES None are unrelated tokenizer problems." }, { "left": "issue:43097", "right": "issue:45588", "accept": false, - "reason": "Both are v5-era regressions, but one is about removed tie_embeddings guidance and the other is a flash_attention.py crash on sink-less models." + "reason": "Embedding-tying warning and flash_attention.py s_aux=None crash are different code paths." }, { "left": "issue:42673", "right": "issue:43577", "accept": false, - "reason": "VRAM leak in threaded dataloading is unrelated to dtype handling for BLIP2 loading." + "reason": "VRAM leak under dataloader threads and Blip2 dtype staying float32 are unrelated." }, { "left": "issue:38617", "right": "issue:44351", "accept": false, - "reason": "Different missing imports from different modules; no shared underlying bug." + "reason": "Different missing imports from different modules; same broad category only, not the same bug." }, { "left": "issue:44704", "right": "issue:45103", "accept": false, - "reason": "Passing kwargs to cached_file is unrelated to auto_docstring crashing with future annotations." + "reason": "Kwarg forwarding to cached_file and auto_docstring AttributeError are unrelated utility bugs." }, { "left": "issue:44038", "right": "issue:44410", "accept": false, - "reason": "Two different Qwen model issues with different missing/mismatched components; not one concrete bug." + "reason": "Qwen3-VL-Moe loading issue and qwen3next missing projections are different model failures." }, { "left": "issue:43931", "right": "issue:44410", "accept": false, - "reason": "A shape-mismatch loading error for Qwen3-VL-30B-A3B is unrelated to missing projections in qwen3next layer 0." + "reason": "Weight-shape mismatch for Qwen3-VL-30B and missing projections in qwen3next are not the same defect." }, { "left": "issue:43986", "right": "issue:44117", "accept": false, - "reason": "AutoProcessor/torchvision crash is unrelated to tokenizer mapping fallback behavior." + "reason": "Video-model torchvision crash and tokenizer-mapping None handling are unrelated." }, { "left": "issue:44261", "right": "issue:44485", "accept": false, - "reason": "Missing rms_norm_eps in MLA q_a_layernorm is unrelated to GLM-5 RoPE implementation questions." + "reason": "Missing rms_norm_eps in MLA q_a_layernorm and GLM-5 RoPE implementation are different issues." }, { "left": "issue:44964", "right": "issue:45341", "accept": false, - "reason": "Model loading failure for Phi-4 multimodal is unrelated to a testing_utils bug." + "reason": "Phi-4 multimodal load failure and a testing_utils bug have no shared underlying cause." }, { "left": "issue:43502", "right": "issue:43992", "accept": false, - "reason": "Offline-mode network requests are unrelated to missing embed_tokens.weight in UMT5Encoder loading." + "reason": "Unexpected API requests with local_files_only and missing embed_tokens.weight on UMT5Encoder are unrelated." }, { "left": "issue:43097", "right": "issue:43873", "accept": false, - "reason": "Removed tie_embeddings behavior is unrelated to quantization/offloading not working as expected." + "reason": "Embedding-tying deprecation warning and quantization offloading behavior are different problems." } ] }, @@ -12855,8 +12756,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -12927,8 +12828,8 @@ "issue:42907|issue:45753", "issue:42907|issue:44655", "issue:43531|issue:45341", - "issue:44483|issue:44514", "issue:43208|issue:43526", + "issue:44483|issue:44514", "issue:43012|issue:44683", "issue:42757|issue:43824", "issue:43618|issue:43761", @@ -12946,250 +12847,253 @@ "issue:42994|issue:44655", "issue:43066|issue:45381" ], - "prepared_review_unit_hash": "1c000879755fb7155722934f96720f7c08f6c268a489ddab48768d44f069b6bc", + "prepared_review_unit_hash": "702b219bd17b29be1bf6a7c13d7b0c1a056e0507f0ce983b37cffe09e6e2460c", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, "soft_pair_count": 35, - "serialized_chars": 37888, - "estimated_input_tokens": 9472, - "estimated_eval_tokens": 19200 + "serialized_chars": 37893, + "estimated_input_tokens": 9474, + "estimated_eval_tokens": 19204 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "Mostly near-misses across different Transformers regressions; the only somewhat central item is the broad testing_utils report, but the cluster is too heterogeneous for duplicate merging. No PRs are present.", - "confidence": 0.78, - "canonical_issue_reason": "issue:45341 is the broadest and most central-looking issue in this set, but it still appears to be just a loose representative rather than a true duplicate anchor.", + "summary": "This cluster is mostly a loose collection of unrelated Transformers bug reports. The strongest duplicate candidate is the pair of CLIP v5 regression issues, but the titles alone do not prove they are the same underlying bug, so it should remain unmatched unless the bodies show the same regression path.", + "confidence": 0.58, + "canonical_issue_reason": "issue:43761 is the cleaner representative of the CLIP regression because it states the concrete `CLIPVisionModel.forward` symptom and the v5 hidden-state failure more explicitly; issue:43618 reads like a narrower symptom report.", "canonical_pr_reason": null, - "best_issue_reason": "issue:45341 is the best general anchor only because it is the most generic/open report and appears in the most soft-similarity candidates.", + "best_issue_reason": "issue:43761 is the best issue to keep as the representative bug report because it is more specific and directly describes the observed regression path, while issue:43618 is less explicit.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43066", "right": "issue:43906", "accept": false, - "reason": "Different failures: tokenizer decoder type mismatch vs an isolated reproduction of a separate import/path issue." + "reason": "Both are model-loading/tokenization-related, but they point to different failures and different code paths." }, { "left": "issue:43986", "right": "issue:45372", "accept": false, - "reason": "Both are import/load-time crashes, but they involve different missing symbols and different model/processor paths." + "reason": "Different root causes: missing `torchvision` versus a `ReasoningEffort` import failure." }, { "left": "issue:43825", "right": "issue:44655", "accept": false, - "reason": "Error-message regression in pipeline task support is unrelated to saving Pipeline objects." + "reason": "`pipeline()` error messaging and `save_pretrained()` for Pipeline objects are different functionalities." }, { "left": "issue:43986", "right": "issue:45290", "accept": false, - "reason": "AutoProcessor/torchvision dependency crash is unrelated to chat-template handling with tool-call messages." + "reason": "Processor loading without torchvision is unrelated to `apply_chat_template` crashing on tool-call messages." }, { "left": "issue:45341", "right": "issue:45561", "accept": false, - "reason": "Same testing-utils area, but one is a vague bug note and the other is a specific xdist file-race bug." + "reason": "A vague testing-utils bug and an xdist race on `captured_info.txt` are not clearly the same bug." }, { "left": "issue:43986", "right": "issue:45341", "accept": false, - "reason": "One is a processor-loading dependency crash; the other is a generic testing-utils bug." + "reason": "Only superficial overlap in testing/utils wording; the symptoms and code paths differ." }, { "left": "issue:43097", "right": "issue:43901", "accept": false, - "reason": "Removed API warning and stale docs entry are different kinds of changes with different targets." + "reason": "A v5 API removal complaint and a docs mismatch for `return_all_scores` are not the same issue." }, { "left": "issue:42617", "right": "issue:45341", "accept": false, - "reason": "3d_parallel execution failure is unrelated to testing_utils behavior." + "reason": "A 3D parallel script failure and a testing-utils race are unrelated." }, { "left": "issue:43992", "right": "issue:44484", "accept": false, - "reason": "Missing embed_tokens.weight during model load is unrelated to the save_pretrained shard-size question." + "reason": "Missing checkpoint weights on load and the `max_shard_size` default in save_pretrained are different bugs." }, { "left": "issue:45341", "right": "issue:45588", "accept": false, - "reason": "Testing-utils bug and flash_attention crash on s_aux=None are different code paths and symptoms." + "reason": "Testing-utils race and a flash-attention crash in sink-less models are different code paths." }, { "left": "issue:43012", "right": "issue:44368", "accept": false, - "reason": "Both mention warnings, but they come from different APIs and different configuration checks." + "reason": "A PyTorch precision warning during compilation is not the same as a `tie_word_embeddings` config warning." }, { "left": "issue:42886", "right": "issue:43381", "accept": false, - "reason": "Offline cache loading and eval-mode gradient checkpointing are unrelated runtime behaviors." + "reason": "Offline cache loading and gradient-checkpointing eval-mode behavior are unrelated." }, { "left": "issue:43012", "right": "issue:43550", "accept": false, - "reason": "Precision warning during compile is not the same bug as Bamba compile failing with SDPA." + "reason": "One is a warning, the other is a `torch.compile` SDPA failure; different bugs." }, { "left": "issue:43208", "right": "issue:43550", "accept": false, - "reason": "xLSTM training blockers and Bamba torch.compile+SDPA failure are different model-specific issues." + "reason": "xLSTM training bugs and Bamba SDPA compile failures do not match." }, { "left": "issue:42907", "right": "issue:45753", "accept": false, - "reason": "Saving dequantized models is unrelated to an infinite-generation loop on a specific image." + "reason": "Saving dequantized models and an infinite-loop image bug are different problems." }, { "left": "issue:42907", "right": "issue:44655", "accept": false, - "reason": "Quantized/dequantized save failure is not the same as saving Pipeline objects." + "reason": "Quantized/dequantized model saving is unrelated to Pipeline serialization." }, { "left": "issue:43531", "right": "issue:45341", "accept": false, - "reason": "Qwen3-MoE sliding-window behavior is unrelated to testing_utils." + "reason": "Qwen3-MoE sliding-window behavior and testing-utils bugs do not share an underlying failure." }, { - "left": "issue:44483", - "right": "issue:44514", + "left": "issue:43208", + "right": "issue:43526", "accept": false, - "reason": "Chat/completions API rejection and processor chat-template batching are distinct failures." + "reason": "xLSTM training issues and BEiT label-reduction behavior are unrelated." }, { - "left": "issue:43208", - "right": "issue:43526", + "left": "issue:44483", + "right": "issue:44514", "accept": false, - "reason": "xLSTM training bugs do not match a BeitImageProcessorFast label-reduction bug." + "reason": "Chat-completions API validation and batched `apply_chat_template` crashes are different failures." }, { "left": "issue:43012", "right": "issue:44683", "accept": false, - "reason": "A warning about bfloat16 precision is not the same as compiled flex_attention failing on torch>=2.9." + "reason": "A bfloat16 warning is unrelated to compiled flex-attention failures on newer torch versions." }, { "left": "issue:42757", "right": "issue:43824", "accept": false, - "reason": "Both are ImportErrors, but they concern different missing symbols in different modules." + "reason": "Both are import errors, but they involve different missing symbols and different packages." }, { "left": "issue:43618", "right": "issue:43761", "accept": false, - "reason": "Related CLIP regressions, but one is missing attentions and the other is missing hidden_states." + "reason": "They are likely related CLIP v5 regressions, but the titles alone do not establish they are definitely the same bug." }, { "left": "issue:44483", "right": "issue:45290", "accept": false, - "reason": "Different layers of the stack: an endpoint validation failure vs a tokenizer/template crash." + "reason": "Chat-completions request validation and `apply_chat_template` tool-call crashes are different issues." }, { "left": "issue:43064", "right": "issue:44380", "accept": false, - "reason": "Optimizer-state corruption under FSDP2+PEFT is unrelated to GPT2 attention scaling being ignored." + "reason": "Trainer/FSDP optimizer-state corruption is unrelated to GPT2 attention scaling being ignored." }, { "left": "issue:44038", "right": "issue:44230", "accept": false, - "reason": "Qwen3-VL/MoE loading regression and fp8 inference support are separate feature/regression reports." + "reason": "A general Qwen3-VL-MoE breakage and an fp8 support request are not the same bug." }, { "left": "issue:43299", "right": "issue:43867", "accept": false, - "reason": "Qwen3VL model loading breakage is unrelated to a sorted-state_dict load error." + "reason": "Qwen3VL-MoE loading breakage and sorted-state_dict load errors are different issues." }, { "left": "issue:38617", "right": "issue:43502", "accept": false, - "reason": "Importing a removed utility is unrelated to API requests happening despite local_files_only=True." + "reason": "An import failure and accidental API requests despite `local_files_only=True` are unrelated." }, { "left": "issue:44483", "right": "issue:45381", "accept": false, - "reason": "Chat/completions request rejection is unrelated to wrong video vision_position_ids." + "reason": "Different Qwen2.5-VL failures with different symptoms and code paths." }, { "left": "issue:43901", "right": "issue:44655", "accept": false, - "reason": "Docs mentioning return_all_scores are unrelated to Pipeline save_pretrained behavior." + "reason": "Pipeline docs mismatch and Pipeline save_pretrained failure are different problems." }, { "left": "issue:36683", "right": "issue:43475", "accept": false, - "reason": "Gemma3Config missing vocab_size and SAM3 VisionEncoderOutput missing fpn_position_embeddings are different model attribute bugs." + "reason": "Gemma3 config missing `vocab_size` and SAM3 output missing `fpn_position_embeddings` are unrelated attribute errors." }, { "left": "issue:44220", "right": "issue:44655", "accept": false, - "reason": "Fbank feature extraction bug is unrelated to saving Pipeline objects." + "reason": "Feature-extraction helper behavior and Pipeline serialization do not share an underlying bug." }, { "left": "issue:43867", "right": "issue:45216", "accept": false, - "reason": "Sorted state_dict load failure and Qwen3.5 save_pretrained checkpoint regression are different directions of the serialization path." + "reason": "A load error tied to state_dict ordering is not the same as a Qwen3.5 save_pretrained checkpoint regression." }, { "left": "issue:43352", "right": "issue:43937", "accept": false, - "reason": "Flash-Attention-2 unsupported error is unrelated to invalid GenerationConfig for GLM-5." + "reason": "Nemotron Flash Attention support and GLM-5 generation-config validation are unrelated." }, { "left": "issue:42994", "right": "issue:44655", "accept": false, - "reason": "Quantized model save failure and Pipeline save_pretrained failure are different object types and code paths." + "reason": "Quantized model saving and Pipeline saving are different serialization paths and bugs." }, { "left": "issue:43066", "right": "issue:45381", "accept": false, - "reason": "Tokenizer decoder type regression is unrelated to Qwen2.5-VL video position-id issues." + "reason": "Wrong tokenizer decoder type and Qwen2.5-VL video position-id issues are unrelated." } ] }, - "evaluator_result": null, - "evaluator_used": false, - "retried": false, + "evaluator_result": { + "accept": false, + "feedback": "Most of the rejects are reasonably conservative, but the one accepted edge is a bit too strong given only the packet titles: issue:43618 (\u2018CLIPOutput attentions is no longer assigned\u2019) and issue:43761 (\u2018CLIPVisionModel.forward returns hidden_states=None even when output_hidden_states=True\u2019) may be related, but the output does not establish they are definitely the same bug. I\u2019d want that verdict softened to \u2018likely related\u2019 or left unmatched unless the bodies show the same regression path. The cluster summary also slightly overcommits by calling this the \u2018only credible duplicate edge\u2019 instead of \u2018the strongest possible duplicate candidate.\u2019" + }, + "evaluator_used": true, + "retried": true, "accepted_nontrivial_soft_edge": false, "error_kind": null, "error_message": null @@ -13202,8 +13106,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -13267,9 +13171,9 @@ "issue:36683|issue:43881", "issue:43299|issue:45753", "issue:43352|issue:43976", - "issue:43992|issue:45310", "issue:41669|issue:44246", "issue:43441|issue:44380", + "issue:43992|issue:45310", "issue:43352|issue:44380", "issue:44380|issue:45479", "issue:43867|issue:45375", @@ -13288,19 +13192,19 @@ "issue:36296|issue:43575", "issue:44077|issue:44380", "issue:43618|issue:44857", - "issue:43618|issue:45588", "issue:43867|issue:44484", + "issue:43618|issue:45588", "issue:28282|issue:33453", "issue:43618|issue:44977" ], - "prepared_review_unit_hash": "ef64ac19bc4746fa4d3a1a7ad237943eeffd0f92a02030f911bdfbc1acc85b44", + "prepared_review_unit_hash": "6e9ce916a630d98b7cc23a92a8105b929410b3cc234f90d5f37625c08c679928", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, @@ -13315,232 +13219,229 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "Mostly a grab-bag of unrelated Transformers bug reports; the only strong duplicate-like link is an import-time slowdown report and its concrete root-cause issue.", - "confidence": 0.66, - "canonical_issue_reason": "issue:41669 is the most concrete statement of the import-performance regression, naming the likely cause (import * in models) and the observed slowdown; it can absorb the vaguer symptom report in issue:44246.", + "summary": "This cluster is largely noisy: several issues share broad keywords like configs, loading, flash-attention, or import/runtime failures, but they do not look like the same underlying bug or change. No pull requests are present.", + "confidence": 0.59, + "canonical_issue_reason": "issue:36683 is the best representative of the small config-attribute/load-path subtheme and is the most connected issue in the set, but the overall cluster is still heterogeneous.", "canonical_pr_reason": null, - "best_issue_reason": "issue:41669 is the best representative because it is specific, actionable, and likely explains the broader import-slowdown symptom better than the generic performance complaint.", + "best_issue_reason": "issue:36683 is the strongest single issue to keep as the cluster representative because it is active and central to the few genuinely related config-attribute reports.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:36010", "right": "issue:42757", "accept": false, - "reason": "Both are import errors, but they involve different missing symbols from different packages, so they are not the same underlying bug." + "reason": "Both are import errors, but they involve different missing symbols from different packages and different failure modes." }, { "left": "issue:44263", "right": "issue:45440", "accept": false, - "reason": "Different model paths and failure modes; one is a DSA indexer tensor-splitting issue, the other is a DeepseekV3MoE implementation divergence." + "reason": "These are unrelated model-path bugs: one is a torch.split/indexer issue, the other is a DeepSeekV3 implementation divergence." }, { "left": "issue:43756", "right": "issue:44485", "accept": false, - "reason": "Both mention RoPE, but they concern different models and different implementation concerns, not one shared bug." + "reason": "Both mention RoPE, but they concern different models and distinct implementation problems." }, { "left": "issue:36683", "right": "issue:43525", "accept": false, - "reason": "Both are missing-config-attribute reports, but for different configs and different fields, so they are not the same defect." + "reason": "Both are missing-config-attribute reports, but on different models and different fields, so they are not the same bug." }, { "left": "issue:43618", "right": "issue:45216", "accept": false, - "reason": "Unrelated issues: CLIP attentions not being assigned versus a Qwen3.5 checkpoint save regression." + "reason": "One is about CLIPOutput attentions no longer being assigned; the other is a Qwen3.5 save_pretrained regression." }, { "left": "issue:43906", "right": "issue:45753", "accept": false, - "reason": "The reproduction issue is too generic to equate with the Qwen3_5 image infinite-loop report; no clear same bug is shown." + "reason": "The first is an isolated reproduction of another issue; the second is a specific-image infinite loop in Qwen3_5, which is a different symptom." }, { "left": "issue:36683", "right": "issue:44242", "accept": false, - "reason": "Different MoE-related bugs: a missing config field versus a load-balancing loss condition." + "reason": "A missing Gemma3Config attribute is unrelated to load-balancing loss handling when output_router_logits=False." }, { "left": "issue:36683", "right": "issue:43881", "accept": false, - "reason": "Different models and failures; a missing Gemma3Config field is not the same as a GLM-4V loading problem." + "reason": "Different models and different loading failures; only the generic \"loading failed\" theme overlaps." }, { "left": "issue:43299", "right": "issue:45753", "accept": false, - "reason": "Both involve Qwen3.5-family models, but one is a MoE loading regression and the other is an image-generation infinite-loop bug." + "reason": "Both involve Qwen variants, but one is a v5.dev0 loading break and the other is an inference infinite-loop bug." }, { "left": "issue:43352", "right": "issue:43976", "accept": false, - "reason": "These are unrelated: a FlashAttention support error for one model versus a Python-version compatibility issue." - }, - { - "left": "issue:43992", - "right": "issue:45310", - "accept": false, - "reason": "Different problems: missing UMT5 encoder weights versus a Qwen3.5 MoE from_pretrained failure." + "reason": "A FlashAttention support limitation is unrelated to the Python-version compatibility failure." }, { "left": "issue:41669", "right": "issue:44246", - "accept": true, - "reason": "Both describe the same import-time slowdown in Transformers; 41669 identifies the likely cause while 44246 reports the symptom." + "accept": false, + "reason": "Import-time performance and occasional import slowness are too broad here; these do not describe the same code path or bug." }, { "left": "issue:43441", "right": "issue:44380", "accept": false, - "reason": "Both touch attention backends, but one is a model-specific FlashAttention crash and the other is an attention-scaling config mismatch." + "reason": "Both touch attention backends, but they are different model-specific bugs with different root causes." + }, + { + "left": "issue:43992", + "right": "issue:45310", + "accept": false, + "reason": "Missing embed_tokens.weight during UMT5Encoder loading is unrelated to the Qwen3.5 MoE from_pretrained regression." }, { "left": "issue:43352", "right": "issue:44380", "accept": false, - "reason": "A model-specific FlashAttention support limitation is not the same as GPT-2 attention scaling being ignored under backend changes." + "reason": "FlashAttention support in Nemotron and GPT2 attention scaling being ignored are separate backend/model issues." }, { "left": "issue:44380", "right": "issue:45479", "accept": false, - "reason": "Different bugs: backend-specific attention scaling versus a zero-loss classification edge case." + "reason": "Attention scaling behavior and a degenerate zero-loss classification bug are unrelated." }, { "left": "issue:43867", "right": "issue:45375", "accept": false, - "reason": "A generic state-dict loading failure is not the same as a missing Qwen3_5MoeVisionConfig field being dropped by strict handling." + "reason": "A sorted-state_dict loading error is not the same as a missing deepstack_visual_indexes config field." }, { "left": "issue:43064", "right": "issue:44492", "accept": false, - "reason": "Completely unrelated: distributed optimizer-state corruption versus a cache-strategy typo." + "reason": "Optimizer-state corruption during FSDP2/PEFT training is unrelated to a cache-strategy typo." }, { "left": "issue:43618", "right": "issue:44380", "accept": false, - "reason": "Different model/output regressions; CLIPOutput attentions missing is unrelated to GPT-2 attention scaling under SDPA/FlashAttention." + "reason": "These affect different models and different attention-related behaviors." }, { "left": "issue:43352", "right": "issue:45478", "accept": false, - "reason": "Same broad flash-attention area, but different models and different concrete failures; not one mergeable fix." + "reason": "Both are Qwen/Nemotron-era model bugs, but the concrete failure modes are unrelated." }, { "left": "issue:44355", "right": "issue:44855", "accept": false, - "reason": "Both are import/runtime failures, but one is about compiled Python files and the other is a Python 3.13 TorchScript parsing issue." + "reason": "Compiled Python file errors and a Python 3.13 torch.jit.script parsing issue are distinct import/runtime problems." }, { "left": "issue:36683", "right": "issue:44117", "accept": false, - "reason": "Different loading bugs: a missing config attribute versus TOKENIZER_MAPPING_NAMES returning None." + "reason": "A missing model config attribute is not the same as TOKENIZER_MAPPING_NAMES returning None." }, { "left": "issue:42890", "right": "issue:45561", "accept": false, - "reason": "Both are test-related, but one is flaky seed usage in integration tests and the other is an xdist race on a captured-info file." + "reason": "Both are test flakiness, but one is missing set_seed and the other is an xdist race on a captured file." }, { "left": "issue:43097", "right": "issue:43352", "accept": false, - "reason": "Unrelated: an API removal in 5.0.0 versus a FlashAttention support error for Nemotron." + "reason": "A removed config option and a FlashAttention support issue are unrelated." }, { "left": "issue:43867", "right": "issue:45357", "accept": false, - "reason": "A generic sorted-state-dict load error is not the same as incorrect Qwen3.5 visual encoder keys being saved." + "reason": "These both concern saving/loading, but one is a state_dict ordering problem and the other is incorrect visual encoder keys." }, { "left": "issue:43976", "right": "issue:44938", "accept": false, - "reason": "Both are Python-version compatibility reports, but they target different version ranges and likely different root causes." + "reason": "Both are Python-version compatibility reports, but they target different versions and likely different breakages." }, { "left": "issue:43992", "right": "issue:45478", "accept": false, - "reason": "Different loading failures affecting different models and fields; no shared underlying bug is evident." + "reason": "A missing embedding weight during load is unrelated to a Qwen3.5 MoE from_pretrained error." }, { "left": "issue:44360", "right": "issue:45440", "accept": false, - "reason": "These are different MoE/DSA issues; one is a missing ReLU in the indexer, the other is a DeepseekV3MoE implementation mismatch." + "reason": "A missing ReLU in the DSA indexer is unrelated to DeepSeekV3 implementation divergence." }, { "left": "issue:43064", "right": "issue:44655", "accept": false, - "reason": "Completely different problems: distributed training optimizer state corruption versus saving Pipeline objects." + "reason": "Training optimizer-state bugs and saving Pipeline objects are different code paths." }, { "left": "issue:36296", "right": "issue:43575", "accept": false, - "reason": "Both mention tensor parallelism, but one is a training bug and the other is an OOM during model loading." + "reason": "Tensor-parallel training bugs and TP-related OOM during model load are not the same defect." }, { "left": "issue:44077", "right": "issue:44380", "accept": false, - "reason": "Unrelated model/config validation issue versus an attention-backend scaling bug." + "reason": "An optional post_init restriction in patchtsmixer has nothing to do with GPT2 attention scaling." }, { "left": "issue:43618", "right": "issue:44857", "accept": false, - "reason": "Different failures in different subsystems: CLIP attentions assignment versus AMP/CUDA crash in LwDetrImageLoss." + "reason": "Missing CLIPOutput attentions and a float16 AMP CUDA crash are unrelated bugs." }, { - "left": "issue:43618", - "right": "issue:45588", + "left": "issue:43867", + "right": "issue:44484", "accept": false, - "reason": "Both mention attention-related code, but one is missing CLIPOutput attentions while the other is a flash_attention.py crash on sink-less models." + "reason": "State_dict sorting failures and the max_shard_size default question are not the same issue." }, { - "left": "issue:43867", - "right": "issue:44484", + "left": "issue:43618", + "right": "issue:45588", "accept": false, - "reason": "A load-model state_dict ordering issue is unrelated to save_pretrained shard-size behavior." + "reason": "The CLIPOutput attentions regression and the s_aux=None flash_attention crash affect different code paths." }, { "left": "issue:28282", "right": "issue:33453", "accept": false, - "reason": "Different import/load failures: missing PyTorch for AutoModel versus a tokenizer-loading regression." + "reason": "AutoModel missing PyTorch and tokenizer-loading regression are completely different failures." }, { "left": "issue:43618", "right": "issue:44977", "accept": false, - "reason": "Different model-specific regressions; CLIPOutput attentions missing is not the same as Qwen3.5 flash-attention generation failure." + "reason": "These are unrelated: one concerns a missing attention field, the other is a Qwen3.5 flash-attention generation bug." } ] }, - "evaluator_result": { - "accept": true, - "feedback": "The analysis is grounded in the packet and stays mostly conservative. The only accepted soft edge, issue:41669 vs issue:44246, is supported by the titles and is phrased cautiously. The remaining negatives generally rely on clear title-level mismatches in model, subsystem, or failure mode, without overstating duplicate evidence." - }, - "evaluator_used": true, + "evaluator_result": null, + "evaluator_used": false, "retried": false, - "accepted_nontrivial_soft_edge": true, + "accepted_nontrivial_soft_edge": false, "error_kind": null, "error_message": null }, @@ -13552,8 +13453,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -13615,24 +13516,24 @@ "issue:43502|issue:44062", "issue:42890|issue:45341", "issue:43010|issue:44075", - "issue:44222|issue:45125", "issue:42548|issue:44351", + "issue:44222|issue:45125", "issue:43502|issue:44485", "issue:44355|issue:44749", - "issue:30990|issue:31795", "issue:45125|issue:45478", + "issue:30990|issue:31795", "issue:45125|issue:45375", "issue:43010|issue:45278", "issue:44387|issue:44556", - "issue:43723|issue:44556", "issue:30333|issue:33453", + "issue:43723|issue:44556", "issue:43010|issue:44368", "issue:43502|issue:44222", "issue:44315|issue:44360", "issue:43502|issue:44961", "issue:44355|issue:44655", - "issue:30333|issue:36296", "issue:42831|issue:43874", + "issue:30333|issue:36296", "issue:44355|issue:45071", "issue:44222|issue:44230", "issue:43827|issue:44961", @@ -13645,284 +13546,274 @@ "issue:43576|issue:43824", "issue:43010|issue:43976", "issue:44961|issue:45468", - "issue:38617|issue:43576", "issue:31795|issue:38617" ], - "prepared_review_unit_hash": "adcb3e7dfeb46384e5096478d8d76f74e478bd7d64bfd2aa296b1b96ee5feb78", + "prepared_review_unit_hash": "0dfd9c2b3bc8ef5417dc950ce3be516006b9d6124d60a8d03096a8fd7b799427", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, - "soft_pair_count": 40, - "serialized_chars": 39162, - "estimated_input_tokens": 9791, - "estimated_eval_tokens": 19838 + "soft_pair_count": 39, + "serialized_chars": 38861, + "estimated_input_tokens": 9716, + "estimated_eval_tokens": 19688 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is mostly a grab bag of Transformers v5 regressions (imports, model/tokenizer loading, FP8, CLI, docs, tests, and training/runtime bugs). The soft links are generally too broad or only vaguely similar; none look like the same concrete underlying bug/change.", - "confidence": 0.9, - "canonical_issue_reason": "issue:45278 is the best representative umbrella issue: it summarizes broad import breakage after upgrading to 5.5.0, which matches the cluster\u2019s dominant v5-regression theme.", + "summary": "Mostly a mixed bag of unrelated bug reports and documentation regressions around Transformers v5; only one soft edge looks like a plausible duplicate pair. I\u2019d use the broad v5 import-regression report as the representative issue, but there is no PR in this cluster.", + "confidence": 0.42, + "canonical_issue_reason": "issue:45278 is the broadest umbrella report here ('Many import errors after update from 4.57.0 to 5.5.0') and best represents the general v5 import/load regression theme, even though the cluster is not a set of true duplicates.", "canonical_pr_reason": null, - "best_issue_reason": "issue:45278 is the strongest global representative because it is broad, version-specific, and captures the common import/load regression pattern better than the narrower or more isolated reports.", + "best_issue_reason": "issue:45278 is the most general and representative issue among the mostly unrelated reports; it captures the shared version-upgrade breakage theme better than the narrower model-specific or doc-only items.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43901", "right": "issue:44938", "accept": false, - "reason": "Docs for TextClassificationPipeline return_all_scores and a Python 3.14 load failure are unrelated symptoms and code paths." + "reason": "Docs about a removed pipeline argument and a Python 3.14 import/load failure are different problems." }, { "left": "issue:42890", "right": "issue:44964", "accept": false, - "reason": "A flaky integration test missing set_seed is unrelated to a Phi-4 multimodal model loading failure." + "reason": "Seed-setting flakiness in a SAM HQ test is unrelated to Phi-4 multimodal model loading." }, { "left": "issue:44655", "right": "issue:44938", "accept": false, - "reason": "Pipeline save_pretrained support and Python 3.14 import/load failure do not describe the same bug." + "reason": "Saving Pipeline objects and Python 3.14 import failure are not the same code-path bug." }, { "left": "issue:42890", "right": "issue:44112", "accept": false, - "reason": "Missing set_seed in tests is unrelated to a stale device-override CI failure." + "reason": "Test nondeterminism from missing set_seed() is separate from a GraniteSpeech CI device-override failure." }, { "left": "issue:42890", "right": "issue:45600", "accept": false, - "reason": "Test seed flakiness and auto_mappings.py CI breakage are different issues in different paths." + "reason": "A flaky integration test is unrelated to stale auto_mappings references causing CI breakage." }, { "left": "issue:43502", "right": "issue:44062", "accept": false, - "reason": "local_files_only network leakage and AddedToken keyword handling are distinct bugs." + "reason": "local_files_only leaking network requests and AddedToken keyword duplication are distinct loader/tokenizer issues." }, { "left": "issue:42890", "right": "issue:45341", "accept": false, - "reason": "A test flake and a bug in testing_utils.py are not the same concrete failure." + "reason": "Both are test-related, but they describe different failures and different fixes." }, { "left": "issue:43010", "right": "issue:44075", "accept": false, - "reason": "@torch.no_grad decoration for cache updates is unrelated to SGD optimizer arguments not being used." - }, - { - "left": "issue:44222", - "right": "issue:45125", - "accept": false, - "reason": "Both mention MoE/FP8/tensor-parallel areas, but one is FP8 save_pretrained and the other is missing _tp_plan for tensor parallelism; not the same fix." + "reason": "Cache update/no_grad behavior is unrelated to SGD optimizer arguments being ignored." }, { "left": "issue:42548", "right": "issue:44351", "accept": false, - "reason": "Missing PreTrainedModel and missing HybridCache are separate import-export regressions." + "reason": "Both are import errors, but for different missing symbols and likely different fixes." }, { - "left": "issue:43502", - "right": "issue:44485", + "left": "issue:44222", + "right": "issue:45125", "accept": false, - "reason": "Unexpected network requests and a GLM-5 RoPE discussion are unrelated." + "reason": "FP8 save_pretrained on MoE models is not the same as missing _tp_plan for tensor parallelism." }, { - "left": "issue:44355", - "right": "issue:44749", + "left": "issue:43502", + "right": "issue:44485", "accept": false, - "reason": "Compiled-Python-file errors and slow filtering after upgrade are different issues." + "reason": "local_files_only network leakage and GLM-5 RoPE discussion are unrelated." }, - { - "left": "issue:30990", - "right": "issue:31795", + { + "left": "issue:44355", + "right": "issue:44749", "accept": false, - "reason": "Sentence Transformers loading stuck and documentation confusion about model.forward arguments are unrelated." + "reason": "Compiled Python file errors are unrelated to a slowdown after upgrading Transformers." }, { "left": "issue:45125", "right": "issue:45478", + "accept": true, + "reason": "Both point to the same Qwen3.5 MoE loading failure path in v5, plausibly caused by missing tensor-parallel support." + }, + { + "left": "issue:30990", + "right": "issue:31795", "accept": false, - "reason": "Both involve Qwen3.5 MoE, but one is a tensor-parallel plan omission and the other is a broader from_pretrained error; too little evidence they share one concrete fix." + "reason": "SentenceTransformer loading issues and model.forward documentation confusion are unrelated." }, { "left": "issue:45125", "right": "issue:45375", "accept": false, - "reason": "Missing _tp_plan for tensor parallelism and a missing config field silently dropped by @strict are different model config bugs." + "reason": "Missing _tp_plan and a missing deepstack_visual_indexes config field are different model-configuration bugs." }, { "left": "issue:43010", "right": "issue:45278", "accept": false, - "reason": "Cache/layer update decoration and broad import errors after upgrade are not the same code-path problem." + "reason": "Cache no_grad/update behavior is unrelated to broad import errors after upgrading to v5.5.0." }, { "left": "issue:44387", "right": "issue:44556", "accept": false, - "reason": "Int4 CUDA reserved-memory OOM and checkpoint reload breakage across versions are different failure modes." + "reason": "Int4 quantization memory regression and checkpoint reload failure after upgrade are different bugs." }, { - "left": "issue:43723", - "right": "issue:44556", + "left": "issue:30333", + "right": "issue:33453", "accept": false, - "reason": "Tokenizer loading regression and checkpoint reload incompatibility are both load-related but not the same underlying issue." + "reason": "MLflow reporting on failed training jobs and tokenizer loading regression are unrelated." }, { - "left": "issue:30333", - "right": "issue:33453", + "left": "issue:43723", + "right": "issue:44556", "accept": false, - "reason": "MLFlow reporting after training failure and tokenizer loading regression are unrelated." + "reason": "Tokenizer loading in v5 and reloading old checkpoints after upgrade are related broadly, but not the same concrete bug." }, { "left": "issue:43010", "right": "issue:44368", "accept": false, - "reason": "@torch.no_grad on cache updates and tie_word_embeddings warning noise do not share a concrete bug." + "reason": "Cache update/no_grad behavior is unrelated to a tie_word_embeddings warning during Qwen LoRA fine-tuning." }, { "left": "issue:43502", "right": "issue:44222", "accept": false, - "reason": "Network requests despite local_files_only and FP8 save_pretrained issues are unrelated." + "reason": "local_files_only network calls and FP8 MoE save_pretrained are different issues." }, { "left": "issue:44315", "right": "issue:44360", "accept": false, - "reason": "Liger Kernel not applied in model_init and a missing ReLU in the DSA indexer are unrelated subsystems." + "reason": "Liger Kernel not applying with model_init is unrelated to the DSA indexer ReLU discussion." }, { "left": "issue:43502", "right": "issue:44961", "accept": false, - "reason": "API requests despite local_files_only and an unrelated 'racoon' issue have no overlap." + "reason": "local_files_only network leakage and a one-word issue title do not describe the same bug." }, { "left": "issue:44355", "right": "issue:44655", "accept": false, - "reason": "Errors with compiled Python files and Pipeline save_pretrained are different problems." + "reason": "Compiled Python file errors and inability to save Pipeline objects are different failure modes." }, { - "left": "issue:30333", - "right": "issue:36296", + "left": "issue:42831", + "right": "issue:43874", "accept": false, - "reason": "MLFlow reporting on failed training jobs and tensor-parallel training bugs are unrelated." + "reason": "FP8 accuracy regression and missing image-processor method are unrelated." }, { - "left": "issue:42831", - "right": "issue:43874", + "left": "issue:30333", + "right": "issue:36296", "accept": false, - "reason": "FineGrainedFP8 accuracy issues and a missing get_number_of_image_patches method are separate model bugs." + "reason": "MLflow reporting failure and tensor-parallel training bug are different training-time issues." }, { "left": "issue:44355", "right": "issue:45071", "accept": false, - "reason": "Compiled-file errors and PretrainedConfig type-checking breakage are distinct." + "reason": "Compiled file loading issues and PretrainedConfig type-checking breakage are unrelated." }, { "left": "issue:44222", "right": "issue:44230", "accept": false, - "reason": "Both mention FP8/MoE, but one is save_pretrained and the other is inference support; not the same change." + "reason": "Both involve FP8 MoE, but one is save_pretrained and the other is inference support; not the same concrete bug." }, { "left": "issue:43827", "right": "issue:44961", "accept": false, - "reason": "Docs still referencing pipeline() and an unrelated 'racoon' issue are not duplicates." + "reason": "Pipeline docs references and a one-word title are unrelated." }, { "left": "issue:43010", "right": "issue:45656", "accept": false, - "reason": "Cache update no_grad decoration and optimizer.step being called twice with deepspeed are different runtime bugs." + "reason": "Cache update decoration and optimizer double-step behavior under deepspeed are different code paths." }, { "left": "issue:33453", "right": "issue:34689", "accept": false, - "reason": "Tokenizer loading regression and Llama 3.2 model loading breakage are both loading-related but not the same concrete issue." + "reason": "Tokenizer loading regression and Llama 3.2 Vision model loading failure are separate model-loading bugs." }, { "left": "issue:43299", "right": "issue:44556", "accept": false, - "reason": "Qwen3VL MoE loading failure and checkpoint reload incompatibility across versions are not clearly the same root cause." + "reason": "Qwen3VL MoE loading failure and old-checkpoint reload failure are both upgrade-related, but not the same underlying bug." }, { "left": "issue:42757", "right": "issue:43576", "accept": false, - "reason": "A missing huggingface_hub symbol import and a broken transformers env command are unrelated." - }, - { - "left": "issue:33453", - "right": "issue:36296", - "accept": false, - "reason": "Tokenizer loading regression and tensor-parallel training bug are different failures." + "reason": "Missing huggingface_hub export and a broken transformers env command are unrelated." }, { "left": "issue:43576", "right": "issue:45071", "accept": false, - "reason": "Broken env CLI and PretrainedConfig type-checking are unrelated." + "reason": "env command breakage and PretrainedConfig type checking are unrelated." }, { "left": "issue:43576", "right": "issue:43824", "accept": false, - "reason": "Broken env CLI and Qwen2_5_VL import failure are not the same bug." + "reason": "Broken env command and missing Qwen2_5_VLForConditionalGeneration import are different regressions." }, { "left": "issue:43010", "right": "issue:43976", "accept": false, - "reason": "Cache update no_grad decoration and Python-version compatibility issues are unrelated." + "reason": "Cache update behavior and Python 3.9 compatibility failure are unrelated." }, { "left": "issue:44961", "right": "issue:45468", "accept": false, - "reason": "'racoon' and Gemma-4 audio positional encoding are unrelated." - }, - { - "left": "issue:38617", - "right": "issue:43576", - "accept": false, - "reason": "layer_type_validation import failure and env CLI breakage are distinct regressions." + "reason": "Uninformative title and Gemma-4 audio positional encoding bug are unrelated." }, { "left": "issue:31795", "right": "issue:38617", "accept": false, - "reason": "Documentation confusion and an ImportError from configuration_utils are unrelated." + "reason": "Documentation confusion in model.forward and a missing import symbol are unrelated." } ] }, - "evaluator_result": null, - "evaluator_used": false, + "evaluator_result": { + "accept": true, + "feedback": "Grounded and conservative overall. The representative issue choice is supported by the packet, the summary correctly notes this is a mixed cluster rather than true duplicates, and the one accepted soft-edge pair looks plausibly mergeable as the same Qwen3.5 MoE loading regression." + }, + "evaluator_used": true, "retried": false, - "accepted_nontrivial_soft_edge": false, + "accepted_nontrivial_soft_edge": true, "error_kind": null, "error_message": null }, @@ -13934,17 +13825,17 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ "issue:31795", "issue:33453", "issue:36296", + "issue:38617", "issue:41628", "issue:41669", - "issue:42371", "issue:42548", "issue:42890", "issue:43010", @@ -13968,11 +13859,11 @@ "issue:44351", "issue:44355", "issue:44360", + "issue:44373", "issue:44393", "issue:44492", "issue:44556", "issue:44623", - "issue:44655", "issue:44704", "issue:44829", "issue:44861", @@ -13990,6 +13881,7 @@ ], "soft_pairs": [ "issue:41628|issue:42548", + "issue:38617|issue:43576", "issue:43576|issue:44351", "issue:44908|issue:44961", "issue:43010|issue:43232", @@ -13998,15 +13890,15 @@ "issue:36296|issue:43526", "issue:44261|issue:44360", "issue:44273|issue:45003", - "issue:44273|issue:45542", "issue:44623|issue:45230", + "issue:44273|issue:45542", "issue:44393|issue:45753", "issue:43576|issue:44556", "issue:41669|issue:43673", "issue:44393|issue:45071", "issue:44038|issue:45125", "issue:44556|issue:45092", - "issue:44961|issue:45092", + "issue:44373|issue:44961", "issue:44704|issue:44908", "issue:42890|issue:44908", "issue:44246|issue:44273", @@ -14026,285 +13918,279 @@ "issue:44077|issue:44861", "issue:43827|issue:44908", "issue:31795|issue:33453", - "issue:41669|issue:44908", - "issue:42371|issue:44492", - "issue:43576|issue:44273", - "issue:44655|issue:44861" + "issue:41669|issue:44908" ], - "prepared_review_unit_hash": "47e58838d3571a7cc2d7a74f78a1d000cae58939c1a40204402838eb8eb4b90e", + "prepared_review_unit_hash": "10a6187d61c96c91a4cd9b4ea586188c5acc338f5507daced519f4dd45865222", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { "node_count": 48, "item_count": 48, - "soft_pair_count": 41, - "serialized_chars": 39327, - "estimated_input_tokens": 9832, - "estimated_eval_tokens": 19920 + "soft_pair_count": 39, + "serialized_chars": 38746, + "estimated_input_tokens": 9687, + "estimated_eval_tokens": 19630 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "Loose heterogeneous cluster of Transformers issues spanning import/lazy-loading, generation/cache, docs, and model-specific bugs. No PRs are present, and none of the soft edges look like the same concrete fix.", - "confidence": 0.39, - "canonical_issue_reason": "issue:41669 is the broadest and most connected import/lazy-loading/performance report in the set, with inbound references and overlap with several import-related artifacts.", + "summary": "This cluster is mostly a set of separate Transformers issues spanning imports, lazy loading, generation/cache behavior, docs, and model-specific regressions. No soft pair is strongly supported as the same underlying bug; the lazy-loading pair in particular lacks enough evidence to merge confidently.", + "confidence": 0.62, + "canonical_issue_reason": "issue:44273 is the broadest open report in the cluster and the safest representative for the lazy-loading subgroup, but the evidence is not strong enough to treat it as the same bug as issue:45003.", "canonical_pr_reason": null, - "best_issue_reason": "issue:41669 is the best representative triage anchor because it sits near the center of the cluster\u2019s import-related subtheme and has the strongest evidence of cross-issue relevance.", + "best_issue_reason": "issue:44273 is the most central open issue and the best cluster representative by scope, even though the duplicate linkage to issue:45003 is unproven.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:41628", "right": "issue:42548", "accept": false, - "reason": "Both are import failures, but for different symbols and different root causes." + "reason": "Both are import failures, but they involve different missing symbols and different export paths; not the same concrete bug." + }, + { + "left": "issue:38617", + "right": "issue:43576", + "accept": false, + "reason": "Missing internal import vs broken env command; different features and failure modes." }, { "left": "issue:43576", "right": "issue:44351", "accept": false, - "reason": "One is the v5 env command, the other is a missing import; not the same bug." + "reason": "An env command regression and a missing top-level cache export are unrelated." }, { "left": "issue:44908", "right": "issue:44961", "accept": false, - "reason": "Scheduler kwargs and an unrelated placeholder issue." + "reason": "Scheduler kwargs bug and an unrelated placeholder issue; no shared code-path." }, { "left": "issue:43010", "right": "issue:43232", "accept": false, - "reason": "Different generation/cache code paths and different symptoms." + "reason": "Both touch generation/cache code, but they describe different defects and different call sites." }, { "left": "issue:36296", "right": "issue:43901", "accept": false, - "reason": "Tensor-parallel training bug versus a documentation mismatch." + "reason": "Tensor-parallel training bug vs documentation mismatch; not the same issue." }, { "left": "issue:41628", "right": "issue:44355", "accept": false, - "reason": "Missing AutoImageProcessor import is unrelated to compiled-file execution errors." + "reason": "Top-level import failure for a symbol vs compiled-file execution error; different problems." }, { "left": "issue:36296", "right": "issue:43526", "accept": false, - "reason": "Different model-specific bugs: tensor parallel training versus BeitImageProcessorFast label reduction." + "reason": "Tensor parallel training bug and a BeitImageProcessorFast label-return bug are unrelated." }, { "left": "issue:44261", "right": "issue:44360", "accept": false, - "reason": "Precision/eps configuration issue versus a DSA indexer activation complaint." + "reason": "Layer norm precision/config issue vs DSA indexer missing ReLU; different subsystems." }, { "left": "issue:44273", "right": "issue:45003", "accept": false, - "reason": "Lazy loading and sys.modules access are related themes, but not the same underlying defect." + "reason": "The titles are loosely related around lazy loading/import behavior, but there is not enough evidence here to say they are the same concrete bug." }, { - "left": "issue:44273", - "right": "issue:45542", + "left": "issue:44623", + "right": "issue:45230", "accept": false, - "reason": "Lazy loading is unrelated to the TensorFlow backend dependency error." + "reason": "Missing files on processor save vs a vague bug report; no evidence of the same defect." }, { - "left": "issue:44623", - "right": "issue:45230", + "left": "issue:44273", + "right": "issue:45542", "accept": false, - "reason": "Processor save_pretrained missing files is not the same as a generic bug report." + "reason": "Lazy-loading failure and TensorFlow backend detection error are different bugs." }, { "left": "issue:44393", "right": "issue:45753", "accept": false, - "reason": "Same model family, but different failure modes: bad bbox output versus infinite looping." + "reason": "Same model family, but different symptoms and likely different root causes; too broad to merge." }, { "left": "issue:43576", "right": "issue:44556", "accept": false, - "reason": "Both mention v5 regressions, but one is the env command and the other is checkpoint reload compatibility." + "reason": "Broken env command vs checkpoint reload incompatibility after v5 upgrade; different user flows." }, { "left": "issue:41669", "right": "issue:43673", "accept": false, - "reason": "Import-star cleanup/performance is unrelated to chunked_prefill cache handling." + "reason": "Import-star performance regression vs missing generation cache in chunked prefill; not the same code-path." }, { "left": "issue:44393", "right": "issue:45071", "accept": false, - "reason": "Model output hallucination versus a PretrainedConfig typing break." + "reason": "Qwen3-VL output bug and PretrainedConfig type-checking regression are unrelated." }, { "left": "issue:44038", "right": "issue:45125", "accept": false, - "reason": "Different Qwen3-family issues with different code paths and fixes." + "reason": "Both are Qwen3-related, but one is a general bug and the other is a tensor-parallel missing-plan issue." }, { "left": "issue:44556", "right": "issue:45092", "accept": false, - "reason": "Both are v5 compatibility complaints, but they concern different models and different breakages." + "reason": "Both concern v5 checkpoint incompatibilities, but they target different models and different failure causes." }, { - "left": "issue:44961", - "right": "issue:45092", + "left": "issue:44373", + "right": "issue:44961", "accept": false, - "reason": "Unrelated placeholder issue versus remote-code checkpoint incompatibility." + "reason": "Docstring issue vs an unrelated placeholder issue; no overlap." }, { "left": "issue:44704", "right": "issue:44908", "accept": false, - "reason": "AutoProcessor cached_file kwargs and inverse_sqrt scheduler kwargs are unrelated." + "reason": "Processor kwargs forwarding bug and scheduler kwargs bug are different APIs." }, { "left": "issue:42890", "right": "issue:44908", "accept": false, - "reason": "Test flakiness from missing seeds is unrelated to scheduler kwargs." + "reason": "Flaky test due to missing seeds vs scheduler configuration bug; unrelated." }, { "left": "issue:44246", "right": "issue:44273", "accept": false, - "reason": "Import slowdown and lazy-loading breakage are adjacent but distinct bugs." + "reason": "Slow import sometimes and lazy loading not working properly are related in theme, but the reports do not describe the same concrete defect." }, { "left": "issue:43408", "right": "issue:44315", "accept": false, - "reason": "Model-type warning and Liger Kernel application are different concerns." + "reason": "Model-type warning vs model_init/Liger integration issue; different code paths." }, { "left": "issue:43976", "right": "issue:44393", "accept": false, - "reason": "Python version compatibility and Qwen3-VL output bugs are unrelated." + "reason": "Python version compatibility and a Qwen3-VL inference bug are unrelated." }, { "left": "issue:43352", "right": "issue:44393", "accept": false, - "reason": "Flash Attention support gating is unrelated to the Qwen3-VL bbox issue." + "reason": "Flash-Attention-2 support error and Qwen3-VL bounding-box hallucination are different issues." }, { "left": "issue:43576", "right": "issue:44908", "accept": false, - "reason": "v5 env-command breakage is unrelated to scheduler kwarg handling." + "reason": "Broken env command and scheduler kwargs regression are unrelated." }, { "left": "issue:44297", "right": "issue:45230", "accept": false, - "reason": "Tokenizer save metadata mismatch is unrelated to the generic bug report." + "reason": "Tokenizer save metadata mismatch vs a vague bug report; no same underlying change." }, { "left": "issue:43673", "right": "issue:44492", "accept": false, - "reason": "Chunked-prefill cache missing and a typo in cache strategies are not the same defect." + "reason": "Generation cache bug and a typo in cache strategies are not the same defect." }, { "left": "issue:44908", "right": "issue:45071", "accept": false, - "reason": "Scheduler kwargs issue versus type-checking regression." + "reason": "Scheduler kwargs handling and config type checking are unrelated." }, { "left": "issue:44295", "right": "issue:45230", "accept": false, - "reason": "position_ids buffer handling is unrelated to the generic bug report." + "reason": "position_ids buffer access error vs vague bug report; no evidence of duplication." }, { "left": "issue:41669", "right": "issue:44351", "accept": false, - "reason": "Import-star removal/performance and missing HybridCache import are different import issues." + "reason": "Import-star cleanup/performance issue vs missing HybridCache export; related to imports, but not the same bug." }, { "left": "issue:41669", "right": "issue:45341", "accept": false, - "reason": "Import performance issue versus a testing_utils bug." + "reason": "Import performance regression and a testing_utils bug are unrelated." }, { "left": "issue:41669", "right": "issue:44492", "accept": false, - "reason": "Import-star cleanup and cache-strategy typo are unrelated." + "reason": "Import-star performance problem and a cache-strategy typo are different issues." }, { "left": "issue:44393", "right": "issue:45478", "accept": false, - "reason": "Different Qwen3.5/Moe regression than the bbox hallucination report." + "reason": "Different Qwen3.5/Moe bugs with different symptoms and no shared concrete failure mode." }, { "left": "issue:44315", "right": "issue:44829", "accept": false, - "reason": "Liger Kernel model_init behavior and flash_attention_3 training degeneration are distinct issues." + "reason": "Liger-kernel application bug vs flash_attention_3 degenerate training; different paths." }, { "left": "issue:44077", "right": "issue:44861", "accept": false, - "reason": "Optional post_init validation and tied-weight key crashes are different code paths." + "reason": "Optional post_init validation issue vs tied-weight-key AttributeError; unrelated." }, { "left": "issue:43827", "right": "issue:44908", "accept": false, - "reason": "Docs still referencing pipeline() is unrelated to inverse_sqrt scheduler kwargs." + "reason": "Docs still referencing pipeline() vs scheduler kwargs bug; different kinds of changes." }, { "left": "issue:31795", "right": "issue:33453", "accept": false, - "reason": "Model.forward docs confusion and tokenizer-loading regression are unrelated." + "reason": "Documentation confusion in model.forward vs tokenizer loading regression; unrelated." }, { "left": "issue:41669", "right": "issue:44908", "accept": false, - "reason": "Import-star cleanup/performance is unrelated to scheduler kwarg handling." - }, - { - "left": "issue:42371", - "right": "issue:44492", - "accept": false, - "reason": "TF32 settings guidance and a cache-strategy typo are unrelated." - }, - { - "left": "issue:44655", - "right": "issue:44861", - "accept": false, - "reason": "Pipeline save_pretrained failure and tied-weight-key AttributeError are different bugs." + "reason": "Import-star performance regression and scheduler kwargs bug are unrelated." } ] }, - "evaluator_result": null, - "evaluator_used": false, - "retried": false, + "evaluator_result": { + "accept": false, + "feedback": "The cluster summary is broadly reasonable, but it overstates the evidence for the one accepted soft pair. In the packet, issue:44273 and issue:45003 only have titles and a modest similarity score; there are no explicit links or body details proving they are the same bug. Describing 45003 as a concrete cause/root-cause for 44273 is too strong on this evidence. The rest of the soft-edge rejections look conservative, but this accepted duplicate verdict is not sufficiently grounded, so I can\u2019t accept the analyst output." + }, + "evaluator_used": true, + "retried": true, "accepted_nontrivial_soft_edge": false, "error_kind": null, "error_message": null @@ -14317,8 +14203,8 @@ "review_unit_count": 44, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-15354-384", "nodes": [ @@ -14339,6 +14225,7 @@ "issue:43519", "issue:43525", "issue:43572", + "issue:43576", "issue:43606", "issue:43704", "issue:43827", @@ -14351,6 +14238,7 @@ "issue:44485", "issue:44492", "issue:44623", + "issue:44655", "issue:44861", "issue:44908", "issue:44998", @@ -14362,6 +14250,9 @@ "issue:45468" ], "soft_pairs": [ + "issue:42371|issue:44492", + "issue:43576|issue:44273", + "issue:44655|issue:44861", "issue:44360|issue:45468", "issue:45230|issue:45310", "issue:43827|issue:44492", @@ -14386,8 +14277,8 @@ "issue:44908|issue:44998", "issue:44861|issue:45290", "issue:41669|issue:44998", - "issue:43704|issue:44998", "issue:43525|issue:44861", + "issue:43704|issue:44998", "issue:43066|issue:44998", "issue:43572|issue:44861", "issue:31515|issue:38617", @@ -14400,256 +14291,274 @@ "issue:31515|issue:33453", "issue:43298|issue:44230" ], - "prepared_review_unit_hash": "da7980170dcebc6f57633fa387326e973a55caf918940764a24aa8167660d519", + "prepared_review_unit_hash": "6d99e37ee71e9b1c63b7091e95ed0a0e99b3cb77020de6f8a3a9afb3445b46b5", "component_budget": { "node_count": 384, "item_count": 384, - "soft_pair_count": 1442, - "serialized_chars": 743103, - "estimated_input_tokens": 185776, - "estimated_eval_tokens": 371808 + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 }, "budget": { - "node_count": 38, - "item_count": 38, - "soft_pair_count": 37, - "serialized_chars": 32374, - "estimated_input_tokens": 8094, - "estimated_eval_tokens": 16444 + "node_count": 40, + "item_count": 40, + "soft_pair_count": 40, + "serialized_chars": 34369, + "estimated_input_tokens": 8593, + "estimated_eval_tokens": 17442 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "This cluster is largely heterogeneous: most items are unrelated bugs, docs, model-loading, or training issues. There is a small tokenizer/v5 subgroup, but no soft pair is close enough to treat as the same underlying change or bug.", - "confidence": 0.95, - "canonical_issue_reason": "issue 43066 is the most concrete and actionable standalone report in the cluster; it describes a specific tokenizer v5 regression rather than a broad symptom or unrelated subsystem issue.", + "summary": "The cluster is highly heterogeneous: the soft pairs span unrelated bugs in tokenizers, loading, docs, schedulers, model internals, and utilities. None of the candidate pairs look like the same underlying issue/change.", + "confidence": 0.96, + "canonical_issue_reason": null, "canonical_pr_reason": null, - "best_issue_reason": "issue 43066 is the best global issue candidate because it is specific, user-facing, and likely to attract/organize related tokenizer-v5 reports more cleanly than the broader alternatives.", + "best_issue_reason": null, "best_pr_reason": null, "soft_edge_verdicts": [ + { + "left": "issue:42371", + "right": "issue:44492", + "accept": false, + "reason": "TF32 behavior/settings API vs a cache-strategy typo; unrelated symptoms and code paths." + }, + { + "left": "issue:43576", + "right": "issue:44273", + "accept": false, + "reason": "Broken env command in v5 vs lazy-loading failure; different feature areas." + }, + { + "left": "issue:44655", + "right": "issue:44861", + "accept": false, + "reason": "Saving Pipeline objects vs tied-weights key handling; both involve save/load but not the same bug." + }, { "left": "issue:44360", "right": "issue:45468", "accept": false, - "reason": "Different model/component bugs: DSA indexer ReLU behavior vs Gemma4 audio positional encoding." + "reason": "DSA indexer ReLU issue vs Gemma4 audio positional encoding; different model internals." }, { "left": "issue:45230", "right": "issue:45310", "accept": false, - "reason": "One is a generic bug report; the other is a specific Qwen3.5-MoE from_pretrained regression. Not the same issue." + "reason": "Generic bug report vs a specific Qwen3.5 MoE from_pretrained regression; not the same concrete issue." }, { "left": "issue:43827", "right": "issue:44492", "accept": false, - "reason": "Docs referencing removed pipeline APIs and a cache-strategy typo are unrelated." + "reason": "Docs still referencing pipeline() vs cache-strategy typo; unrelated." }, { "left": "issue:44273", "right": "issue:44908", "accept": false, - "reason": "Lazy loading failure and inverse_sqrt scheduler kwarg handling affect different code paths." + "reason": "Lazy loading bug vs inverse_sqrt scheduler kwargs bug; different subsystems." }, { "left": "issue:43519", "right": "issue:44485", "accept": false, - "reason": "Qwen3VL timestamp calculation vs GLM-5 RoPE implementation are different model-specific bugs." + "reason": "Qwen3VL timestamp calculation vs GLM-5 RoPE implementation; both model-specific but distinct defects." }, { "left": "issue:41669", "right": "issue:43704", "accept": false, - "reason": "Import * cleanup/performance issue is unrelated to VRAM leakage in multithreaded dataloaders." + "reason": "Import-star cleanup/performance issue vs VRAM leak in Qwen3ForCausalLM; unrelated." }, { "left": "issue:43519", "right": "issue:44908", "accept": false, - "reason": "Processor timestamp math and scheduler kwargs are unrelated subsystems." + "reason": "Timestamp math in processor vs scheduler kwargs handling; no overlap in code path." }, { "left": "issue:31795", "right": "issue:34689", "accept": false, - "reason": "Documentation confusion about forward args is unrelated to a Llama 3.2 model-loading regression." + "reason": "Documentation confusion in model.forward vs a model-loading regression; unrelated." }, { "left": "issue:31795", "right": "issue:37428", "accept": false, - "reason": "Forward-doc clarification and flash-attention import failure are different problems." + "reason": "Docs around input_ids/past_key_values vs flash-attention import error; different topics." }, { "left": "issue:44315", "right": "issue:45092", "accept": false, - "reason": "Both involve model creation/loading, but one is Liger Kernel application with model_init and the other is remote-code/meta-init incompatibility; not the same bug." + "reason": "Liger kernel not applied with model_init vs meta-init incompatibility for old remote-code checkpoints; not the same bug." }, { "left": "issue:43704", "right": "issue:44485", "accept": false, - "reason": "VRAM leak in threaded inference and RoPE implementation are unrelated." + "reason": "VRAM leak in dataloader-threaded causal LM use vs GLM-5 RoPE implementation; distinct issues." }, { "left": "issue:44485", "right": "issue:45468", "accept": false, - "reason": "Different model components: RoPE implementation vs audio relative positional encoding." + "reason": "GLM-5 RoPE math vs Gemma4 audio positional encoding; different models and failures." }, { "left": "issue:42371", "right": "issue:43704", "accept": false, - "reason": "TF32 configuration guidance is unrelated to a memory leak in dataloader threads." + "reason": "TF32 configuration warning vs VRAM leak; unrelated." }, { "left": "issue:43502", "right": "issue:43519", "accept": false, - "reason": "Local-files-only network access and Qwen3VL timestamp errors are different bugs." + "reason": "local_files_only still making API calls vs timestamp miscalculation; different code paths." }, { "left": "issue:43519", "right": "issue:44623", "accept": false, - "reason": "Timestamp calculation in a processor and missing processor save files are separate issues." + "reason": "Processor timestamp bug vs processor.save_pretrained missing files; both processor-related but not the same defect." }, { "left": "issue:43519", "right": "issue:43704", "accept": false, - "reason": "Processor timestamp bug and multithreaded VRAM leak do not share a concrete code path." + "reason": "Timestamp calculation vs VRAM leak; unrelated." }, { "left": "issue:44263", "right": "issue:45468", "accept": false, - "reason": "A torch.split return-value issue in one indexer is unrelated to Gemma4 audio positional encoding." + "reason": "torch.split return handling in GlmMoeDsaIndexer vs Gemma4 audio positional encoding; different components." }, { "left": "issue:43606", "right": "issue:45468", "accept": false, - "reason": "CPU offload device mismatch for bark-small is unrelated to Gemma4 audio positional encoding." + "reason": "CPU offload device mismatch for bark-small vs Gemma4 audio positional encoding; different models/problems." }, { "left": "issue:44261", "right": "issue:45468", "accept": false, - "reason": "Layernorm epsilon/config precision discussion is unrelated to Gemma4 audio positional encoding." + "reason": "Missing rms_norm_eps in MLA q_a_layernorm vs Gemma4 audio positional encoding; unrelated." }, { "left": "issue:43704", "right": "issue:45341", "accept": false, - "reason": "VRAM leak in dataloader threads and a testing_utils bug are unrelated." + "reason": "VRAM leak in model use vs a testing_utils bug; no concrete overlap." }, { "left": "issue:42913", "right": "issue:43066", "accept": false, - "reason": "Both mention tokenizer v5 behavior, but one is broad behavior drift and the other is a specific decoder-type mismatch; not enough evidence of the same underlying bug." + "reason": "General tokenizer behavior difference across v4/v5 vs wrong decoder type in v5; related area but not clearly the same change." }, { "left": "issue:44908", "right": "issue:44998", "accept": false, - "reason": "Scheduler kwargs handling and an unrelated placeholder issue are not the same." + "reason": "Scheduler kwarg bug vs unrelated issue titled 'Unemployment'; no relation." }, { "left": "issue:44861", "right": "issue:45290", "accept": false, - "reason": "Tied-weight-key handling and chat template crashes involve different code paths." + "reason": "Tied-weight-key crash during loading vs chat-template crash with tool calls; different APIs and failure modes." }, { "left": "issue:41669", "right": "issue:44998", "accept": false, - "reason": "Import cleanup/performance issue is unrelated to the placeholder issue." + "reason": "Import-star cleanup/performance vs unrelated issue; no shared bug." }, { - "left": "issue:43704", - "right": "issue:44998", + "left": "issue:43525", + "right": "issue:44861", "accept": false, - "reason": "VRAM leak bug is unrelated to the placeholder issue." + "reason": "Missing pad_token_id on Llama4Config vs tied-weight-key AttributeError; unrelated." }, { - "left": "issue:43525", - "right": "issue:44861", + "left": "issue:43704", + "right": "issue:44998", "accept": false, - "reason": "Missing pad_token_id on Llama4Config and tied-weight-key attribute handling are different failures." + "reason": "VRAM leak vs unrelated issue; no overlap." }, { "left": "issue:43066", "right": "issue:44998", "accept": false, - "reason": "Tokenizer decoder-type regression is unrelated to the placeholder issue." + "reason": "Tokenizer decoder type bug vs unrelated issue; no relation." }, { "left": "issue:43572", "right": "issue:44861", "accept": false, - "reason": "StableLmConfig pad_token_idx regression and tied-weight-key handling are different issues." + "reason": "StableLmConfig pad_token_idx regression vs tied-weight-key crash; different config and code paths." }, { "left": "issue:31515", "right": "issue:38617", "accept": false, - "reason": "Slow checkpoint loading and an ImportError for layer_type_validation are not the same bug." + "reason": "Slow from_pretrained checkpoint loading vs ImportError for layer_type_validation; unrelated loader problems." }, { "left": "issue:44360", "right": "issue:44998", "accept": false, - "reason": "DSA indexer behavior is unrelated to the placeholder issue." + "reason": "DSA indexer math issue vs unrelated issue; no common bug." }, { "left": "issue:44230", "right": "issue:44623", "accept": false, - "reason": "FP8 inference support for Qwen models and processor save_pretrained missing files are unrelated." + "reason": "FP8 inference support for Qwen3-VL/Qwen3.5 MoE vs processor.save_pretrained file omission; unrelated." }, { "left": "issue:30333", "right": "issue:31515", "accept": false, - "reason": "MLFlow job-status reporting and slow from_pretrained checkpoint loading are different problems." + "reason": "MLFlow reporting on failed training jobs vs slow checkpoint loading; unrelated." }, { "left": "issue:31515", "right": "issue:36296", "accept": false, - "reason": "Slow checkpoint loading is unrelated to a tensor-parallel training bug." + "reason": "Slow checkpoint loading vs tensor parallel training bug; different subsystems." }, { "left": "issue:31515", "right": "issue:31795", "accept": false, - "reason": "Loading performance and forward-API documentation confusion are not the same issue." + "reason": "Loading performance regression vs documentation confusion in model.forward; unrelated." }, { "left": "issue:44230", "right": "issue:45310", "accept": false, - "reason": "FP8 inference support and Qwen3.5-MoE from_pretrained failure are separate Qwen-related reports with different concrete failures." + "reason": "FP8 support request for Qwen3-VL/Qwen3.5 MoE infer vs a from_pretrained error on transformers>=5.4.0; different concrete problems." }, { "left": "issue:31515", "right": "issue:33453", "accept": false, - "reason": "Slow checkpoint loading and tokenizer loading regression are distinct bugs." + "reason": "Slow from_pretrained loading vs tokenizer-loading regression; not the same bug." }, { "left": "issue:43298", "right": "issue:44230", "accept": false, - "reason": "Continuous batching for audio models and Qwen fp8 inference support are unrelated." + "reason": "Continuous batching not supporting audio models vs FP8 support for Qwen3-VL/Qwen3.5 MoE; different features and models." } ] }, @@ -14668,8 +14577,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-26350-2", "nodes": [ @@ -14679,12 +14588,12 @@ "soft_pairs": [ "issue:26350|issue:33260" ], - "prepared_review_unit_hash": "2770400a9e467b8f30b6d6d65fa3cd3d0a8edbfab7eb4a7129a90ad1be3ca6ee", + "prepared_review_unit_hash": "b16df79519325d4d44019b581c01e8b7983539a7c1b09b039b5d16e5f1202548", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 1589, + "serialized_chars": 1590, "estimated_input_tokens": 398, "estimated_eval_tokens": 1052 }, @@ -14692,7 +14601,7 @@ "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 1589, + "serialized_chars": 1590, "estimated_input_tokens": 398, "estimated_eval_tokens": 1052 }, @@ -14701,8 +14610,8 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These are two separate community feature requests, not duplicates: one asks for Flash Attention 2 support for more architectures, the other for GGUF support for more architectures. The soft similarity signal is too broad to treat them as the same underlying issue.", - "confidence": 0.97, + "summary": "These are two separate feature requests for different model/runtime integration work: one asks for Flash Attention 2 support across more architectures, the other for GGUF support across more architectures. They share the same contribution framing but not the same underlying change.", + "confidence": 0.98, "canonical_issue_reason": null, "canonical_pr_reason": null, "best_issue_reason": null, @@ -14712,7 +14621,7 @@ "left": "issue:26350", "right": "issue:33260", "accept": false, - "reason": "Different feature areas and codepaths (Flash Attention 2 vs GGUF); they share only a broad \u2018more architectures\u2019 theme, so they are not the same underlying issue." + "reason": "Different features and code paths: Flash Attention 2 support vs GGUF support. Similar wording about architecture coverage does not make them duplicates." } ] }, @@ -14731,8 +14640,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-27088-4", "nodes": [ @@ -14749,7 +14658,7 @@ "issue:43834|issue:44365", "issue:44365|issue:44370" ], - "prepared_review_unit_hash": "6ce48d1491b884fcc9c0ee1936c7fddc7cc03167a8eb6addbd4efed853a5a838", + "prepared_review_unit_hash": "536033348200ca16ef36127a4729cd4bce6a19ee705e4804a5339a55471f796e", "component_budget": { "node_count": 4, "item_count": 4, @@ -14771,48 +14680,48 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These are issue-template style requests for translating docs, but they appear to target different languages rather than the same underlying task. No PRs are present.", - "confidence": 0.79, - "canonical_issue_reason": "issue:27088 is the only specific, fully named request (Turkish) and is the most concrete representative of the cluster.", + "summary": "This cluster looks like repeated i18n-doc translation request templates, not true duplicates. The only concrete request is Turkish; the others are generic placeholder-language variants that likely refer to different languages, so they should stay separate.", + "confidence": 0.87, + "canonical_issue_reason": "issue:27088 is the most concrete and identifiable request: it names Turkish explicitly, while the others are generic template-style placeholders with no specific language shown.", "canonical_pr_reason": null, - "best_issue_reason": "issue:27088 is the clearest issue to keep as the representative because it is specific and not a placeholder-style template title.", + "best_issue_reason": "issue:27088 is the best representative to keep because it is specific, human-readable, and clearly scoped to one translation request; the others appear to be less concrete template instances.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:27088", "right": "issue:44370", "accept": false, - "reason": "Both are translation-request issues, but they do not indicate the same target language or the same concrete change; this looks like template similarity only." + "reason": "Same translation-request template, but no evidence they target the same language; placeholder titles suggest different underlying requests." }, { "left": "issue:27088", "right": "issue:44365", "accept": false, - "reason": "Same broad i18n-docs theme, but no evidence they refer to the same language or duplicate request." + "reason": "Both are i18n-doc translation requests, but the specific language is different/unspecified, so this is template similarity rather than a duplicate." }, { "left": "issue:27088", "right": "issue:43834", "accept": false, - "reason": "The titles suggest different language-specific translation requests, so they should not be merged as duplicates." + "reason": "Turkish is a concrete language request; the other issue is a generic placeholder. They do not appear to be the same underlying issue." }, { "left": "issue:43834", "right": "issue:44370", "accept": false, - "reason": "Highly similar template structure, but no proof they are the same translation task rather than separate language requests." + "reason": "Highly similar template wording, but nothing indicates the same language target or the same exact translation task." }, { "left": "issue:43834", "right": "issue:44365", "accept": false, - "reason": "These look like repeated template issues, not the same underlying bug/change." + "reason": "These look like separate instances of the same template for different languages, not one duplicated issue." }, { "left": "issue:44365", "right": "issue:44370", "accept": false, - "reason": "No concrete overlap beyond the generic translation-docs template; reject as duplicate." + "reason": "Only template-level similarity is visible; without a shared concrete language or task, they should not be merged as duplicates." } ] }, @@ -14831,8 +14740,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-36817-14", "nodes": [ @@ -14881,7 +14790,7 @@ "issue:43837|issue:45027", "issue:43837|issue:45177" ], - "prepared_review_unit_hash": "19087cc10efede8e85b35f9957c304b1f327961cbb2b15e296c0015728584579", + "prepared_review_unit_hash": "168ed3567329b21338d6486f8c4ce8b16cb462c8b4212c63d7c878503e11e042", "component_budget": { "node_count": 14, "item_count": 14, @@ -14903,180 +14812,180 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "All items are distinct model-request issues spanning different architectures, model families, or modalities. There is no true duplicate cluster here; the closest pair is Qwen3-TTS vs Qwen3-ASR, but those are separate code paths and should stay separate.", - "confidence": 0.97, - "canonical_issue_reason": null, + "summary": "These are all model-request issues in the same broad transformers ecosystem, but they describe different concrete models/features rather than duplicate bugs or the same change.", + "confidence": 0.28, + "canonical_issue_reason": "Issue 41115 is the most representative broad architecture-add request in the set, with stronger activity than most others and a generic title that fits the cluster\u2019s overall theme.", "canonical_pr_reason": null, - "best_issue_reason": null, + "best_issue_reason": "41115 is the best standalone issue to anchor the cluster because it is concrete, broadly phrased, and closest to the shared pattern of requesting new model support.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:36817", "right": "issue:41211", "accept": false, - "reason": "Different model families (EuroBert vs DEIMv2); same broad request type only." + "reason": "Different model families (EuroBERT vs DEIMv2); same add-model pattern only." }, { "left": "issue:36817", "right": "issue:43671", "accept": false, - "reason": "EuroBert config request vs Qwen3-TTS support; unrelated models." + "reason": "EuroBERT config request and Qwen3-TTS support are different models and modalities." }, { "left": "issue:37015", "right": "issue:43837", "accept": false, - "reason": "NeoBERT and Qwen3-ASR are different model requests and code paths." + "reason": "NeoBERT and Qwen3-ASR are unrelated concrete additions." }, { "left": "issue:37015", "right": "issue:44121", "accept": false, - "reason": "NeoBERT vs weight-sparse transformer; different architectures." + "reason": "NeoBERT support and weight-sparse transformer support are different code/model changes." }, { "left": "issue:37015", "right": "issue:43909", "accept": false, - "reason": "NeoBERT vs LFM2.5 Audio; unrelated model families." + "reason": "NeoBERT vs LFM2.5 Audio are different model requests." }, { "left": "issue:37015", "right": "issue:42503", "accept": false, - "reason": "NeoBERT and ModernVBERT are separate model additions." + "reason": "NeoBERT and ModernVBERT are distinct model additions." }, { "left": "issue:37015", "right": "issue:45027", "accept": false, - "reason": "NeoBERT vs Voxtral TTS; different model targets." + "reason": "NeoBERT and Voxtral TTS are different model families and tasks." }, { "left": "issue:41115", "right": "issue:43837", "accept": false, - "reason": "MiniCPM3 architecture request vs Qwen3-ASR support; not the same change." + "reason": "MiniCPM3 architecture work and Qwen3-ASR support are not the same underlying change." }, { "left": "issue:41115", "right": "issue:44121", "accept": false, - "reason": "MiniCPM3 vs sparse-transformer request; different concrete model work." + "reason": "MiniCPM3 support and the sparse-transformer request target different implementations." }, { "left": "issue:41115", "right": "issue:42503", "accept": false, - "reason": "MiniCPM3 and ModernVBERT are distinct model requests." + "reason": "MiniCPM3 and ModernVBERT are separate model-support requests." }, { "left": "issue:41115", "right": "issue:45027", "accept": false, - "reason": "MiniCPM3 vs Voxtral TTS; unrelated model families." + "reason": "MiniCPM3 architecture support and Voxtral-TTS support are unrelated changes." }, { "left": "issue:41211", "right": "issue:43671", "accept": false, - "reason": "DEIMv2 vs Qwen3-TTS; separate model additions." + "reason": "DEIMv2 and Qwen3-TTS are different model additions." }, { "left": "issue:42503", "right": "issue:43837", "accept": false, - "reason": "ModernVBERT vs Qwen3-ASR; different model targets." + "reason": "ModernVBERT and Qwen3-ASR are distinct model requests." }, { "left": "issue:42503", "right": "issue:43671", "accept": false, - "reason": "ModernVBERT vs Qwen3-TTS; distinct requests." + "reason": "ModernVBERT and Qwen3-TTS are different concrete supports." }, { "left": "issue:42503", "right": "issue:42738", "accept": false, - "reason": "ModernVBERT request vs a generic BERT-like RoPE proposal; not the same issue." + "reason": "One is a specific model request, the other is a broad BERT/RoPE feature proposal." }, { "left": "issue:42738", "right": "issue:45027", "accept": false, - "reason": "Generic BERT-like RoPE discussion vs Voxtral TTS support; unrelated." + "reason": "BERT-like models with RoPE and Voxtral-TTS are unrelated requests." }, { "left": "issue:43671", "right": "issue:43837", "accept": false, - "reason": "Same Qwen3 family, but TTS vs ASR are different modalities and likely separate code paths." + "reason": "Both are Qwen3 audio requests, but TTS and ASR are separate concrete changes." }, { "left": "issue:43671", "right": "issue:45027", "accept": false, - "reason": "Qwen3-TTS vs Voxtral TTS; different model families." + "reason": "Qwen3-TTS and Voxtral-TTS are different model integrations." }, { "left": "issue:43671", - "right": "issue:45177", + "right": "issue:43909", "accept": false, - "reason": "Qwen3-TTS vs DeepSeek-OCR2; unrelated model requests." + "reason": "TTS support for Qwen3 and LFM2.5 Audio 1.5B are different model additions." }, { "left": "issue:43671", - "right": "issue:44121", + "right": "issue:45177", "accept": false, - "reason": "Qwen3-TTS vs weight-sparse transformer; different changes." + "reason": "Qwen3-TTS support and DeepSeek-OCR2 are different modalities and code paths." }, { - "left": "issue:43837", - "right": "issue:43909", + "left": "issue:43671", + "right": "issue:44121", "accept": false, - "reason": "Qwen3-ASR support vs LFM2.5 Audio; separate audio model requests." + "reason": "Audio model support and sparse-transformer support are unrelated." }, { "left": "issue:43837", - "right": "issue:44121", + "right": "issue:45322", "accept": false, - "reason": "Qwen3-ASR vs sparse-transformer request; not the same bug/change." + "reason": "Qwen3-ASR and EUPE are different model requests." }, { "left": "issue:43837", - "right": "issue:45027", + "right": "issue:43909", "accept": false, - "reason": "Qwen3-ASR vs Voxtral TTS; different models and tasks." + "reason": "ASR support and LFM2.5 Audio support are related only at a broad audio level, not the same change." }, { - "left": "issue:43837", - "right": "issue:45177", + "left": "issue:43909", + "right": "issue:44205", "accept": false, - "reason": "Qwen3-ASR vs DeepSeek-OCR2; unrelated." + "reason": "LFM2.5 Audio and SAM3-LiteText are different model families." }, { - "left": "issue:43837", + "left": "issue:44121", "right": "issue:45322", "accept": false, - "reason": "Qwen3-ASR vs EUPE; different model families." + "reason": "Sparse-transformer support and EUPE support are unrelated changes." }, { - "left": "issue:43909", - "right": "issue:44205", + "left": "issue:43837", + "right": "issue:44121", "accept": false, - "reason": "LFM2.5 Audio vs SAM3-LiteText; distinct model support requests." + "reason": "Qwen3-ASR and the sparse-transformer request target different code paths." }, { - "left": "issue:43671", - "right": "issue:43909", + "left": "issue:43837", + "right": "issue:45027", "accept": false, - "reason": "Qwen3-TTS vs LFM2.5 Audio; different audio models and code paths." + "reason": "Qwen3-ASR and Voxtral-TTS are different concrete audio-model integrations." }, { - "left": "issue:44121", - "right": "issue:45322", + "left": "issue:43837", + "right": "issue:45177", "accept": false, - "reason": "Sparse transformer request vs EUPE; unrelated model additions." + "reason": "Qwen3-ASR and DeepSeek-OCR2 are different tasks and models." } ] }, @@ -15095,8 +15004,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-37934-6", "nodes": [ @@ -15117,82 +15026,82 @@ "issue:43668|issue:44031", "issue:43668|issue:43813" ], - "prepared_review_unit_hash": "04e2237cec87510ce1e840ce44f7f7109d957411609a42436a0b0b5f7864c5a4", + "prepared_review_unit_hash": "0ab1a9b3f48581e5ce6637fec5422564f2d1d4e7401a049d1bdd9149206412c1", "component_budget": { "node_count": 6, "item_count": 6, "soft_pair_count": 8, - "serialized_chars": 6294, - "estimated_input_tokens": 1574, - "estimated_eval_tokens": 3404 + "serialized_chars": 6291, + "estimated_input_tokens": 1573, + "estimated_eval_tokens": 3402 }, "budget": { "node_count": 6, "item_count": 6, "soft_pair_count": 8, - "serialized_chars": 6294, - "estimated_input_tokens": 1574, - "estimated_eval_tokens": 3404 + "serialized_chars": 6291, + "estimated_input_tokens": 1573, + "estimated_eval_tokens": 3402 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "No duplicate cluster here: the issues span unrelated bugs across different models and tokenizer/config code paths. All soft links look like false positives from superficial wording overlap.", - "confidence": 0.98, - "canonical_issue_reason": "No canonical duplicate stands out; these six issues describe different problems in different components rather than the same underlying bug.", + "summary": "These issues are all superficially similar only at the wording level, but they describe different bugs across unrelated components (norm/config typing, Mask2Former loss handling, tokenizer regex warnings, and a typo in a quantization attribute). I do not see a true duplicate cluster here.", + "confidence": 0.93, + "canonical_issue_reason": "No single issue clearly represents a duplicate set; the items span distinct subsystems and problem types, so there is no good canonical duplicate issue.", "canonical_pr_reason": null, - "best_issue_reason": "issue:44031 is the broadest and most impactful report in the set, with the most discussion and the clearest user-facing regression, so it is the best representative issue if one must be chosen.", + "best_issue_reason": "Issue 43668 is the cleanest representative only in the sense that it is a narrowly scoped, concrete bug report, but it is not a cluster canonical because the other issues are unrelated.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:37934", "right": "issue:43668", "accept": false, - "reason": "Different problems: one questions Llama4TextL2Norm semantics, the other is a ModernBERT config type-hint issue." + "reason": "Different problems: Llama4TextL2Norm/RMS norm semantics vs ModernBERTConfig type-hint error. No shared code path or concrete duplicate evidence." }, { "left": "issue:42936", "right": "issue:43240", "accept": false, - "reason": "Mask2Former's ignore_value handling and missing kwargs in loss calculation are separate training bugs with different code paths." + "reason": "Mask2Former ignore_value handling and kwargs not reaching loss calculation are different bugs, even though both touch loss-related code." }, { "left": "issue:42936", "right": "issue:44031", "accept": false, - "reason": "Unrelated areas: model-specific ignore_value usage versus a tokenizer regex warning regression." + "reason": "Mask2Former loss ignore_value bug is unrelated to tokenizer regex warning behavior." }, { "left": "issue:42936", "right": "issue:43813", "accept": false, - "reason": "Mask2Former ignore_value handling is unrelated to a quantization typo in a different component." + "reason": "Different subsystems and failure modes: model loss parameter use vs a typo in an attribute name." }, { "left": "issue:42936", "right": "issue:43668", "accept": false, - "reason": "Different subsystems and bug types: Mask2Former behavior versus a ModernBERTConfig annotation error." + "reason": "Mask2Former loss issue and ModernBERT config type hint issue are unrelated." }, { "left": "issue:43240", "right": "issue:43668", "accept": false, - "reason": "Missing kwargs in loss computation is not the same as an incorrect config type hint." + "reason": "Loss kwargs forwarding bug is not the same as an incorrect config type hint." }, { "left": "issue:43668", "right": "issue:44031", "accept": false, - "reason": "A config type-hint bug and a tokenizer regex warning are distinct issues affecting different paths." + "reason": "ModernBERTConfig typing issue and tokenizer regex warning issue do not describe the same underlying defect." }, { "left": "issue:43668", "right": "issue:43813", "accept": false, - "reason": "ModernBERTConfig typing and a typo in orig_conversion.quantization_operation are unrelated." + "reason": "A type-hint mistake in ModernBERTConfig is unrelated to a typo in quantization-operation naming." } ] }, @@ -15211,8 +15120,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-41084-2", "nodes": [ @@ -15222,12 +15131,12 @@ "soft_pairs": [ "issue:41084|issue:43037" ], - "prepared_review_unit_hash": "72dd3f4e2809d48ad23ef35083a350e95413d8ad40a0658e58ddd96f64918f47", + "prepared_review_unit_hash": "d32cdff39c4e1d0b88ebeb119b69fc9c0518059ad9a72492dbebb8435786e533", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 1479, + "serialized_chars": 1480, "estimated_input_tokens": 370, "estimated_eval_tokens": 996 }, @@ -15235,7 +15144,7 @@ "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 1479, + "serialized_chars": 1480, "estimated_input_tokens": 370, "estimated_eval_tokens": 996 }, @@ -15244,18 +15153,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "The two issues are unrelated feature requests: one about block decoding and the other about DeepSeek v3.2 support. They do not describe the same underlying bug or change, so this cluster should not be merged as duplicates.", + "summary": "The two issues are unrelated: one requests Block Decoding support, the other requests DeepSeek v3.2 support. There is no shared codepath, filename, or explicit linkage suggesting duplication.", "confidence": 0.97, - "canonical_issue_reason": null, + "canonical_issue_reason": "Issue 41084 is the more concrete, feature-oriented request (\u201cSet Block Decoding\u201d) and is not dependent on a model-specific support request.", "canonical_pr_reason": null, - "best_issue_reason": null, + "best_issue_reason": "Issue 41084 is the better canonical anchor because it names a specific functionality, whereas 43037 is a broader model-support request.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:41084", "right": "issue:43037", "accept": false, - "reason": "Different requests affecting different areas: block decoding vs model support. No shared code-path, explicit target, or concrete duplicate behavior." + "reason": "Different underlying requests: block decoding vs DeepSeek v3.2 support. The similarity is only superficial and not enough to treat them as duplicates." } ] }, @@ -15274,8 +15183,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-43598-2", "nodes": [ @@ -15285,12 +15194,12 @@ "soft_pairs": [ "issue:43598|issue:43600" ], - "prepared_review_unit_hash": "f8cd2e0d1413e2e4edde95a4a79432b7ed2b2e8c11921d12706e9f32cfb3388e", + "prepared_review_unit_hash": "4103cf883f3a49d9259d06e6f1f5bc9e9d370f32c18573bdf5bd7eede20b75a0", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 1584, + "serialized_chars": 1582, "estimated_input_tokens": 396, "estimated_eval_tokens": 1048 }, @@ -15298,7 +15207,7 @@ "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 1584, + "serialized_chars": 1582, "estimated_input_tokens": 396, "estimated_eval_tokens": 1048 }, @@ -15307,18 +15216,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These two issues are related to custom trainer support, but they describe different problems: one is about `remove_unused_column` customizability, the other about FLOPS estimation for custom trainers. I would not merge them as duplicates.", - "confidence": 0.89, - "canonical_issue_reason": "No true canonical issue: the reports point to separate fixes in Trainer behavior, so they should remain distinct.", + "summary": "The two issues are related only in that both mention custom trainers, but they describe different problems: one is about `remove_unused_column` behavior in `Trainer`, the other is about FLOPS estimation. They should not be treated as duplicates.", + "confidence": 0.93, + "canonical_issue_reason": null, "canonical_pr_reason": null, - "best_issue_reason": "Issue #43598 is the broader customizability discussion and is the better representative of the cluster theme, if one must be chosen.", + "best_issue_reason": "Issue 43598 is the more concrete Trainer behavior/customizability report, but it still does not subsume 43600; the latter is a separate FLOPS-accounting concern.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43598", "right": "issue:43600", "accept": false, - "reason": "They share a general Trainer/custom-trainer theme, but the underlying bugs are different: column removal behavior vs FLOPS estimation logic." + "reason": "Same broad area (custom trainers), but different underlying bugs: `remove_unused_column` customizability versus FLOPS estimation. They would not be merged as one issue." } ] }, @@ -15337,8 +15246,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-43684-2", "nodes": [ @@ -15348,12 +15257,12 @@ "soft_pairs": [ "issue:43684|issue:44405" ], - "prepared_review_unit_hash": "ac7721c472cca869f1db6fc83d974a20f8a39fba28d7bf51c8e9960e043276fd", + "prepared_review_unit_hash": "718ff7d60e2025fb7923a1f824c9e1fc0d0b79990e64c22f00360e466157782a", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 1612, + "serialized_chars": 1611, "estimated_input_tokens": 403, "estimated_eval_tokens": 1062 }, @@ -15361,7 +15270,7 @@ "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 1612, + "serialized_chars": 1611, "estimated_input_tokens": 403, "estimated_eval_tokens": 1062 }, @@ -15370,18 +15279,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "The two issues are related only by the Qwen ecosystem, but they request different capabilities: one is about registering Qwen3-Omni with AutoModel/AutoModelForConditionalGeneration, while the other asks for AutoModelForSequenceClassification support for Qwen3.5. They are not duplicates of the same underlying bug or change.", - "confidence": 0.96, - "canonical_issue_reason": "issue:43684 is the broader, more central request around model registration and auto-class support, making it the best representative issue in this small cluster.", + "summary": "These are not duplicates: one requests AutoModel/AutoModelForConditionalGeneration registration for Qwen3-Omni, while the other requests AutoModelForSequenceClassification support for Qwen3.5. They touch different model families and different auto-class entry points.", + "confidence": 0.94, + "canonical_issue_reason": "issue:43684 is the broader, more central registration request and best represents the cluster\u2019s general theme of adding missing AutoModel support.", "canonical_pr_reason": null, - "best_issue_reason": "issue:43684 is the strongest issue to keep as the cluster representative because it targets a core registration gap rather than a narrower task-specific addition.", + "best_issue_reason": "issue:43684 is the best representative because it is the broader core-model registration task; issue:44405 is a narrower, different head-specific support request.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43684", "right": "issue:44405", "accept": false, - "reason": "Different model variants and different auto-class targets: Qwen3-Omni registration for AutoModel/AutoModelForConditionalGeneration vs Qwen3.5 sequence-classification support. They do not look like the same underlying fix." + "reason": "Different model families and different features: Qwen3-Omni registration for AutoModel/AutoModelForConditionalGeneration vs Qwen3.5 sequence-classification support. Not the same underlying bug/change." } ] }, @@ -15400,8 +15309,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-43835-4", "nodes": [ @@ -15418,80 +15327,77 @@ "issue:43845|issue:43859", "issue:43846|issue:43859" ], - "prepared_review_unit_hash": "fc28462ff87a90654ab3fe0d45e8016496c6df5db65099ad863a74bbfe5f6ed0", + "prepared_review_unit_hash": "2e8c6528fa82edb81e5cb9233bd1f2fc53677184b19d347f9d3fc994a0e0a9d5", "component_budget": { "node_count": 4, "item_count": 4, "soft_pair_count": 6, - "serialized_chars": 4266, - "estimated_input_tokens": 1067, - "estimated_eval_tokens": 2390 + "serialized_chars": 4264, + "estimated_input_tokens": 1066, + "estimated_eval_tokens": 2388 }, "budget": { "node_count": 4, "item_count": 4, "soft_pair_count": 6, - "serialized_chars": 4266, - "estimated_input_tokens": 1067, - "estimated_eval_tokens": 2390 + "serialized_chars": 4264, + "estimated_input_tokens": 1066, + "estimated_eval_tokens": 2388 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Three issues share the exact title \"huggingface \" and appear to be duplicates; the \"Fett\" issue is unrelated by title and should remain separate.", - "confidence": 0.78, - "canonical_issue_reason": "Issue #43845 is the earliest of the identical-title \"huggingface \" issues, making it the best representative for the duplicate cluster.", + "summary": "The cluster is mostly low-signal: three nearly empty issues titled \"huggingface\" and one unrelated short issue titled \"Fett\". The title match alone is not enough to confirm duplicate underlying bugs, so I would not merge any of the soft edges.", + "confidence": 0.58, + "canonical_issue_reason": "Issue 43845 is the earliest of the three near-identical \"huggingface\" reports, but the bodies are too sparse to establish a concrete duplicate target with confidence.", "canonical_pr_reason": null, - "best_issue_reason": "Issue #43845 is the cleanest canonical choice because it is earliest and matches the other near-identical issues exactly.", + "best_issue_reason": "If one issue must represent the cluster, 43845 is the safest pick because it is the earliest repeated title; however, the content is too thin to treat it as a strong canonical bug report.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:43835", "right": "issue:43845", "accept": false, - "reason": "Different titles with no shared target or code evidence; not enough to treat as the same issue." + "reason": "Different titles and no body/detail evidence tying them to the same underlying bug." }, { "left": "issue:43835", "right": "issue:43846", "accept": false, - "reason": "Different titles with no shared target or code evidence; not enough to treat as the same issue." + "reason": "Different titles and no concrete shared issue description beyond weak similarity." }, { "left": "issue:43835", "right": "issue:43859", "accept": false, - "reason": "Different titles with no shared target or code evidence; not enough to treat as the same issue." + "reason": "Different titles and no explicit target or technical detail indicating the same problem." }, { "left": "issue:43845", "right": "issue:43846", - "accept": true, - "reason": "Same exact title and very close timing strongly suggest the same underlying issue." + "accept": false, + "reason": "Same short title, but the reports are too empty to confirm they describe the same concrete bug." }, { "left": "issue:43845", "right": "issue:43859", - "accept": true, - "reason": "Same exact title and same topic strongly indicate a duplicate report." + "accept": false, + "reason": "Same short title, but there is not enough detail to safely treat these as duplicate reports of one bug." }, { "left": "issue:43846", "right": "issue:43859", - "accept": true, - "reason": "Same exact title and same topic strongly indicate a duplicate report." + "accept": false, + "reason": "Same short title, but the available evidence is too thin to confirm a shared underlying issue." } ] }, - "evaluator_result": { - "accept": true, - "feedback": "The summary is grounded in the packet: three issues have the exact same title, and #43845 is the earliest among them. The soft-edge decisions are conservative enough given the limited evidence, and the unrelated 'Fett' issue is correctly kept separate." - }, - "evaluator_used": true, + "evaluator_result": null, + "evaluator_used": false, "retried": false, - "accepted_nontrivial_soft_edge": true, + "accepted_nontrivial_soft_edge": false, "error_kind": null, "error_message": null }, @@ -15503,8 +15409,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-44371-5", "nodes": [ @@ -15525,12 +15431,12 @@ "issue:44572|issue:44937", "issue:44573|issue:44937" ], - "prepared_review_unit_hash": "188c3a1f9cc64a40eff79ff8f53f947ffeebfa5859770ac13a8bc48440c4b596", + "prepared_review_unit_hash": "616eca8ac804bb082d3b2a9b45fd5f7259c8ef66031a1e7cb6561c55445a91d7", "component_budget": { "node_count": 5, "item_count": 5, "soft_pair_count": 9, - "serialized_chars": 5785, + "serialized_chars": 5788, "estimated_input_tokens": 1447, "estimated_eval_tokens": 3150 }, @@ -15538,7 +15444,7 @@ "node_count": 5, "item_count": 5, "soft_pair_count": 9, - "serialized_chars": 5785, + "serialized_chars": 5788, "estimated_input_tokens": 1447, "estimated_eval_tokens": 3150 }, @@ -15547,76 +15453,73 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "This cluster is mostly unrelated low-information spam/issues. The only plausible duplicates are the three exact \"\" reports; the documentation suggestion and \"Google LLC\" post look separate.", - "confidence": 0.77, - "canonical_issue_reason": "issue:44371 is the best representative of the exact \"\" trio: it is one of the identical spam-style reports, appears earliest, and sits at the center of the strongest soft matches.", + "summary": "These are low-information spam/documentation-like issues with no concrete shared bug or change, so they should not be deduplicated into one canonical thread.", + "confidence": 0.93, + "canonical_issue_reason": "No clear canonical duplicate exists: the issues do not describe the same problem, and the available text is too generic/incoherent to establish a single underlying report.", "canonical_pr_reason": null, - "best_issue_reason": "issue:44371 is the best issue to anchor the cluster because it matches the other two exact spam items and is the earliest of that duplicate-like set.", + "best_issue_reason": "No issue is a good global canonical because the cluster does not form a real duplicate set; the titles point to unrelated spam/marketing-style reports rather than one defect.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:44371", "right": "issue:44572", - "accept": true, - "reason": "Same exact spam-style issue: identical title pattern and strongest token overlap; likely the same duplicate report." + "accept": false, + "reason": "Both are generic spam markers, but there is no concrete shared bug or change to prove they are the same underlying issue." }, { "left": "issue:44371", "right": "issue:44573", - "accept": true, - "reason": "Same exact spam-style issue as the other \"\" reports; no meaningful distinction is visible." + "accept": false, + "reason": "Identical-looking spam titles are not enough to confirm a duplicate; no substantive issue details are present." }, { "left": "issue:44371", "right": "issue:44418", "accept": false, - "reason": "Different subject matter: one is spam-like, the other is a documentation enhancement request." + "reason": "One is spam-like and the other is a documentation suggestion; they do not describe the same problem." }, { "left": "issue:44371", "right": "issue:44937", "accept": false, - "reason": "Different titles and apparent intent; \"Check out \\\"Google LLC\\\"\" is a separate promotional/spam-like post, not the same report." + "reason": "The titles refer to different topics, and there is no evidence of a shared underlying defect or request." }, { "left": "issue:44418", "right": "issue:44572", "accept": false, - "reason": "Documentation enhancement and spam placeholder are not the same underlying issue." + "reason": "Documentation enhancement and spam are not the same user problem." }, { "left": "issue:44418", "right": "issue:44573", "accept": false, - "reason": "Documentation enhancement and spam placeholder are not the same underlying issue." + "reason": "Documentation enhancement and spam are not the same user problem." }, { "left": "issue:44572", "right": "issue:44573", - "accept": true, - "reason": "Both are identical \"\" reports with matching high similarity, so they look like duplicates." + "accept": false, + "reason": "Both are generic spam titles, but the content is too empty to safely treat them as the same underlying report." }, { "left": "issue:44572", "right": "issue:44937", "accept": false, - "reason": "Both may be low-quality posts, but they are not the same concrete issue or report." + "reason": "Spam and a Google LLC mention are different topics with no concrete shared bug/change." }, { "left": "issue:44573", "right": "issue:44937", "accept": false, - "reason": "Both may be low-quality posts, but they are not the same concrete issue or report." + "reason": "Spam and a Google LLC mention are different topics with no concrete shared bug/change." } ] }, - "evaluator_result": { - "accept": true, - "feedback": "The cluster summary is grounded in the packet: the three \"\" issues are the only clearly duplicate-like items, and the documentation/\"Google LLC\" entries are conservatively treated as separate. The soft-edge accepts/rejects are appropriately cautious given the available evidence." - }, - "evaluator_used": true, + "evaluator_result": null, + "evaluator_used": false, "retried": false, - "accepted_nontrivial_soft_edge": true, + "accepted_nontrivial_soft_edge": false, "error_kind": null, "error_message": null }, @@ -15628,8 +15531,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-45145-2", "nodes": [ @@ -15639,7 +15542,7 @@ "soft_pairs": [ "issue:45145|issue:45151" ], - "prepared_review_unit_hash": "6efd60e378117c5052bbd8ed701f0af63c471165fe99f0ac392aaba79fbb3375", + "prepared_review_unit_hash": "edc315d7efaaef003d33479330b683c205cf7fccfa2e14cedd97f52410a3c64c", "component_budget": { "node_count": 2, "item_count": 2, @@ -15661,28 +15564,28 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These two closed issues look highly similar from metadata alone: identical titles, close body lengths, and a soft-similarity signal. That supports treating them as likely duplicates, but the available evidence is not strong enough to confidently say they are the same underlying issue without body text, targets, or file-level detail.", - "confidence": 0.74, - "canonical_issue_reason": "#45145 is the better canonical anchor because it has the inbound reference, so it appears to be the more linked of the two otherwise similar issues.", + "summary": "The two closed issues appear to be duplicates of the same request/problem: identical titles, nearly identical body sizes, and strong text similarity. Issue 45145 is the better canonical choice because it was created earlier and has an inbound reference.", + "confidence": 0.98, + "canonical_issue_reason": "Issue 45145 is the best canonical issue: same title/topic as 45151, created earlier, and it already has an inbound reference.", "canonical_pr_reason": null, - "best_issue_reason": "#45145 is the best issue candidate for the cluster because it is the referenced thread and serves as the most visible anchor among the two similar issues.", + "best_issue_reason": "Issue 45145 is the strongest representative of the duplicate set because it is earlier and has slightly more link activity.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:45145", "right": "issue:45151", - "accept": false, - "reason": "The metadata suggests a likely duplicate, but we do not have the actual body content, targets, or filenames. With only identical titles, similar body sizes, and a soft-similarity signal, the evidence is too thin to accept this edge deterministically." + "accept": true, + "reason": "Identical titles and very similar bodies strongly indicate the same underlying issue/request." } ] }, "evaluator_result": { - "accept": false, - "feedback": "The duplicate-cluster conclusion is directionally reasonable, but the analyst is a bit too confident for the evidence available. The packet only shows identical titles, similar body lengths, and a soft-similarity signal; it does not expose the actual body content, targets, or filenames, so calling them the same underlying issue is stronger than the metadata alone supports. The canonical choice for #45145 is grounded by its inbound reference, but the soft-edge acceptance should be more conservative given deterministic_accept=false." + "accept": true, + "feedback": "The summary is grounded in the packet: both issues have identical titles, similar body sizes, and the pair is marked with soft_similarity. Choosing 45145 as canonical is also supported by its earlier creation time and inbound reference. The soft-edge verdict is conservative enough given the evidence available." }, "evaluator_used": true, - "retried": true, - "accepted_nontrivial_soft_edge": false, + "retried": false, + "accepted_nontrivial_soft_edge": true, "error_kind": null, "error_message": null }, @@ -15748,8 +15651,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-45672-2", "nodes": [ @@ -15759,7 +15662,7 @@ "soft_pairs": [ "issue:45672|issue:45674" ], - "prepared_review_unit_hash": "51149c8c814def8804a0702c4e49d266c1cafa1343b015421b5a4e293a42c355", + "prepared_review_unit_hash": "48cb58b1df25447b25d967ba80fc36ec4ed4c1cd1fdcaff98a33700102be2c0c", "component_budget": { "node_count": 2, "item_count": 2, @@ -15781,18 +15684,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These are two separate 4-bit inference bugs: one in Gemma4 audio modules hitting `torch.finfo()` on `uint8` weights, and another in `BitsAndBytesConfig` where `llm_int8_skip_modules` drops the default `lm_head` exclusion and triggers an `AssertionError`.", + "summary": "The two issues both concern 4-bit inference regressions, but they describe different root causes and different failure paths: one is Gemma4/audio-module specific torch.finfo() on uint8 weights, the other is a BitsAndBytesConfig skip_modules regression that drops the default lm_head exclusion.", "confidence": 0.95, - "canonical_issue_reason": "Issue #45674 is the more general, configuration-specific regression with a clear failure mode and likely broader user impact.", + "canonical_issue_reason": "Issue 45674 is the better cluster representative because it describes a self-contained configuration regression with a clear trigger and failure mode; issue 45672 is a separate, model-specific Gemma4 bug.", "canonical_pr_reason": null, - "best_issue_reason": "Issue #45674 is the clearest and most actionable representative of the cluster, with a concrete repro path tied to a specific config interaction.", + "best_issue_reason": "Issue 45674 is the strongest standalone issue to keep because it is more general and actionable, while 45672 appears to be a narrower, unrelated model-specific defect.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:45672", "right": "issue:45674", "accept": false, - "reason": "Different bugs and code paths: Gemma4 audio-module dtype handling vs. BitsAndBytes skip-module exclusion logic. They both occur in 4-bit inference, but the failure mechanisms are unrelated." + "reason": "Reject: both mention 4-bit inference, but the symptoms and causes differ. 45672 is about torch.finfo() on uint8 weights in Gemma4 audio modules; 45674 is about BitsAndBytesConfig llm_int8_skip_modules clearing the default lm_head exclusion and triggering an AssertionError. These are not the same underlying bug." } ] }, @@ -15811,8 +15714,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-45762-2", "nodes": [ @@ -15822,7 +15725,7 @@ "soft_pairs": [ "issue:45762|issue:45763" ], - "prepared_review_unit_hash": "c2430bfebdc0033abdd26da99f9057562eeb68dc08c6234293e49755150b6481", + "prepared_review_unit_hash": "984d883ec9e367730dd9ec8978243c3136573b858c38d78562b05b793fd65f76", "component_budget": { "node_count": 2, "item_count": 2, @@ -15844,7 +15747,7 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These two closed issues appear unrelated; their similarity is only superficial (personal names plus dates), with no shared context or technical content.", + "summary": "The two closed issues are just person-name/date entries with no body or shared context, so they do not look like duplicates or the same underlying problem.", "confidence": 0.99, "canonical_issue_reason": null, "canonical_pr_reason": null, @@ -15855,7 +15758,7 @@ "left": "issue:45762", "right": "issue:45763", "accept": false, - "reason": "Different names and dates, no shared problem description, and no evidence they refer to the same underlying issue." + "reason": "Different names and dates, with no description, references, or shared context indicating the same underlying issue." } ] }, @@ -15874,8 +15777,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-45764-2", "nodes": [ @@ -15885,40 +15788,40 @@ "soft_pairs": [ "issue:45764|issue:45765" ], - "prepared_review_unit_hash": "2a07fe28106928ab5482f27604446a3b309f13fad86abb40dfc1e559409f7bf1", + "prepared_review_unit_hash": "55ee6551cd9d69cb5bc35db67c2536ee0d9ce188d65a7369315c1bf7ddb30caf", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 1480, - "estimated_input_tokens": 370, - "estimated_eval_tokens": 996 + "serialized_chars": 1481, + "estimated_input_tokens": 371, + "estimated_eval_tokens": 998 }, "budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 1480, - "estimated_input_tokens": 370, - "estimated_eval_tokens": 996 + "serialized_chars": 1481, + "estimated_input_tokens": 371, + "estimated_eval_tokens": 998 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "The two closed issues are not duplicates; they appear to be unrelated personal-name/date entries with no shared problem statement or actionable issue content.", + "summary": "These two closed issues are not duplicates or part of the same bug/change; they are unrelated title-only entries with different names and dates.", "confidence": 0.99, - "canonical_issue_reason": "No clear canonical issue: the titles refer to different people and different dates, with no evidence of the same underlying report.", + "canonical_issue_reason": null, "canonical_pr_reason": null, - "best_issue_reason": "Neither issue is a good global representative of a duplicate cluster because there is no real cluster here; if one must be picked, issue 45764 is only marginally simpler/earlier.", + "best_issue_reason": "No issue is a good canonical representative here; both are unrelated, body-less placeholder-style issues, so neither should be preferred as a cluster canonical.", "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "issue:45764", "right": "issue:45765", "accept": false, - "reason": "Reject: these are different named entries with different dates and no indication they describe the same bug or request." + "reason": "Different personal names and dates, no body or shared problem description; not the same underlying issue." } ] }, @@ -15980,178 +15883,178 @@ "pull_request:44827|pull_request:45770", "pull_request:43251|pull_request:43254" ], - "prepared_review_unit_hash": "9326002ff67b9751f2c06facb98e4abe8010b6dc7f8185dc2f78a5cc6d62875b", + "prepared_review_unit_hash": "25829ef5c0f193de131203f2bb21e58e34de0d326dc6ca44f01760b74bcb2194", "component_budget": { "node_count": 13, "item_count": 13, "soft_pair_count": 22, - "serialized_chars": 36419, - "estimated_input_tokens": 9105, - "estimated_eval_tokens": 18466 + "serialized_chars": 36416, + "estimated_input_tokens": 9104, + "estimated_eval_tokens": 18464 }, "budget": { "node_count": 13, "item_count": 13, "soft_pair_count": 22, - "serialized_chars": 36419, - "estimated_input_tokens": 9105, - "estimated_eval_tokens": 18466 + "serialized_chars": 36416, + "estimated_input_tokens": 9104, + "estimated_eval_tokens": 18464 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Only one soft edge looks like a true duplicate: the two `fixed_cross_entropy` PRs both implement the same kwargs/label_smoothing/weight pass-through on the same function and target the same issue. The rest are unrelated model additions, refactors, or test/documentation changes in different code paths.", - "confidence": 0.97, + "summary": "This cluster is mostly heterogeneous: model additions, auto/processing refactors, tests, and docs are mixed together. The only true duplicate-like pair is the two fixed_cross_entropy PRs (43251 and 43254); all other soft edges look unrelated and should stay separate.", + "confidence": 0.96, "canonical_issue_reason": null, - "canonical_pr_reason": "PR #43251 is the better canonical representative: it has the explicit issue link, broader/clearer wording of the same `cross_entropy` fix, and the same code path as #43254.", + "canonical_pr_reason": null, "best_issue_reason": null, - "best_pr_reason": "PR #43251 is the best global representative because it is the clearest, issue-linked version of the shared `fixed_cross_entropy` fix; the only other true duplicate (#43254) is a near-identical variant.", + "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "pull_request:39895", "right": "pull_request:43424", "accept": false, - "reason": "Unrelated PRs: VideoPrism model addition vs ExecuTorch dynamic-shape test coverage." + "reason": "VideoPrism model addition vs an Executorch dynamic-shape test; different code paths and goals." }, { "left": "pull_request:41855", "right": "pull_request:43823", "accept": false, - "reason": "Different changes: Mistral tokenizer method additions vs adding a new MobileLLM model." + "reason": "Mistral tokenizer helper methods vs adding a new MobileLLM model; unrelated areas." }, { "left": "pull_request:41855", "right": "pull_request:43424", "accept": false, - "reason": "Tokenizer API fix vs ExecuTorch export test; no shared underlying bug." + "reason": "Tokenizer API changes vs Executorch export test coverage; not the same bug or change." }, { "left": "pull_request:42668", "right": "pull_request:43823", "accept": false, - "reason": "Processor robustness refactor is unrelated to MobileLLM model implementation." + "reason": "Processor auto-loading robustness vs a new model implementation; separate problems." }, { "left": "pull_request:43251", "right": "pull_request:43823", "accept": false, - "reason": "`fixed_cross_entropy` kwargs bugfix is unrelated to adding MobileLLM." + "reason": "Cross-entropy kwargs fix vs MobileLLM model addition; no shared underlying change." }, { "left": "pull_request:43251", "right": "pull_request:44827", "accept": false, - "reason": "Different code paths: loss utility fix vs Mistral4 test/integration changes." + "reason": "Loss-function API fix vs Mistral4 test fixes; unrelated code paths." }, { "left": "pull_request:43424", "right": "pull_request:44827", "accept": false, - "reason": "ExecuTorch dynamic-shape test and Mistral4 test fixes are not the same underlying change." + "reason": "Executorch export test vs Mistral4 tests; both are tests, but not the same concrete issue." }, { "left": "pull_request:43424", "right": "pull_request:43823", "accept": false, - "reason": "ExecuTorch export test vs MobileLLM model addition; no duplicate bug/change." + "reason": "Executorch dynamic-shape export test vs model addition; unrelated." }, { "left": "pull_request:43823", "right": "pull_request:45770", "accept": false, - "reason": "New MobileLLM model vs AutoModel `text_config` unwrap logic; different functionality." + "reason": "New MobileLLM model files vs AutoModel text_config unwrapping; different subsystems." }, { "left": "pull_request:43823", "right": "pull_request:45655", "accept": false, - "reason": "MobileLLM model addition is unrelated to `cls.config` resolution order fix." + "reason": "MobileLLM model addition vs config resolution-order fix; not the same change." }, { "left": "pull_request:43823", "right": "pull_request:43995", "accept": false, - "reason": "New model implementation vs Falcon output-collection refactor; distinct changes." - }, - { - "left": "pull_request:43823", - "right": "pull_request:44827", - "accept": false, - "reason": "MobileLLM model addition and Mistral4 test fix do not address the same bug." + "reason": "New model implementation vs Falcon output-collection refactor; unrelated." }, { "left": "pull_request:43995", "right": "pull_request:45770", "accept": false, - "reason": "Falcon refactor and AutoModel `text_config` unwrap are different code paths." + "reason": "Falcon model refactor vs AutoFactory config unwrapping; different code paths." }, { "left": "pull_request:43995", "right": "pull_request:45213", "accept": false, - "reason": "Falcon model refactor is unrelated to the model-creation skill PR." + "reason": "Falcon refactor vs model-creation skill documentation; clearly unrelated." }, { "left": "pull_request:43995", "right": "pull_request:44827", "accept": false, - "reason": "Falcon output interface refactor vs Mistral4 test fix; no shared concrete fix." + "reason": "Falcon refactor vs Mistral4 test fixes; not the same bug or patch." }, { "left": "pull_request:44827", "right": "pull_request:45170", "accept": false, - "reason": "Mistral4 test/integration work is unrelated to the `layrnorm` spelling correction across CLIP-like models." + "reason": "Mistral4 test fixes vs a layernorm typo rename sweep; unrelated changes." }, { "left": "pull_request:44827", "right": "pull_request:45655", "accept": false, - "reason": "Mistral4 test fixes and `cls.config` resolution order are separate issues." + "reason": "Mistral4 tests vs cls.config resolution order fix; different issues." }, { "left": "pull_request:43823", "right": "pull_request:44827", "accept": false, - "reason": "MobileLLM model addition and Mistral4 test fix are different changes." + "reason": "MobileLLM model addition vs Mistral4 test fixes; unrelated." }, { "left": "pull_request:44827", "right": "pull_request:45213", "accept": false, - "reason": "Mistral4 test fixes are unrelated to the model-creation skill/meta PR." + "reason": "Mistral4 tests vs AI skill/docs for model creation; not a duplicate." }, { "left": "pull_request:43823", "right": "pull_request:45213", "accept": false, - "reason": "New model implementation vs meta skill/doc PR; not the same underlying change." + "reason": "Model implementation PR vs workflow skill/documentation PR; unrelated." }, { - "left": "pull_request:45655", + "left": "pull_request:43823", "right": "pull_request:45770", "accept": false, - "reason": "Both touch config handling, but one fixes `cls.config` resolution order while the other unwraps `text_config`; not the same concrete bug." + "reason": "MobileLLM model addition vs AutoFactory text_config unwrap; different areas." }, { - "left": "pull_request:44827", + "left": "pull_request:43995", + "right": "pull_request:45655", + "accept": false, + "reason": "Falcon refactor vs Python 3.14 config-resolution fix; separate concerns." + }, + { + "left": "pull_request:45655", "right": "pull_request:45770", "accept": false, - "reason": "Mistral4 test fixes are unrelated to AutoModel `text_config` unwrapping." + "reason": "Both touch config handling, but one fixes __init_subclass__ resolution order and the other unwraps text_config in AutoModelFor.from_config; not the same underlying bug." }, { "left": "pull_request:43251", "right": "pull_request:43254", "accept": true, - "reason": "Both PRs fix the same `fixed_cross_entropy` API gap in `src/transformers/loss/loss_utils.py` by passing supported kwargs (`weight`, `label_smoothing`) through to `nn.functional.cross_entropy`." + "reason": "Both PRs patch fixed_cross_entropy to forward supported nn.functional.cross_entropy kwargs (weight/label_smoothing) in the same function and code path; they are effectively the same fix." } ] }, "evaluator_result": { "accept": true, - "feedback": "The analyst output is well grounded in the packet and stays conservative on the soft edges. The only accepted duplicate pair, PR #43251 and #43254, is supported by the shared file, same function, and shared issue target 43240. The other verdicts are consistently limited to unrelated model additions, refactors, or tests, with no overstatement of duplicate evidence." + "feedback": "The summary is grounded in the packet: the cluster is heterogeneous, and the only clearly duplicate-like pair is 43251/43254, which both patch `fixed_cross_entropy` in the same file to forward the same supported `nn.functional.cross_entropy` kwargs. The other soft-edge rejections are conservative and do not overclaim a shared bug or mergeable single-PR fix." }, "evaluator_used": true, "retried": false, @@ -16178,7 +16081,7 @@ "soft_pairs": [ "pull_request:40546|pull_request:42781" ], - "prepared_review_unit_hash": "745c6dc5e0c79b23001ad329081e4a080b55a68f260b4138da889ead134beb79", + "prepared_review_unit_hash": "5aa00e3b995dedc48a6f9df50c5a085ee1f1dc814b444063a0e42ad708a976e3", "component_budget": { "node_count": 2, "item_count": 2, @@ -16200,18 +16103,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Two VibeVoice PRs overlap on the base model/docs/auto-registration scaffolding, but they target different deliverables: a full VibeVoice implementation vs. a later VibeVoice Realtime extension. They are related, not duplicates.", - "confidence": 0.93, + "summary": "These are related VibeVoice PRs, but they are not duplicates: PR 40546 is the base VibeVoice integration, while PR 42781 adds VibeVoice Realtime plus an acoustic tokenizer and other realtime-specific changes. The overlap is mostly model-family plumbing, not the same concrete fix/change.", + "confidence": 0.95, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 40546 is the broader foundational VibeVoice implementation and is the better representative of the base feature work. PR 42781 is a separate realtime follow-up with additional tokenizer/generation changes, not the same change set.", + "canonical_pr_reason": "PR 40546 is the better canonical PR because it is the main, broader VibeVoice implementation and has substantial review activity; PR 42781 is a separate realtime-focused extension with additional components and a different scope.", "best_issue_reason": null, - "best_pr_reason": "PR 40546 is the stronger cluster representative because it is the original, more complete implementation and is not marked draft. PR 42781 appears to be a distinct extension rather than a duplicate of the base VibeVoice PR.", + "best_pr_reason": "PR 40546 is the best representative of the cluster since it appears to be the foundational VibeVoice addition. PR 42781 is draft and targets a different feature set, so it should not be treated as the same PR.", "soft_edge_verdicts": [ { "left": "pull_request:40546", "right": "pull_request:42781", "accept": false, - "reason": "Both touch VibeVoice infrastructure, but 42781 adds realtime-specific components and extra generation/tokenizer work; it is a follow-up feature, not the same concrete fix/change as 40546." + "reason": "Shared VibeVoice files indicate the same model family, but the changes are not the same underlying code-path problem. 40546 implements VibeVoice generally; 42781 adds realtime support and an acoustic tokenizer, which is a separate extension and not plausibly mergeable as one duplicate PR." } ] }, @@ -16241,7 +16144,7 @@ "soft_pairs": [ "pull_request:40755|pull_request:41224" ], - "prepared_review_unit_hash": "71de2fd1b4c153e5fc1784e73f43a94f41e6a55f77db4c96906f17dc3915fb61", + "prepared_review_unit_hash": "29ac0ff1a0017db82c84920c2cfc934dcc5d17751ac02c8874fddd77931d5134", "component_budget": { "node_count": 2, "item_count": 2, @@ -16263,18 +16166,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These are two unrelated pull requests: one adds TimesFM covariate forecasting support, the other adds DINOv3 image-classification support. They share only a broad \u201cmodel feature addition\u201d similarity, not the same code path or fix.", - "confidence": 0.97, + "summary": "The two pull requests target completely different models and code paths: TimesFM covariate forecasting vs. DINOv3 image classification support. They should not be clustered as duplicates.", + "confidence": 0.98, "canonical_issue_reason": null, - "canonical_pr_reason": null, + "canonical_pr_reason": "[TimesFM] Add support for forecasting with covariates is the broader, more self-contained PR in this cluster and is the better representative if a single PR must be chosen.", "best_issue_reason": null, - "best_pr_reason": null, + "best_pr_reason": "[TimesFM] Add support for forecasting with covariates is the best overall representative because it has the larger, more substantive model change and clearer feature scope.", "soft_edge_verdicts": [ { - "left": "pull_request:40755", - "right": "pull_request:41224", + "left": "[TimesFM] Add support for forecasting with covariates", + "right": "Add DINOv3ViTForImageClassification support", "accept": false, - "reason": "Reject: they modify different models, different files, and different functionality. TimesFM covariate forecasting and DINOv3 classification support are separate feature additions, not the same underlying change or bug fix." + "reason": "Different model families, different tasks, and different implementation paths; they do not fix the same underlying bug or add the same feature." } ] }, @@ -16367,7 +16270,7 @@ "soft_pairs": [ "pull_request:41797|pull_request:45186" ], - "prepared_review_unit_hash": "aa039cd1adc9376a7cb012028955f19db45c1e3361eeb05bd6d5a40a65b87cc8", + "prepared_review_unit_hash": "dd632b20348ef0297b4280709455a2d6ca01c645f51e1ca897e30a61eae44643", "component_budget": { "node_count": 2, "item_count": 2, @@ -16389,18 +16292,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These are not duplicates: one PR adds DeepSeek OCR, the other adds Isaac. They share framework scaffolding for new model integrations, but the actual model code, docs, tests, and behavior are for different underlying models and code paths.", - "confidence": 0.98, + "summary": "These are two separate model-addition PRs (DeepSeek OCR vs Isaac) that share framework registration plumbing but do not fix the same code-path or bug.", + "confidence": 0.97, "canonical_issue_reason": null, - "canonical_pr_reason": "No canonical PR: the two pull requests implement different model additions (DeepSeek OCR vs Isaac), not the same fix/change.", + "canonical_pr_reason": null, "best_issue_reason": null, - "best_pr_reason": "Neither PR is a global best duplicate candidate because they are unrelated model-integration PRs rather than alternate solutions to the same problem.", + "best_pr_reason": "Neither PR is a duplicate of the other. If forced to pick one representative, 45186 is the cleaner model-addition PR, but it is not a canonical merge target for this cluster because the changes are unrelated.", "soft_edge_verdicts": [ { "left": "pull_request:41797", "right": "pull_request:45186", "accept": false, - "reason": "Shared boilerplate for model registration/docs is incidental; the PRs add different models with different configs, processors, modeling code, and tests, so they do not fix the same underlying bug/change." + "reason": "Different model families and separate integrations. The overlap is only in shared Transformers scaffolding (auto registries, docs, repo checks), not the same concrete bug or change, so they should not be merged as duplicates." } ] }, @@ -16430,7 +16333,7 @@ "soft_pairs": [ "pull_request:42975|pull_request:42976" ], - "prepared_review_unit_hash": "1b52a3839903aacf2d143dbf9bc7cc626d09c592886d9dc7a23c7e49cace8e78", + "prepared_review_unit_hash": "4134162635f963596c7da92e60ed426224de5a5084e030abfa8d2d12d1a4b273", "component_budget": { "node_count": 2, "item_count": 2, @@ -16452,18 +16355,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These are related only at a broad maintenance level: both update GitHub Actions versions, but they target different action families and different upgrade goals. They should not be deduplicated.", - "confidence": 0.88, + "summary": "Two workflow-maintenance PRs overlap in several GitHub Actions files, but they make different upgrades: PR 42975 is a broad Node 24 compatibility bump for core actions, while PR 42976 is a separate latest-version update centered on docker/build-push-action and related workflows.", + "confidence": 0.94, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 42975 is the broader and more specific workflow-maintenance change: it explicitly targets Node 24 compatibility and updates many GitHub Actions usages across workflows.", + "canonical_pr_reason": "PR 42975 is the clearest representative of the broader GitHub Actions compatibility upgrade; PR 42976 is a separate maintenance change focused on different action versions, so they are not duplicates.", "best_issue_reason": null, - "best_pr_reason": "PR 42975 is the best representative of the cluster because it covers the main GitHub Actions compatibility upgrade theme more comprehensively than PR 42976, which is a separate action-version bump focused on docker/build-push-action and a few related workflows.", + "best_pr_reason": "PR 42975 is the better representative for the cluster because it has a clear, specific rationale (Node 24 compatibility) and the broader workflow surface; PR 42976 is a distinct action-version bump rather than the same change.", "soft_edge_verdicts": [ { "left": "pull_request:42975", "right": "pull_request:42976", "accept": false, - "reason": "Both PRs update workflow action versions, but they change different actions and solve different upgrade needs. PR 42975 is about Node 24 compatibility for multiple actions; PR 42976 is about bumping docker/build-push-action and similar workflow dependencies. They are not the same concrete fix and would not plausibly be merged into one PR." + "reason": "Shared workflow filenames alone are not enough: 42975 updates actions for Node 24 compatibility, while 42976 upgrades docker/build-push-action and other actions to newer versions. Different concrete maintenance changes, so not the same underlying PR." } ] }, @@ -16502,76 +16405,76 @@ "pull_request:43451|pull_request:45490", "pull_request:45490|pull_request:45630" ], - "prepared_review_unit_hash": "82bc1568491a1571e6796e975027e0581ccf49ebd7fc949a99843cca377e319b", + "prepared_review_unit_hash": "5ef69bba952317c3a888a426673ef7d4d9f91a1826e2112882f73ebb15d3da71", "component_budget": { "node_count": 5, "item_count": 5, "soft_pair_count": 7, - "serialized_chars": 17617, - "estimated_input_tokens": 4405, - "estimated_eval_tokens": 9066 + "serialized_chars": 17616, + "estimated_input_tokens": 4404, + "estimated_eval_tokens": 9064 }, "budget": { "node_count": 5, "item_count": 5, "soft_pair_count": 7, - "serialized_chars": 17617, - "estimated_input_tokens": 4405, - "estimated_eval_tokens": 9066 + "serialized_chars": 17616, + "estimated_input_tokens": 4404, + "estimated_eval_tokens": 9064 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "All five items are separate model-addition pull requests for different models (ViT NEPA, Molmo, Molmo2, CTSM, Kimi2_6). They share only boilerplate integration files and docs patterns, not the same underlying change or code path.", + "summary": "These are separate model-addition PRs, not duplicates. They only overlap on shared boilerplate areas like docs and auto-registry updates, but each targets a different model/code path (ViT NEPA, Molmo, Molmo2, CTSM, Kimi2-6).", "confidence": 0.99, "canonical_issue_reason": null, - "canonical_pr_reason": "If one PR must represent the cluster, 43451 (Molmo2) is the most complete standalone integration, with the broadest implementation surface (modeling, processing, docs, tests, and auto-mapping updates).", + "canonical_pr_reason": null, "best_issue_reason": null, - "best_pr_reason": "43451 is the strongest representative for triage because it is the most substantive single integration; the others are distinct model ports and should not be collapsed into it.", + "best_pr_reason": null, "soft_edge_verdicts": [ { "left": "pull_request:42978", "right": "pull_request:43448", "accept": false, - "reason": "Different models and different change goals: ViT NEPA vs Molmo. Shared auto/docs wiring is boilerplate, not the same bug or feature." + "reason": "Different model additions (ViT NEPA vs Molmo); shared docs/auto wiring is boilerplate, not the same fix." }, { "left": "pull_request:42978", "right": "pull_request:43451", "accept": false, - "reason": "Different model integrations (ViT NEPA vs Molmo2). They touch some of the same framework files, but the concrete code paths and artifacts are unrelated." + "reason": "ViT NEPA and Molmo2 are distinct architectures with different files and tests; not the same underlying change." }, { "left": "pull_request:42978", "right": "pull_request:45490", "accept": false, - "reason": "ViT NEPA and CTSM are distinct model additions with no evidence of a shared underlying fix or mergeable single change." + "reason": "These add different models (ViT NEPA vs CTSM) and do not fix one shared bug or implement one shared feature." }, { "left": "pull_request:43448", "right": "pull_request:43451", "accept": false, - "reason": "Molmo and Molmo2 are separate model ports. Similar filenames only reflect shared registration/docs scaffolding, not duplicate work." + "reason": "Molmo and Molmo2 are separate model families; overlap is limited to common integration scaffolding." }, { "left": "pull_request:43448", "right": "pull_request:45490", "accept": false, - "reason": "Molmo and CTSM are unrelated integrations; they do not fix the same concrete problem or overlap beyond generic framework registration." + "reason": "Molmo and CTSM are unrelated model additions; similar registry/docs edits are not enough to merge." }, { "left": "pull_request:43451", "right": "pull_request:45490", "accept": false, - "reason": "Molmo2 and CTSM are different model additions with distinct implementations and tests, so they should not be merged as duplicates." + "reason": "Molmo2 and CTSM touch different model implementations and test suites, so they are not duplicates." }, { "left": "pull_request:45490", "right": "pull_request:45630", "accept": false, - "reason": "CTSM and Kimi2_6 are separate model additions; shared auto/modeling file edits are broad subsystem overlap, not the same code-path change." + "reason": "CTSM and Kimi2-6 are different model additions with different code paths; they should remain separate PRs." } ] }, @@ -16609,7 +16512,7 @@ "pull_request:43665|pull_request:45749", "pull_request:44815|pull_request:45747" ], - "prepared_review_unit_hash": "94fc491d0cc10075d0bbae640e0eb4346224bae0e8255efe481e77302e21284b", + "prepared_review_unit_hash": "71667fba30a134eafec6910aed568191e2ede3cd51942216d6011056408ec2eb", "component_budget": { "node_count": 5, "item_count": 5, @@ -16631,48 +16534,48 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These PRs are not duplicates: they span unrelated areas (Whisper decoding/tests, CLIP/ViT test edits, FP8 dequant/model-loading, continuous batching docs, and an unrelated split-batch-size fix). The only substantive code change is the FP8/dequant PR; the others look like test-only or cosmetic changes, or lack enough evidence to connect them to the same bug/change.", - "confidence": 0.93, + "summary": "These PRs are only superficially similar (mostly small fixes with generic titles). They do not appear to address the same underlying bug or change, so the soft links should be rejected. Among them, the FP8/dequant work is the strongest standalone code fix.", + "confidence": 0.9, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 44815 is the strongest representative because it makes the only clearly substantive production-code changes across multiple source files and includes a real bug-fix theme (FP8/dequant/model loading), unlike the others which are test-only, cosmetic, or too underspecified.", + "canonical_pr_reason": "PR #44815 is the most substantive, concrete code fix in the set: it changes loading and FP8 integration logic across multiple files, with tests, and has a clear bug-fix scope compared with the others.", "best_issue_reason": null, - "best_pr_reason": "PR 44815 is the best standalone PR in the set: it has concrete code-path changes, touches relevant runtime files, and appears to address an actual functional issue rather than test noise or a spelling/doc cleanup.", + "best_pr_reason": "PR #44815 is the best representative of the cluster because it has the clearest technical problem statement and the most meaningful implementation, while the other PRs are either test-only, cosmetic, or too underspecified.", "soft_edge_verdicts": [ { "left": "pull_request:43291", "right": "pull_request:43665", "accept": false, - "reason": "Both are labeled as fixes, but they target different models and different kinds of changes: Whisper tokenizer/runtime behavior versus CLIP/ViT test edits. No shared code path or same underlying bug is evident." + "reason": "Different subsystems and different behavior: Whisper tokenizer decoding vs CLIP/ViT tests. Similarity is only in being small fixes, not the same bug." }, { "left": "pull_request:43291", "right": "pull_request:44815", "accept": false, - "reason": "Whisper decoding/tokenization changes are unrelated to FP8 dequant/model-loading fixes. Different subsystems, different failure modes." + "reason": "Unrelated changes: Whisper decoding logic/tests versus FP8 dequant/loading code. No shared code path or concrete defect." }, { "left": "pull_request:43665", "right": "pull_request:44815", "accept": false, - "reason": "This is test-only CLIP/ViT churn versus a concrete FP8/runtime bug fix. They do not look like the same underlying change and would not plausibly merge as one PR." + "reason": "CLIP/ViT test edits are not the same underlying issue as FP8 loading/dequant changes. Shared 'fix' wording is too weak to merge." }, { "left": "pull_request:43665", "right": "pull_request:45747", "accept": false, - "reason": "The CLIP/ViT test edits are unrelated to a split-batch-size fix; no shared code path or concrete bug is visible." + "reason": "No evidence of a shared bug or code path; one is test-file changes, the other is an unspecified batch-size fix." }, { "left": "pull_request:43665", "right": "pull_request:45749", "accept": false, - "reason": "Test-file edits in vision models and a docstring spelling correction in continuous batching are clearly separate cosmetic changes." + "reason": "Test edits in vision models and a docstring spelling correction are clearly unrelated changes." }, { "left": "pull_request:44815", "right": "pull_request:45747", "accept": false, - "reason": "FP8 dequant/model-loading work is unrelated to an unspecified split-batch-size PR; there is not enough evidence of the same bug or change." + "reason": "FP8 loading/dequant changes and a split-batch-size fix are different problems with no concrete overlap." } ] }, @@ -16702,12 +16605,12 @@ "soft_pairs": [ "pull_request:43488|pull_request:43532" ], - "prepared_review_unit_hash": "237bb86038e7c63be626a66b75af36fe9c836a44964ecd16a4db74503714532b", + "prepared_review_unit_hash": "bc87674bcfacdf3a6004bbc7ad3e7a28853a54b7424532c30efc671742a872c0", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 4562, + "serialized_chars": 4563, "estimated_input_tokens": 1141, "estimated_eval_tokens": 2538 }, @@ -16715,7 +16618,7 @@ "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 4562, + "serialized_chars": 4563, "estimated_input_tokens": 1141, "estimated_eval_tokens": 2538 }, @@ -16724,18 +16627,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These two open PRs are not duplicates: one is a setup/version-bump + test dependency change, while the other is a workflow/testing-utils change around CI/comment handling. They only overlap on a shared Vit test file, which is not enough to indicate the same underlying fix.", - "confidence": 0.96, + "summary": "These two PRs are not duplicates: #43488 is a small bot-format/versioning test change around setup.py and Vit test scaffolding, while #43532 rewires several GitHub workflows plus helper/notification code. The shared Vit test filename is incidental, not evidence of the same underlying fix.", + "confidence": 0.97, "canonical_issue_reason": null, - "canonical_pr_reason": "PR #43532 is the broader and more representative change in this pair, but it is not a duplicate of #43488; the two PRs address different code paths and should not be merged into one canonical change.", + "canonical_pr_reason": "No canonical PR: the PRs target different code paths and different maintenance purposes, so they should not be merged into one duplicate cluster.", "best_issue_reason": null, - "best_pr_reason": "PR #43532 is the better standalone representative of the cluster because it contains the more substantial multi-file behavior change, whereas #43488 looks like a narrow bot/version-format check.", + "best_pr_reason": "If one PR must represent the cluster, #43532 is the broader and more substantive change, but it is still not a duplicate of #43488.", "soft_edge_verdicts": [ { "left": "pull_request:43488", "right": "pull_request:43532", "accept": false, - "reason": "Reject: the diffs touch different concerns. #43488 updates setup/versioning and test dependencies, while #43532 rewires GitHub workflows and testing utilities. A shared test filename alone does not mean they fix the same concrete bug or should be merged together." + "reason": "Different underlying changes: #43488 is a setup/versioning and test-bot check, while #43532 changes CI/workflow behavior and notification utilities. They do not fix the same concrete problem and would not be merged as one PR." } ] }, @@ -16765,7 +16668,7 @@ "soft_pairs": [ "pull_request:43612|pull_request:43613" ], - "prepared_review_unit_hash": "9eb3f17e054b4dc20d017e73b3ab71edafbb818b76396ba5a1f9d957eec25c1f", + "prepared_review_unit_hash": "650d45821801a3666ffaa3cfe5b0bb77aad4f96932887409a164d4257e43d21b", "component_budget": { "node_count": 2, "item_count": 2, @@ -16787,18 +16690,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Two separate PRs add adjacent but distinct pipeline features: one for Promptable Concept Segmentation and one for Promptable Visual Segmentation. They share scaffolding and docs plumbing, but they target different tasks and model integrations, so they should not be deduplicated.", - "confidence": 0.95, + "summary": "Two open PRs add similarly named but distinct pipelines: one for promptable concept segmentation and one for promptable visual segmentation. They share repository scaffolding, but they target different feature sets and model integrations, so they should not be treated as duplicates.", + "confidence": 0.96, "canonical_issue_reason": null, - "canonical_pr_reason": "No single canonical PR: these are parallel feature additions for different segmentation tasks, not the same underlying change.", + "canonical_pr_reason": "No canonical PR: the two PRs implement different pipelines with different model backends and task docs, not the same underlying change.", "best_issue_reason": null, - "best_pr_reason": "Neither PR is a true duplicate of the other. If forced to pick a representative, PR #43613 is slightly broader in model/test coverage, but it still represents a different feature than PR #43612.", + "best_pr_reason": "No single best PR for deduplication; each PR is the primary candidate for its own distinct feature area.", "soft_edge_verdicts": [ { "left": "pull_request:43612", "right": "pull_request:43613", "accept": false, - "reason": "Both add segmentation-related pipelines and share some docs/metadata files, but they implement different concrete features with different pipeline names, model backends, and test suites. They are related, not the same underlying change." + "reason": "Reject: these are separate pipeline additions. One introduces Promptable Concept Segmentation (SAM3-focused), the other Promptable Visual Segmentation (SAM/SAM2/EdgeTam/tracker-focused). Shared docs/registry edits do not make them the same code-path fix or something that could be merged into one PR." } ] }, @@ -16828,7 +16731,7 @@ "soft_pairs": [ "pull_request:43651|pull_request:43663" ], - "prepared_review_unit_hash": "67fee8ffee0ab63772978e41931e57141d029424ac932c4f6dd74874efe22a6e", + "prepared_review_unit_hash": "e69588df2d965c49574fbf07583655a432c49b0a7145660ae1d99720ff7458ba", "component_budget": { "node_count": 2, "item_count": 2, @@ -16850,18 +16753,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "The two open PRs both touch trainer.py and add extension hooks, but they address different behaviors: gradient-accumulation loss scaling vs. signature-column filtering. They should not be treated as duplicates.", - "confidence": 0.97, + "summary": "Both items are open PRs in `trainer.py`, but they address different Trainer extensibility points: one changes gradient-accumulation loss scaling, the other adds an override hook for signature-column filtering. They are not the same fix and should not be clustered as duplicates.", + "confidence": 0.96, "canonical_issue_reason": null, - "canonical_pr_reason": "No single canonical PR: 43651 adds a GA-loss-scaling hook, while 43663 adds a signature-column override hook; they solve separate trainer extensibility problems.", + "canonical_pr_reason": "No single canonical PR stands out: `#43651` and `#43663` modify different code paths and solve unrelated Trainer customization problems.", "best_issue_reason": null, - "best_pr_reason": "If a representative is needed, 43651 is slightly more central because it changes the core training-step loss-scaling path, but it is still not a substitute for 43663.", + "best_pr_reason": "Neither PR is a good global representative for the other; each targets a separate behavior change in `Trainer` rather than a shared underlying bug.", "soft_edge_verdicts": [ { "left": "pull_request:43651", "right": "pull_request:43663", "accept": false, - "reason": "Different underlying changes and code paths: 43651 controls loss normalization during gradient accumulation, while 43663 exposes signature-column filtering for model inputs. They are related only by file/subsystem, not by the same bug or fix." + "reason": "Reject: same file, but different functionality and code paths. `#43651` is about loss normalization under gradient accumulation; `#43663` is about extracting signature columns for filtering. They are not plausibly merged as one fix." } ] }, @@ -16893,7 +16796,7 @@ "pull_request:43656|pull_request:43842", "pull_request:43836|pull_request:43842" ], - "prepared_review_unit_hash": "27686481e71678a329faa06c384b48a860f2015dc3f059bcc1987fa7c8a071be", + "prepared_review_unit_hash": "bc99fb03682d89a03d7b4e6e342bd0d78c4215353840669a655441578aacfafc", "component_budget": { "node_count": 3, "item_count": 3, @@ -16915,30 +16818,30 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "All three PRs target the same CLI serve.py TypeAdapter annotation NameError class of bug; 43842 is the most focused fix, while 43656 is a similar variant and 43836 appears to include the same fix alongside broader cleanup/refactor work.", - "confidence": 0.81, + "summary": "All three PRs target the same `src/transformers/cli/serve.py` runtime annotation/TypeAdapter NameError in the CLI server. PR 43842 is the cleanest, most targeted fix; the other two are broader or alternate annotation-workaround variants.", + "confidence": 0.86, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 43842 is the clearest, most narrowly scoped fix for the concrete TypeAdapter NameError when pydantic is absent; it directly addresses the runtime annotation problem in serve.py with minimal unrelated churn.", + "canonical_pr_reason": "PR 43842 is the most precise fix: it guards `TypeAdapter` behind `TYPE_CHECKING` and quotes the affected annotation, directly addressing the reported pydantic-missing NameError with minimal scope.", "best_issue_reason": null, - "best_pr_reason": "PR 43842 best represents the cluster because it fixes the exact NameError path with the smallest, most explicit change set and is easiest to reason about as the canonical duplicate target.", + "best_pr_reason": "PR 43842 best matches the concrete crash path and has the smallest, clearest patch while still fixing the underlying runtime NameError.", "soft_edge_verdicts": [ { "left": "pull_request:43656", "right": "pull_request:43842", "accept": true, - "reason": "Both PRs fix the same runtime TypeAdapter annotation NameError in src/transformers/cli/serve.py; the edits are different implementations of the same underlying bug and could plausibly be merged into one fix." + "reason": "Both are fixes for the same CLI server annotation NameError around `TypeAdapter`/forward references in `serve.py`; they address the same underlying runtime problem and could be merged into one PR." }, { "left": "pull_request:43836", "right": "pull_request:43842", "accept": true, - "reason": "Despite 43836 showing broader file churn, its title and touched file indicate the same TypeAdapter string-literal/annotation NameError fix in the same code path, so it is close enough to treat as the same underlying change." + "reason": "Both patch the same `serve.py` `TypeAdapter` NameError path with annotation-only changes; they are alternative implementations of the same bug fix and could plausibly be combined." } ] }, "evaluator_result": { "accept": true, - "feedback": "The cluster summary is adequately grounded in the packet: all three PRs touch the same file and are tied by titles/diffs to TypeAdapter/annotation NameError fixes in `serve.py`, with 43836 clearly broader than 43842. The soft-edge accepts are conservative enough, especially since 43836 and 43842 share an issue target and 43656 is at least plausibly the same annotation-related bugfix family." + "feedback": "The summary is grounded in the packet: all three PRs touch `src/transformers/cli/serve.py` and are aimed at preventing annotation-related `NameError`s, with PR 43842 being the narrowest fix. The soft-edge pair verdicts are also conservative enough: 43836 and 43842 clearly address the same `TypeAdapter` annotation issue, and 43656 is plausibly another implementation of the same underlying forward-reference/runtime annotation problem in the same file." }, "evaluator_used": true, "retried": false, @@ -17103,8 +17006,8 @@ "pull_request:43997|pull_request:44056", "pull_request:44071|pull_request:44076", "pull_request:44074|pull_request:44101", - "pull_request:44056|pull_request:44161", "pull_request:43997|pull_request:44018", + "pull_request:44056|pull_request:44161", "pull_request:43997|pull_request:44161", "pull_request:44074|pull_request:44161", "pull_request:43997|pull_request:44071", @@ -17116,7 +17019,7 @@ "pull_request:44010|pull_request:44018", "pull_request:44071|pull_request:44129", "pull_request:43996|pull_request:44161", - "pull_request:43996|pull_request:44017", + "pull_request:43996|pull_request:44074", "pull_request:44010|pull_request:44044", "pull_request:44044|pull_request:44074", "pull_request:44066|pull_request:44722", @@ -17128,385 +17031,343 @@ "pull_request:44044|pull_request:44154", "pull_request:44024|pull_request:44076" ], - "prepared_review_unit_hash": "f3eb77e844e532f8eef743d71efd201a378d2630cb18b6cfdc66cda45021a458", + "prepared_review_unit_hash": "7ae479abfc365a6763527dbfc1f0619b7ea1781d5a75fa633416ca225bff432b", "component_budget": { "node_count": 28, "item_count": 28, "soft_pair_count": 77, - "serialized_chars": 84498, - "estimated_input_tokens": 21125, - "estimated_eval_tokens": 42506 + "serialized_chars": 84521, + "estimated_input_tokens": 21131, + "estimated_eval_tokens": 42518 }, "budget": { "node_count": 28, "item_count": 28, "soft_pair_count": 72, - "serialized_chars": 76931, - "estimated_input_tokens": 19233, - "estimated_eval_tokens": 38722 + "serialized_chars": 76955, + "estimated_input_tokens": 19239, + "estimated_eval_tokens": 38734 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "The packet contains two separate duplicate subclusters, not one: a GPT-J/CodeGen family centered on 44722, and a distinct GPT-Neo pair around 44018/44068. Most other PRs are model-specific output-tracing refactors and should stay separate.", - "confidence": 0.82, + "summary": "This cluster is mostly a set of independent model-specific output-tracing refactors tied to the same tracking issue, not one duplicate bug. The only strong duplicate/superseding pairs are within the same model family and concrete code path (notably GPT-Neo, and GPT-J/CodeGen).", + "confidence": 0.9, "canonical_issue_reason": null, - "canonical_pr_reason": "44722 is the best canonical PR for the GPT-J/CodeGen subcluster because it spans both model files, explicitly targets the tracking issue, and is the broadest overlapping implementation. The GPT-Neo duplicates form a separate family and should not be folded into this canonical choice.", + "canonical_pr_reason": "44722 is the best canonical PR for the cluster\u2019s main theme: it is a later, concrete GPT-J/CodeGen output-tracing refactor with the same issue target and the clearest, most complete code-path change among the overlapping GPT-J work.", "best_issue_reason": null, - "best_pr_reason": "44722 is the strongest single PR in the packet, but only for the GPT-J/CodeGen family; 44068 is the better representative for the separate GPT-Neo family.", + "best_pr_reason": "44722 is the strongest representative PR: it has explicit issue linkage, touches the actual implementation files, and most clearly reflects the standardized output-tracing refactor pattern seen across the cluster.", "soft_edge_verdicts": [ { "left": "pull_request:43996", "right": "pull_request:44085", "accept": false, - "reason": "Different model families and code paths; same output-tracing theme is too broad." + "reason": "Different models/code paths (CVT/FNet vs GPT-J/CodeGen); same tracking issue is not enough." }, { "left": "pull_request:43996", "right": "pull_request:44044", "accept": false, - "reason": "Different backbones and implementations; not the same concrete change." + "reason": "Different model families and different implementation paths; only the output-tracing theme overlaps." }, { "left": "pull_request:44066", "right": "pull_request:44085", "accept": true, - "reason": "Both are GPT-J output-tracing refactors on the same code path and could plausibly be merged into one PR." + "reason": "Same GPT-J output-tracing refactor in the same file family; one is effectively a variant/superset of the other." }, { "left": "pull_request:44007", "right": "pull_request:44072", "accept": false, - "reason": "RegNet/ResNet-style tracing work vs EfficientNet are unrelated model paths." + "reason": "ResNet vs EfficientNet are separate model implementations, so not the same underlying change." }, { "left": "pull_request:44072", "right": "pull_request:44722", "accept": false, - "reason": "EfficientNet vs GPT-J/CodeGen are unrelated implementations." + "reason": "Different models and code paths; shared issue target alone does not make them duplicates." }, { "left": "pull_request:44066", "right": "pull_request:44072", "accept": false, - "reason": "GPT-J/CodeGen and EfficientNet are different code paths." + "reason": "GPT-J/CodeGen refactor vs EfficientNet refactor; different underlying change." }, { "left": "pull_request:44013", "right": "pull_request:44044", "accept": false, - "reason": "MobileNetV2 and DeBERTa are unrelated models." + "reason": "MobileNetV2 vs DeBERTa V2 are unrelated implementations." }, { "left": "pull_request:44066", "right": "pull_request:44086", "accept": false, - "reason": "GPT-J and MGP-STR are unrelated model code paths." + "reason": "GPT-J/CodeGen output tracing vs MGP-STR output tracing; different code paths." }, { "left": "pull_request:44018", "right": "pull_request:44068", "accept": true, - "reason": "Same GPT-Neo output-tracing refactor on the same file; these look like duplicate/superset variants." + "reason": "Both are GPT-Neo output-tracing refactors over the same file and tests; 44068 looks like a later superseding version." }, { "left": "pull_request:44066", "right": "pull_request:44071", "accept": false, - "reason": "GPT-J/CodeGen and MPT are different models; shared decorator migration is not enough." + "reason": "GPT-J/CodeGen vs MPT; separate model implementations." }, { "left": "pull_request:44066", "right": "pull_request:44068", "accept": false, - "reason": "GPT-J/CodeGen and GPT-Neo are separate model families." + "reason": "Different models (GPT-J/CodeGen vs GPT-Neo), so not the same concrete change." }, { "left": "pull_request:44018", "right": "pull_request:44066", "accept": false, - "reason": "GPT-Neo and GPT-J/CodeGen are different concrete code paths." + "reason": "Different model families; only the output-tracing migration pattern is shared." }, { "left": "pull_request:44019", "right": "pull_request:44722", "accept": false, - "reason": "ResNet and GPT-J/CodeGen are unrelated changes." + "reason": "ResNet output tracing is unrelated to GPT-J/CodeGen output tracing." }, { "left": "pull_request:44086", "right": "pull_request:44722", "accept": false, - "reason": "MGP-STR and GPT-J/CodeGen are unrelated models." + "reason": "MGP-STR vs GPT-J/CodeGen are different model-specific changes." }, { "left": "pull_request:44019", "right": "pull_request:44085", "accept": false, - "reason": "ResNet and GPT-J are unrelated implementations." + "reason": "ResNet vs GPT-J; separate code paths." }, { "left": "pull_request:44018", "right": "pull_request:44019", "accept": false, - "reason": "GPT-Neo and ResNet are unrelated models." + "reason": "GPT-Neo and ResNet are unrelated model implementations." }, { "left": "pull_request:44019", "right": "pull_request:44071", "accept": false, - "reason": "ResNet and MPT are unrelated code paths." + "reason": "ResNet vs MPT; different underlying refactors." }, { "left": "pull_request:44068", "right": "pull_request:44116", "accept": false, - "reason": "GPT-Neo and Flaubert/XLM are unrelated models." + "reason": "GPT-Neo vs Flaubert/XLM output tracing are different model changes." }, { "left": "pull_request:44068", "right": "pull_request:44722", "accept": false, - "reason": "GPT-Neo is a separate duplicate family from GPT-J/CodeGen." + "reason": "GPT-Neo is a different model/code path from GPT-J/CodeGen." }, { "left": "pull_request:44018", "right": "pull_request:44116", "accept": false, - "reason": "GPT-Neo and Flaubert/XLM are unrelated." + "reason": "GPT-Neo vs Flaubert/XLM are unrelated implementations." }, { "left": "pull_request:44071", "right": "pull_request:44722", "accept": false, - "reason": "MPT and GPT-J/CodeGen are unrelated models." + "reason": "MPT vs GPT-J/CodeGen are different concrete changes." }, { "left": "pull_request:44018", "right": "pull_request:44086", "accept": false, - "reason": "GPT-Neo and MGP-STR are unrelated." + "reason": "GPT-Neo vs MGP-STR; different model code paths." }, { "left": "pull_request:44018", "right": "pull_request:44085", "accept": false, - "reason": "GPT-Neo and GPT-J are different code paths." + "reason": "GPT-Neo vs GPT-J/CodeGen are not the same underlying change." }, { "left": "pull_request:44068", "right": "pull_request:44086", "accept": false, - "reason": "GPT-Neo and MGP-STR are unrelated models." - }, - { - "left": "pull_request:44071", - "right": "pull_request:44085", - "accept": false, - "reason": "MPT and GPT-J are unrelated." + "reason": "GPT-Neo and MGP-STR are different model-specific refactors." }, { "left": "pull_request:44059", "right": "pull_request:44086", "accept": false, - "reason": "GPT-2 and MGP-STR are unrelated." + "reason": "GPT-2 vs MGP-STR; same output-tracing theme but different code paths." }, { "left": "pull_request:44085", "right": "pull_request:44086", "accept": false, - "reason": "GPT-J and MGP-STR are unrelated model paths." + "reason": "GPT-J/CodeGen vs MGP-STR are unrelated model changes." }, { "left": "pull_request:44098", "right": "pull_request:44101", "accept": false, - "reason": "ViLT and Flaubert/XLM are unrelated." + "reason": "ViLT vs Flaubert/XLM are different model families." }, { "left": "pull_request:44059", "right": "pull_request:44722", "accept": false, - "reason": "GPT-2 and GPT-J/CodeGen are unrelated models." - }, - { - "left": "pull_request:44018", - "right": "pull_request:44071", - "accept": false, - "reason": "GPT-Neo and MPT are unrelated implementations." - }, - { - "left": "pull_request:44017", - "right": "pull_request:44161", - "accept": false, - "reason": "SegFormer and LongT5 are unrelated models." + "reason": "GPT-2 refactor is separate from GPT-J/CodeGen." }, { - "left": "pull_request:44073", - "right": "pull_request:44101", - "accept": false, - "reason": "VisualBert and Flaubert/XLM are different code paths." - }, - { - "left": "pull_request:44059", - "right": "pull_request:44071", - "accept": false, - "reason": "GPT-2 and MPT are unrelated models." - }, - { - "left": "pull_request:44056", - "right": "pull_request:44101", - "accept": false, - "reason": "MPNet and Flaubert/XLM are unrelated." - }, - { - "left": "pull_request:44073", - "right": "pull_request:44161", + "left": "pull_request:44071", + "right": "pull_request:44085", "accept": false, - "reason": "VisualBert and LongT5 are unrelated." + "reason": "MPT vs GPT-J; different underlying changes." }, { "left": "pull_request:43997", "right": "pull_request:44017", "accept": false, - "reason": "RegNet and SegFormer are different backbones." + "reason": "RegNet vs Segformer are unrelated model implementations." }, { "left": "pull_request:43997", - "right": "pull_request:44056", + "right": "44056", "accept": false, - "reason": "RegNet and MPNet are unrelated." + "reason": "RegNet vs MPNet are separate code paths." }, { "left": "pull_request:44071", "right": "pull_request:44076", "accept": false, - "reason": "MPT and ImageGPT are unrelated model implementations." + "reason": "MPT vs ImageGPT are different models; shared output-tracing pattern is too broad." }, { "left": "pull_request:44074", "right": "pull_request:44101", "accept": false, - "reason": "TextNet and Flaubert/XLM are unrelated." - }, - { - "left": "pull_request:44056", - "right": "pull_request:44161", - "accept": false, - "reason": "MPNet and LongT5 are unrelated." + "reason": "TextNet vs Flaubert/XLM are unrelated changes." }, { "left": "pull_request:43997", "right": "pull_request:44018", "accept": false, - "reason": "RegNet and GPT-Neo are unrelated model code paths." - }, - { - "left": "pull_request:43997", - "right": "pull_request:44161", - "accept": false, - "reason": "RegNet and LongT5 are unrelated." + "reason": "RegNet vs GPT-Neo are different model-specific refactors." }, { "left": "pull_request:44071", "right": "pull_request:44129", "accept": false, - "reason": "MPT and SpeechT5 are unrelated." + "reason": "MPT vs SpeechT5 are separate code paths." }, { - "left": "pull_request:43997", - "right": "pull_request:44071", + "left": "pull_request:43996", + "right": "pull_request:44161", "accept": false, - "reason": "RegNet and MPT are unrelated." + "reason": "CVT/FNet vs LongT5 are unrelated model changes." }, { - "left": "pull_request:44010", - "right": "pull_request:44071", + "left": "pull_request:43996", + "right": "pull_request:44074", "accept": false, - "reason": "SqueezeBERT and MPT are unrelated." + "reason": "CVT/FNet vs TextNet are different underlying changes." }, { "left": "pull_request:44010", - "right": "pull_request:44073", + "right": "pull_request:44044", "accept": false, - "reason": "SqueezeBERT and VisualBert are unrelated." + "reason": "SqueezeBert vs DeBERTa V2 are separate model implementations." }, { - "left": "pull_request:44018", - "right": "pull_request:44129", + "left": "pull_request:44044", + "right": "pull_request:44074", "accept": false, - "reason": "GPT-Neo and SpeechT5 are unrelated." + "reason": "DeBERTa V2 vs TextNet are unrelated changes." }, { - "left": "pull_request:44074", - "right": "pull_request:44161", - "accept": false, - "reason": "TextNet and LongT5 are unrelated." + "left": "pull_request:44066", + "right": "pull_request:44722", + "accept": true, + "reason": "Same GPT-J/CodeGen output-tracing refactor family; 44722 is a broader later version of the same change." }, { - "left": "pull_request:43997", - "right": "pull_request:44074", - "accept": false, - "reason": "RegNet and TextNet are unrelated." + "left": "pull_request:44085", + "right": "pull_request:44722", + "accept": true, + "reason": "Both are GPT-J output-tracing refactors on the same code path; 44722 appears to supersede/extend the earlier PR." }, { - "left": "pull_request:43997", + "left": "pull_request:44085", "right": "pull_request:44129", "accept": false, - "reason": "RegNet and SpeechT5 are unrelated." + "reason": "GPT-J vs SpeechT5 are unrelated model changes." }, { - "left": "pull_request:44010", - "right": "pull_request:44018", + "left": "pull_request:43996", + "right": "pull_request:44007", "accept": false, - "reason": "SqueezeBERT and GPT-Neo are unrelated." + "reason": "CVT/FNet vs ResNet are different model-specific refactors." }, { - "left": "pull_request:44071", + "left": "pull_request:43996", "right": "pull_request:44129", "accept": false, - "reason": "MPT and SpeechT5 are unrelated." + "reason": "CVT/FNet vs SpeechT5 are different underlying changes." }, { - "left": "pull_request:43996", - "right": "pull_request:44161", + "left": "pull_request:44044", + "right": "pull_request:44129", "accept": false, - "reason": "CVT/FNet and LongT5 are unrelated." + "reason": "DeBERTa V2 vs SpeechT5 are unrelated model changes." }, { - "left": "pull_request:43996", - "right": "pull_request:44017", + "left": "pull_request:44044", + "right": "pull_request:44154", "accept": false, - "reason": "CVT/FNet and SegFormer are different backbones." + "reason": "DeBERTa V2 vs Vits are separate model-specific refactors." }, { - "left": "pull_request:44010", - "right": "pull_request:44044", + "left": "pull_request:44024", + "right": "pull_request:44076", "accept": false, - "reason": "SqueezeBERT and DeBERTa are unrelated models." + "reason": "FocalNet vs ImageGPT are different models; same refactor style does not make them duplicates." }, { - "left": "pull_request:44044", - "right": "pull_request:44074", + "left": "pull_request:44017", + "right": "pull_request:44161", "accept": false, - "reason": "DeBERTa and TextNet are unrelated." + "reason": "Segformer vs LongT5 are unrelated code paths." }, { - "left": "pull_request:44024", - "right": "pull_request:44076", + "left": "pull_request:44019", + "right": "pull_request:44101", "accept": false, - "reason": "FocalNet and ImageGPT are unrelated model implementations." + "reason": "ResNet vs Flaubert/XLM are different model implementations." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44068", + "accept": true, + "reason": "Same GPT-Neo output-tracing refactor with the same file/test surface; likely a later duplicate/superseding PR." } ] }, "evaluator_result": { - "accept": false, - "feedback": "The pair-level verdicts are mostly conservative, but the cluster summary overstates the evidence. The packet supports two separate duplicate subclusters: GPT-J/CodeGen (44066, 44085, 44722) and GPT-Neo (44018, 44068). Saying there is \u201conly real duplicate family\u201d and collapsing GPT-J/GPT-Neo into one family is not grounded. 44722 is a plausible canonical PR for the GPT-J/CodeGen subcluster, but the overall summary should acknowledge the second GPT-Neo duplicate pair as well." + "accept": true, + "feedback": "The summary is grounded in the packet, and the conservative theme distinction is appropriate: most PRs are separate model-specific output-tracing refactors sharing only the tracking issue. The accepted soft-edge pairs are limited to same-family, same-code-path overlaps (GPT-Neo and GPT-J/CodeGen), which do look mergeable as one bugfix PR. I did not see any clear overstatement of evidence." }, "evaluator_used": true, - "retried": true, + "retried": false, "accepted_nontrivial_soft_edge": true, "error_kind": null, "error_message": null @@ -17538,20 +17399,20 @@ "pull_request:43997|pull_request:44010", "pull_request:44010|pull_request:44129" ], - "prepared_review_unit_hash": "6a2ba859ba3645c59dd92b1424f0673e181fb14fc8ce0774852f9dfad307c8a8", + "prepared_review_unit_hash": "52fdf146c1522d186152b3265164bfc207c5c6a49734db5fae454e7c6b8a1ab2", "component_budget": { "node_count": 28, "item_count": 28, "soft_pair_count": 77, - "serialized_chars": 84498, - "estimated_input_tokens": 21125, - "estimated_eval_tokens": 42506 + "serialized_chars": 84521, + "estimated_input_tokens": 21131, + "estimated_eval_tokens": 42518 }, "budget": { "node_count": 6, "item_count": 6, "soft_pair_count": 5, - "serialized_chars": 13067, + "serialized_chars": 13066, "estimated_input_tokens": 3267, "estimated_eval_tokens": 6790 }, @@ -17560,42 +17421,42 @@ "aggressively_trimmed": false, "split": true, "analyst_result": { - "summary": "All items are model-specific PRs that migrate different architectures to the same output-tracing/capture pattern, but they are not duplicates of one another; they touch different code paths and would not be merged into a single PR.", - "confidence": 0.9, + "summary": "A cluster of similar-sounding but model-specific refactors to migrate output tracing/capture across different Transformer implementations. They share a pattern, not the same concrete bug or change.", + "confidence": 0.69, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 44129 is the strongest representative because it has an explicit issue target and a clearly scoped refactor to standardized output capture.", + "canonical_pr_reason": "PR 44129 is the best representative: it has an explicit issue link, a clear title, and a substantive output-capture refactor in one model.", "best_issue_reason": null, - "best_pr_reason": "PR 44129 best captures the cluster theme while remaining a concrete, self-contained change.", + "best_pr_reason": "PR 44129 is the strongest single PR to represent the cluster because it is the most concretely scoped and best tracked, but it is still model-specific rather than duplicate of the others.", "soft_edge_verdicts": [ { "left": "pull_request:44073", "right": "pull_request:44074", "accept": false, - "reason": "Same refactor pattern, but different models and files; these are separate model-specific changes, not one concrete bug fix." + "reason": "Both are output-tracing refactors, but for different models (VisualBert vs TextNet) and different code paths; not the same change." }, { "left": "pull_request:44056", "right": "pull_request:44073", "accept": false, - "reason": "Both standardize output tracing, but MPNet and VisualBert have distinct implementations and code paths; too broad to merge as duplicates." + "reason": "MPNet and VisualBert are separate implementations with independent decorator migrations, not one shared fix." }, { "left": "pull_request:44056", "right": "pull_request:44074", "accept": false, - "reason": "Shared helper migration theme only; the PRs modify different model internals and are not the same underlying change." + "reason": "MPNet vs TextNet are unrelated model code paths; this is only a thematic similarity." }, { "left": "pull_request:43997", "right": "pull_request:44010", "accept": false, - "reason": "RegNet and SqueezeBert are unrelated model implementations; similar decorator/capture refactors do not make them duplicates." + "reason": "RegNet and SqueezeBert refactor different model internals; they do not fix the same concrete issue." }, { "left": "pull_request:44010", "right": "pull_request:44129", "accept": false, - "reason": "Different architectures and different forward paths; the overlap is only in output-tracing cleanup, not the same fix." + "reason": "These touch different models and output APIs; they are separate migrations, not mergeable duplicate PRs." } ] }, @@ -17675,274 +17536,274 @@ "pull_request:44028|pull_request:44030", "pull_request:44029|pull_request:44030" ], - "prepared_review_unit_hash": "5d987a5a3454e225355b4972891993c9df239f47f625858a8e17c6f269882b55", + "prepared_review_unit_hash": "c3919d02f3fbf8f474c9c504e44a56d4d295f4af20d359637347970b2e3071b6", "component_budget": { "node_count": 13, "item_count": 13, "soft_pair_count": 40, - "serialized_chars": 40176, - "estimated_input_tokens": 10044, - "estimated_eval_tokens": 20344 + "serialized_chars": 40161, + "estimated_input_tokens": 10041, + "estimated_eval_tokens": 20338 }, "budget": { "node_count": 13, "item_count": 13, "soft_pair_count": 40, - "serialized_chars": 40176, - "estimated_input_tokens": 10044, - "estimated_eval_tokens": 20344 + "serialized_chars": 40161, + "estimated_input_tokens": 10041, + "estimated_eval_tokens": 20338 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These PRs are thematically similar output-tracing refactors, but they target different model implementations and code paths, so they are not duplicates of one another. No issue artifacts are present.", - "confidence": 0.91, + "summary": "This cluster is a set of similar-but-separate output-tracing refactors across different model families. They share a theme, but they are not duplicates of the same underlying bug/change.", + "confidence": 0.88, "canonical_issue_reason": null, - "canonical_pr_reason": "44025 is the best representative PR in this set: it applies the output-tracing refactor to a model plus its prompt variant, which captures the cluster\u2019s core pattern without being the same change as the other model-specific PRs.", + "canonical_pr_reason": "PR 44025 is the cleanest representative of the cluster: it applies the same output-tracing refactor across two closely related Depth Anything files and is a focused, coherent change.", "best_issue_reason": null, - "best_pr_reason": "44025 is the most suitable cluster representative for triage because it shows the common refactor pattern cleanly and spans a related pair of files, but it still is not a duplicate of the other PRs.", + "best_pr_reason": "PR 44025 is the best overall representative for triage because it is narrow, self-contained, and clearly exemplifies the cluster\u2019s output-tracing modernization.", "soft_edge_verdicts": [ { "left": "pull_request:43998", "right": "pull_request:44002", "accept": false, - "reason": "Same refactor theme, but different model code paths (`timm_backbone` vs `upernet`); not the same underlying change." + "reason": "Different model families and files (timm_backbone vs upernet); same refactor pattern, not the same concrete change." }, { "left": "pull_request:43998", "right": "pull_request:44000", "accept": false, - "reason": "Both are output-tracing refactors, but they affect unrelated models and forward paths." + "reason": "Different model families (timm_backbone vs vision_text_dual_encoder); only the output-tracing theme matches." }, { "left": "pull_request:43998", "right": "pull_request:44001", "accept": false, - "reason": "Different architectures (`timm_backbone` vs `univnet`) and different forward logic; only the pattern matches." + "reason": "Different architectures and files; both are refactors of return handling, not one underlying fix." }, { "left": "pull_request:43998", "right": "pull_request:44003", "accept": false, - "reason": "`timm_backbone` and `falcon_mamba/mamba` are separate implementations with different output handling, so this is not one duplicate fix." + "reason": "timm_backbone and mamba/falcon_mamba are unrelated code paths; shared refactor motif is too broad." }, { "left": "pull_request:43998", "right": "pull_request:44029", "accept": false, - "reason": "Both touch tracing-like behavior, but they are distinct model families and code paths." + "reason": "Different models and execution paths (timm_backbone vs rwkv); not mergeable as one concrete PR." }, { "left": "pull_request:43999", "right": "pull_request:44001", "accept": false, - "reason": "`mobilenet_v1` and `univnet` are unrelated models; shared refactor wording is not enough to merge them." + "reason": "MobileNetV1 and UnivNet are unrelated models; same tracing cleanup pattern only." }, { "left": "pull_request:43999", "right": "pull_request:44002", "accept": false, - "reason": "Different model subsystems and different forward signatures; not the same bug/change." + "reason": "Different subsystems (MobileNetV1 vs UperNet); no shared bug or code path." }, { "left": "pull_request:43999", "right": "pull_request:44025", "accept": false, - "reason": "Both are refactors around output tracing, but they target different model code paths and cannot plausibly be one PR." + "reason": "MobileNetV1 and Depth Anything are separate model families; both refactor outputs but do not fix the same thing." }, { "left": "pull_request:43999", "right": "pull_request:44026", "accept": false, - "reason": "`mobilenet_v1` vs `vision_encoder_decoder` are unrelated implementations; no shared concrete fix." + "reason": "MobileNetV1 and VisionEncoderDecoder are unrelated code paths; only the refactor style is similar." }, { "left": "pull_request:43999", "right": "pull_request:44027", "accept": false, - "reason": "Same general refactor pattern, but distinct encoder-decoder wrapper code paths." + "reason": "MobileNetV1 and SpeechEncoderDecoder are different architectures; not the same concrete change." }, { "left": "pull_request:44000", "right": "pull_request:44001", "accept": false, - "reason": "`vision_text_dual_encoder` and `univnet` do not share the same underlying output-tracing bug." + "reason": "Vision-Text Dual Encoder and UnivNet are unrelated models; same output-tracing modernization, not a duplicate." }, { "left": "pull_request:44000", "right": "pull_request:44002", "accept": false, - "reason": "Different models and different forward implementations; only the refactor theme is shared." + "reason": "Different model families and file scopes; no shared underlying bug." }, { "left": "pull_request:44000", "right": "pull_request:44003", "accept": false, - "reason": "`vision_text_dual_encoder` and `falcon_mamba/mamba` are separate code paths, so they are not duplicates." - }, - { - "left": "pull_request:44000", - "right": "pull_request:44004", - "accept": false, - "reason": "`vision_text_dual_encoder` and `codegen` change different model logic; same broad subsystem does not imply same change." + "reason": "Vision-Text Dual Encoder vs Mamba/FalconMamba are distinct code paths; too broad to combine." }, { "left": "pull_request:44000", "right": "pull_request:44025", "accept": false, - "reason": "Both are output-tracing refactors, but they are for different models and different runtime behavior." + "reason": "Different model families and behaviors; only the tracing refactor is common." }, { "left": "pull_request:44001", "right": "pull_request:44002", "accept": false, - "reason": "`univnet` and `upernet` are unrelated model families with distinct output semantics." + "reason": "UnivNet and UperNet are unrelated; same refactor theme does not make them duplicates." }, { "left": "pull_request:44001", "right": "pull_request:44003", "accept": false, - "reason": "Different architectures and different forward/capture behavior; not mergeable as one fix." - }, - { - "left": "pull_request:44001", - "right": "pull_request:44004", - "accept": false, - "reason": "`univnet` and `codegen` touch different code paths and different output shapes." + "reason": "UnivNet and Mamba/FalconMamba fix different model code paths; not mergeable as one PR." }, { "left": "pull_request:44001", "right": "pull_request:44025", "accept": false, - "reason": "Same theme, but the actual change is model-specific and not the same underlying bug." - }, - { - "left": "pull_request:44002", - "right": "pull_request:44004", - "accept": false, - "reason": "`upernet` segmentation forward logic and `codegen` causal LM tracing are different concrete changes." + "reason": "UnivNet and Depth Anything are separate model implementations; only the API cleanup is shared." }, { "left": "pull_request:44002", "right": "pull_request:44029", "accept": false, - "reason": "`upernet` and `rwkv` do not share the same code path or fix." + "reason": "UperNet and RWKV are unrelated architectures; not the same underlying issue." }, { "left": "pull_request:44002", "right": "pull_request:44003", "accept": false, - "reason": "Different model internals; the overlap is only in the refactor pattern." + "reason": "UperNet and Mamba/FalconMamba are different code paths; shared pattern is insufficient." }, { "left": "pull_request:44003", "right": "pull_request:44028", "accept": false, - "reason": "`falcon_mamba/mamba` and `superpoint` are unrelated implementations and cannot be merged as one PR." + "reason": "Mamba/FalconMamba and SuperPoint are unrelated model families; these are separate refactors." }, { "left": "pull_request:44003", "right": "pull_request:44004", "accept": false, - "reason": "`falcon_mamba/mamba` and `codegen` change different model code paths and output conventions." + "reason": "Mamba/FalconMamba and CodeGen are different implementations with different output-shape changes; not one concrete fix." }, { "left": "pull_request:43998", "right": "pull_request:44004", "accept": false, - "reason": "`timm_backbone` and `codegen` are separate model implementations; not the same underlying fix." + "reason": "Timm Backbone and CodeGen are unrelated model paths; same tracing cleanup only." + }, + { + "left": "pull_request:44001", + "right": "pull_request:44004", + "accept": false, + "reason": "UnivNet and CodeGen are different model types and outputs; not a duplicate change." + }, + { + "left": "pull_request:44002", + "right": "pull_request:44004", + "accept": false, + "reason": "UperNet and CodeGen are unrelated; the refactor pattern alone is not enough." + }, + { + "left": "pull_request:44000", + "right": "pull_request:44004", + "accept": false, + "reason": "Vision-Text Dual Encoder and CodeGen address different forward paths; no shared concrete bug." }, { "left": "pull_request:44025", "right": "pull_request:44026", "accept": false, - "reason": "`depth_anything` and `vision_encoder_decoder` share a refactor motif, but they are different model wrappers and different forward code." + "reason": "Depth Anything and VisionEncoderDecoder are unrelated model families; only the output-tracing refactor pattern overlaps." }, { "left": "pull_request:44025", "right": "pull_request:44027", "accept": false, - "reason": "`depth_anything` and `speech_encoder_decoder` are distinct code paths; no single duplicate PR here." + "reason": "Depth Anything and SpeechEncoderDecoder are different code paths; not the same underlying change." }, { "left": "pull_request:44025", "right": "pull_request:44028", "accept": false, - "reason": "Different models and different output-tracing implementations; not the same bug/change." + "reason": "Depth Anything and SuperPoint are distinct models; same modernized return handling, not a duplicate." }, { "left": "pull_request:44025", "right": "pull_request:44029", "accept": false, - "reason": "`depth_anything` and `rwkv` modify unrelated forward logic." + "reason": "Depth Anything and RWKV are unrelated implementations; too broad to merge as one PR." }, { "left": "pull_request:44025", "right": "pull_request:44030", "accept": false, - "reason": "`depth_anything` and `dpr` are different model families; only the refactor style is similar." + "reason": "Depth Anything and DPR are separate model families; same refactor theme only." }, { "left": "pull_request:44026", "right": "pull_request:44029", "accept": false, - "reason": "`vision_encoder_decoder` and `rwkv` are unrelated code paths and not a duplicate fix." + "reason": "VisionEncoderDecoder and RWKV are unrelated model paths; not the same concrete problem." }, { "left": "pull_request:44026", "right": "pull_request:44027", "accept": false, - "reason": "Both are encoder-decoder wrappers, but they are different model families and separate changes." + "reason": "VisionEncoderDecoder and SpeechEncoderDecoder are different wrapper models; no shared bug." }, { "left": "pull_request:44026", "right": "pull_request:44028", "accept": false, - "reason": "`vision_encoder_decoder` and `superpoint` do not share the same underlying code change." + "reason": "VisionEncoderDecoder and SuperPoint are unrelated; both just adopt the same tracing API changes." }, { "left": "pull_request:44026", "right": "pull_request:44030", "accept": false, - "reason": "Different model implementations and different output semantics; not mergeable as one PR." + "reason": "VisionEncoderDecoder and DPR are distinct code paths; not duplicates." }, { "left": "pull_request:44027", "right": "pull_request:44028", "accept": false, - "reason": "`speech_encoder_decoder` and `superpoint` are unrelated models; same broad tracing theme is insufficient." + "reason": "SpeechEncoderDecoder and SuperPoint are unrelated model families; same output-tracing theme only." }, { "left": "pull_request:44027", "right": "pull_request:44029", "accept": false, - "reason": "Different architectures and different forward paths; not a duplicate." + "reason": "SpeechEncoderDecoder and RWKV are different architectures; not one underlying change." }, { "left": "pull_request:44027", "right": "pull_request:44030", "accept": false, - "reason": "`speech_encoder_decoder` and `dpr` change different model code and output handling." + "reason": "SpeechEncoderDecoder and DPR are unrelated; no shared concrete fix." }, { "left": "pull_request:44028", "right": "pull_request:44029", "accept": false, - "reason": "`superpoint` and `rwkv` are unrelated model implementations with no shared concrete bug." + "reason": "SuperPoint and RWKV are different model paths; not the same bug or change." }, { "left": "pull_request:44028", "right": "pull_request:44030", "accept": false, - "reason": "Different models and different output logic; only the refactor pattern matches." + "reason": "SuperPoint and DPR are unrelated implementations; only the refactor style matches." }, { "left": "pull_request:44029", "right": "pull_request:44030", "accept": false, - "reason": "`rwkv` and `dpr` are distinct code paths with no evidence of the same underlying change." + "reason": "RWKV and DPR are different model families; same output-tracing motif is too broad to treat as duplicate." } ] }, @@ -17975,52 +17836,52 @@ "pull_request:44662|pull_request:45075", "pull_request:45075|pull_request:45401" ], - "prepared_review_unit_hash": "dc9d3f85e75d98aa8c3e8a257b74d02c72e8deca7bb00a537be74a82e15a48f7", + "prepared_review_unit_hash": "9ed0a4d6e72151d80cc757f8aabedde7367ea672ef04932995acd51146e3c693", "component_budget": { "node_count": 3, "item_count": 3, "soft_pair_count": 3, - "serialized_chars": 10912, - "estimated_input_tokens": 2728, - "estimated_eval_tokens": 5712 + "serialized_chars": 10914, + "estimated_input_tokens": 2729, + "estimated_eval_tokens": 5714 }, "budget": { "node_count": 3, "item_count": 3, "soft_pair_count": 3, - "serialized_chars": 10912, - "estimated_input_tokens": 2728, - "estimated_eval_tokens": 5712 + "serialized_chars": 10914, + "estimated_input_tokens": 2729, + "estimated_eval_tokens": 5714 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "All three items are separate model-addition PRs for different models (PenguinVL, DeepSeek-OCR-2, and Voxtral TTS). They share Transformers boilerplate and auto-registration files, but they do not describe the same code-path bug or the same feature change, so none are duplicates.", - "confidence": 0.93, + "summary": "These are three separate model-addition PRs (PenguinVL, DeepSeek-OCR-2, Voxtral TTS). They share common Transformers scaffolding and docs/auto-registry edits, but they do not address the same model or concrete code-path bug, so they are not duplicates.", + "confidence": 0.97, "canonical_issue_reason": null, - "canonical_pr_reason": "No single PR is a true duplicate-canonical candidate because each adds a different model. If one representative must be chosen for the cluster, PR #45075 is the best anchor since it is the most complete/established model-addition PR and also links an explicit issue target.", + "canonical_pr_reason": null, "best_issue_reason": null, - "best_pr_reason": "PR #45075 is the most representative of the cluster only in the sense that it has the clearest scope and explicit issue linkage; however, it is still not a duplicate of the others because the model, files, and implementation are distinct.", + "best_pr_reason": "No single PR is a duplicate anchor here; each introduces a different model implementation with its own files and scope.", "soft_edge_verdicts": [ { "left": "pull_request:44662", "right": "pull_request:45401", "accept": false, - "reason": "PenguinVL implementation vs Voxtral TTS support are unrelated model additions. Shared docs/auto-registry files are expected boilerplate, not evidence of the same underlying change." + "reason": "Different models and different code paths; only generic model-registration/docs overlap." }, { "left": "pull_request:44662", "right": "pull_request:45075", "accept": false, - "reason": "Both are model implementation PRs, but for different models with different configs, processors, and tests. They do not fix or add the same concrete code path." + "reason": "Separate model implementations (PenguinVL vs DeepSeek-OCR-2); shared auto/docs plumbing is not enough to merge as one PR." }, { "left": "pull_request:45075", "right": "pull_request:45401", "accept": false, - "reason": "DeepSeek-OCR-2 and Voxtral TTS are distinct model integrations. Overlap in shared registration and documentation files is too broad to treat them as the same PR." + "reason": "Different model families and functionality; overlap is limited to shared Transformers boilerplate." } ] }, @@ -18050,12 +17911,12 @@ "soft_pairs": [ "pull_request:44979|pull_request:45363" ], - "prepared_review_unit_hash": "78d8311b9302eff1d60754b0ae5793baaa0313470dc9281a558135942402c868", + "prepared_review_unit_hash": "f6c5b2dbc9d5006d4549d38c3eef47bad39c7fc3b50f36d1d92a952477e0a08a", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 4408, + "serialized_chars": 4407, "estimated_input_tokens": 1102, "estimated_eval_tokens": 2460 }, @@ -18063,7 +17924,7 @@ "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 4408, + "serialized_chars": 4407, "estimated_input_tokens": 1102, "estimated_eval_tokens": 2460 }, @@ -18072,18 +17933,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Related but distinct PRs: one adds a generic module-fusion API, while the other implements KernelConfig-driven n-to-1 kernel fusion plus integration plumbing and an example. They overlap conceptually, but they are not the same underlying change.", - "confidence": 0.88, + "summary": "These PRs are related conceptually around module/kernel fusion, but they implement different layers of the stack: #44979 adds a new generic Module Fusion API, while #45363 adds KernelConfig-based n-to-1 kernel fusion plumbing and an example. They are not the same underlying change and should stay separate.", + "confidence": 0.86, "canonical_issue_reason": null, - "canonical_pr_reason": "45363 is the clearer representative of the set because it delivers the concrete KernelConfig-based fusion workflow and integration points, whereas 44979 is broader foundational API scaffolding.", + "canonical_pr_reason": "#44979 is the better canonical PR because it is the more foundational, self-contained introduction of the fusion API and tests; #45363 is a broader, more specific follow-on around KernelConfig integration and example usage.", "best_issue_reason": null, - "best_pr_reason": "45363 is the most end-user-facing and concrete PR here: it wires fusion through model loading, integration helpers, and a working example, so it best summarizes the cluster theme.", + "best_pr_reason": "#44979 is the best PR representative for the cluster since it defines the core abstraction being introduced. #45363 is a downstream implementation/integration PR rather than the same concrete code-path fix or mergeable change set.", "soft_edge_verdicts": [ { "left": "pull_request:44979", "right": "pull_request:45363", "accept": false, - "reason": "Reject: 44979 introduces a generic module-fusion API, while 45363 adds a specific n-to-1 kernel-fusion implementation via KernelConfig and related integration code. Similar area, but not the same concrete change and not plausibly one merged PR." + "reason": "Related topic, but not the same change: #44979 introduces the generic Module Fusion API, while #45363 wires in KernelConfig-based n-to-1 kernel fusion and example code. They touch different code paths and would not plausibly be merged as one PR." } ] }, @@ -18113,7 +17974,7 @@ "soft_pairs": [ "pull_request:45086|pull_request:45317" ], - "prepared_review_unit_hash": "e3005195c7cb66424c7400ec40f5a0b8b77ef2a654964c71948c34c7f431d8ce", + "prepared_review_unit_hash": "07996f7c637634ea9322ce3656527843f98c8799c76a290ca3866d560dd48483", "component_budget": { "node_count": 2, "item_count": 2, @@ -18135,18 +17996,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Both pull requests fix the same _patch_mistral_regex AttributeError in the Mistral tokenizer patch path and target the same issue; 45317 is the more complete variant with tests.", + "summary": "Both pull requests address the same Mistral regex AttributeError in `_patch_mistral_regex`; 45317 is the more complete version because it includes the fix plus tests, so it is the best representative of the cluster.", "confidence": 0.97, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 45317 is the better canonical representative because it contains the same core fix as 45086 plus a test update and a clearer, more specific title.", + "canonical_pr_reason": "PR 45317 is the stronger canonical PR: it implements the same core code-path fix as 45086 and adds test coverage, making it the more complete representative of the duplicate pair.", "best_issue_reason": null, - "best_pr_reason": "PR 45317 is the strongest overall representative of the cluster: same bug fix, same file, and it includes validation coverage.", + "best_pr_reason": "PR 45317 is the best PR to keep because it contains the same bug fix as 45086 plus validation, so it is the most useful single record for this cluster.", "soft_edge_verdicts": [ { "left": "pull_request:45086", "right": "pull_request:45317", "accept": true, - "reason": "Same concrete AttributeError fix in _patch_mistral_regex on the same code path; 45317 is essentially the tested refinement of 45086, so these look mergeable into one PR." + "reason": "Same underlying AttributeError fix in `_patch_mistral_regex` on the same file and shared issue target; the second PR appears to be a superset with tests, so they are mergeable duplicates." } ] }, @@ -18176,7 +18037,7 @@ "soft_pairs": [ "pull_request:45115|pull_request:45569" ], - "prepared_review_unit_hash": "16032a8c74bdf26305b4cbf8ec8470130c8855d0e81db4387316ee16bd694695", + "prepared_review_unit_hash": "3e7e75e904d3a0d100ab4c19578907fbba5e573c92ee02f7e68e716d82a02cd1", "component_budget": { "node_count": 2, "item_count": 2, @@ -18198,18 +18059,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Both pull requests are about NemotronH model refactoring, but they target different concrete changes: one is a conversion/mapping and inheritance cleanup, the other introduces separate dense/sparse model implementations and docs updates. They are related, but not the same underlying change set.", - "confidence": 0.87, + "summary": "These are related NemotronH refactors, but they are not the same change: one is a narrow inheritance/mapping refactor, the other is a broader model split introducing NemotronHDense/NemotronHSparse plus docs and tests.", + "confidence": 0.89, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 45569 is the better canonical PR because it is the broader, more complete NemotronH restructuring: it adds the dedicated dense/sparse model split plus auto-mapping, docs, and tests, while 45115 is a narrower conversion/inheritance refactor.", + "canonical_pr_reason": "pull_request:45569 is the broader, more complete implementation: it adds the dedicated dense/sparse model packages, updates auto-mapping/docs, and includes tests. pull_request:45115 is a narrower support refactor for conversion/mapping and inheritance.", "best_issue_reason": null, - "best_pr_reason": "PR 45569 is the best overall PR candidate for this cluster because it represents the main user-facing architecture change and subsumes more of the NemotronH work than the narrower mapping-focused PR.", + "best_pr_reason": "pull_request:45569 is the best representative PR because it captures the full architectural change and user-visible packaging/docs updates, whereas pull_request:45115 is only a smaller internal refactor.", "soft_edge_verdicts": [ { "left": "pull_request:45115", "right": "pull_request:45569", "accept": false, - "reason": "Related NemotronH work, but not the same concrete bug/change: 45115 focuses on checkpoint conversion and class renaming/inheritance, while 45569 splits NemotronH into dense/sparse variants with docs/tests. They would not naturally merge into one PR as a single fix." + "reason": "Shared NemotronH filenames do not imply duplicate scope: 45115 focuses on conversion mapping and inheritance refactoring, while 45569 introduces separate dense/sparse model implementations, docs, auto-mapping, and tests. They address related but different code paths and would not plausibly collapse into one PR." } ] }, @@ -18239,7 +18100,7 @@ "soft_pairs": [ "pull_request:45332|pull_request:45333" ], - "prepared_review_unit_hash": "12a2dcf035a6dbaa361e70043a7ef3db7560976af018ae0cb944407ac9bdaa08", + "prepared_review_unit_hash": "7351ab2a7b013fcc34fe7972ac6f0c83bba09ce692e7b20e3ca17b177b19f36d", "component_budget": { "node_count": 2, "item_count": 2, @@ -18261,18 +18122,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These are closely related feature PRs for heterogeneous model support, but they are not duplicates: 45333 adds per-layer config plumbing, while 45332 adds the downstream modeling/cache/masking changes that consume that plumbing.", - "confidence": 0.87, + "summary": "These are related but distinct PRs: one adds per-layer heterogeneous config plumbing, the other adds broader heterogeneous model/runtime support. They are not the same concrete change, so the soft edge should be rejected.", + "confidence": 0.92, "canonical_issue_reason": null, - "canonical_pr_reason": "45332 is the broader representative PR because it covers the end-to-end heterogeneous modeling path, not just config scaffolding.", + "canonical_pr_reason": "No true canonical duplicate here; PR 45333 is the more foundational per-layer config piece, while PR 45332 builds additional modeling/runtime support on top of that theme.", "best_issue_reason": null, - "best_pr_reason": "45332 is the best overall PR to represent the cluster since it includes the more complete change set; 45333 is narrower groundwork for the same feature area.", + "best_pr_reason": "PR 45333 is the cleaner representative for the shared heterogeneity/configuration thread because it introduces the core per-layer config abstraction used by the broader support work.", "soft_edge_verdicts": [ { "left": "pull_request:45332", "right": "pull_request:45333", "accept": false, - "reason": "Related but not the same concrete change: 45333 adds heterogeneous config support, while 45332 adds modeling/cache handling that depends on that config. They are complementary, not duplicate fixes, and would not naturally collapse into one PR." + "reason": "Related but not duplicate: 45333 is config-layer plumbing, while 45332 is a larger heterogeneous model/runtime change. They touch the same area but do not appear to fix the same concrete code-path problem, and would not naturally collapse into one PR." } ] }, @@ -18302,40 +18163,40 @@ "soft_pairs": [ "pull_request:45415|pull_request:45425" ], - "prepared_review_unit_hash": "4fee7047114a985ccf633afd33d0422cc5ee0319096c46a692b227271d747ffc", + "prepared_review_unit_hash": "088834af3a3151c52c465b2346ff99f7e3d7474af9b58464bc4a4702a7dc72a4", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 5514, - "estimated_input_tokens": 1379, - "estimated_eval_tokens": 3014 + "serialized_chars": 5512, + "estimated_input_tokens": 1378, + "estimated_eval_tokens": 3012 }, "budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 5514, - "estimated_input_tokens": 1379, - "estimated_eval_tokens": 3014 + "serialized_chars": 5512, + "estimated_input_tokens": 1378, + "estimated_eval_tokens": 3012 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These are related typing-focused PRs, but they are not duplicates: one is a broad repo-wide type-checking cleanup, while the other adds/adjusts typing protocols and helpers in specific code paths.", - "confidence": 0.9, + "summary": "Both pull requests are type-checking related, but they address different changes: #45415 is a broad pass adding type checking and suppressing many type-ignore warnings across numerous files, while #45425 introduces typing protocols/helpers and adjusts a smaller set of call sites. They are not the same underlying fix and should not be merged as duplicates.", + "confidence": 0.94, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 45415 is the broader, more representative typing cleanup across many `src/transformers/*` files and the type-checking config, so it fits best as the cluster\u2019s canonical PR.", + "canonical_pr_reason": "#45425 is the better canonical PR for this cluster because it is narrower and more concrete, centered on new typing abstractions and targeted call-site updates rather than a wide repository-wide type-check sweep.", "best_issue_reason": null, - "best_pr_reason": "PR 45415 is the strongest overall representative because it covers the widest scope of the typing work and is closer to the cluster\u2019s apparent theme than the narrower helper-type refactor in PR 45425.", + "best_pr_reason": "#45425 is the better representative PR: it has a clearer, more specific change set and a tighter code-path focus than #45415's broad type-checking cleanup.", "soft_edge_verdicts": [ { "left": "pull_request:45415", "right": "pull_request:45425", "accept": false, - "reason": "Both are typing-related, but they address different changes: 45415 is a broad type-checking sweep, while 45425 introduces specific typing protocol/helper definitions. They do not look like the same concrete bug or change and would not plausibly merge as one PR." + "reason": "They share a typing theme and a few filenames, but the changes are substantively different: #45415 is a broad type-checking/config cleanup across many modules, while #45425 adds typing protocols and targeted adjustments for specific interfaces. Not the same concrete bug/change." } ] }, @@ -18365,12 +18226,12 @@ "soft_pairs": [ "pull_request:45487|pull_request:45615" ], - "prepared_review_unit_hash": "e0368674c8d01752fcef3010dab830c751cb9885c1eb9648addb11c5d4cf9fcc", + "prepared_review_unit_hash": "4d8f8e6fc38d36dde1c43d759efc5af5b24d0e92cb6567d7d9803e6468bbc7d8", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 5159, + "serialized_chars": 5160, "estimated_input_tokens": 1290, "estimated_eval_tokens": 2836 }, @@ -18378,7 +18239,7 @@ "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 5159, + "serialized_chars": 5160, "estimated_input_tokens": 1290, "estimated_eval_tokens": 2836 }, @@ -18387,18 +18248,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Two pull requests: one fixes a real model-parallel device placement bug in AltCLIP/ChineseCLIP-related model code; the other only adds XPU test expectations for Qianfan OCR. They are not the same underlying change.", - "confidence": 0.96, + "summary": "These two PRs are unrelated: one fixes a real device-placement bug in model code for AltCLIP/ChineseClip, while the other only updates Qianfan OCR XPU test expectations.", + "confidence": 0.97, "canonical_issue_reason": null, - "canonical_pr_reason": "PR 45487 is the stronger canonical PR because it contains the substantive code fix across model implementations, not just test-data updates.", + "canonical_pr_reason": "pull_request:45487 is the canonical PR because it changes the actual forward path/device handling in model implementation files across AltCLIP and ChineseClip, addressing a concrete runtime bug.", "best_issue_reason": null, - "best_pr_reason": "PR 45487 is the best representative of the cluster since it addresses the actual bug in model forward logic and has broader functional impact than the test-only PR.", + "best_pr_reason": "pull_request:45487 is the best PR candidate for the cluster since it contains the substantive code fix; pull_request:45615 is test-only and does not fix the underlying code-path problem.", "soft_edge_verdicts": [ { "left": "pull_request:45487", "right": "pull_request:45615", "accept": false, - "reason": "Different subsystems and different change types: 45487 fixes device placement in AltCLIP/ChineseCLIP model code, while 45615 only adds XPU expectations to Qianfan OCR tests. They do not appear mergeable into one PR or to address the same concrete bug." + "reason": "Reject: these do not fix the same bug. 45487 patches model code to move token_type_ids to the correct device; 45615 only adds XPU test expectations for a different model." } ] }, @@ -18428,12 +18289,12 @@ "soft_pairs": [ "pull_request:45548|pull_request:45662" ], - "prepared_review_unit_hash": "d980997f34a5a13858a7d7bcb1e998c652b53d4e79a484adda636ef444c61fbe", + "prepared_review_unit_hash": "01b3cde4d662b289184fb928740237ba651cce03d593a2190748ec164bf777ae", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 4508, + "serialized_chars": 4507, "estimated_input_tokens": 1127, "estimated_eval_tokens": 2510 }, @@ -18441,7 +18302,7 @@ "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 4508, + "serialized_chars": 4507, "estimated_input_tokens": 1127, "estimated_eval_tokens": 2510 }, @@ -18450,18 +18311,18 @@ "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Both items are expert-parallelism loading fixes, but they address different failure modes: one removes `distributed_config` from serialization and adds `has_ep` for DeepSpeed ZeRO-3/accelerate loading, while the other prevents expert parameters from being overwritten during FSDP2 loading by post-shard wrapping. The shared `modeling_utils.py` touch is superficial, not evidence of the same bug.", - "confidence": 0.96, + "summary": "Both PRs are about expert parallelism, but they fix different concrete failure modes: one is EP + ZeRO-3/accelerate config loading, the other is EP + FSDP2 expert parameters being overwritten during sharded loading. They are related, not duplicates.", + "confidence": 0.92, "canonical_issue_reason": null, - "canonical_pr_reason": "No single canonical PR: these are distinct EP-loading bugs with different code paths and fixes, so neither PR is a duplicate of the other.", + "canonical_pr_reason": "PR 45662 is the better representative because it targets the concrete expert-overwrite path in sharded loading/runtime, with broader EP/FSDP2 handling changes. PR 45548 is a separate accelerate/ZeRO-3 loading/config serialization fix.", "best_issue_reason": null, - "best_pr_reason": "PR 45662 is the stronger representative of an EP-loading bug because it fixes a concrete parameter materialization issue in the loader path, but it still does not subsume 45548.", + "best_pr_reason": "PR 45662 is the strongest standalone candidate: it addresses a specific silent corruption bug for experts and includes the relevant loading and DTensor/sharding changes. PR 45548 looks like a different EP-related loading bug.", "soft_edge_verdicts": [ { "left": "pull_request:45548", "right": "pull_request:45662", "accept": false, - "reason": "Different underlying problems: 45548 is about config serialization and EP detection for ZeRO-3/accelerate, while 45662 is about FSDP2 load-time overwriting of experts and shard wrapping. They are related but not the same fix." + "reason": "Different bug paths and fixes: 45548 is about EP + DeepSpeed ZeRO-3 loading via accelerate launch and config serialization; 45662 is about EP + FSDP2 experts being overwritten by rank-0 broadcast during sharded loading. Shared EP context is not enough to merge them as one PR." } ] }, @@ -18483,55 +18344,73 @@ "source": "llm", "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", - "cluster_id": "cluster-45639-2", + "cluster_id": "cluster-45639-3", "nodes": [ "pull_request:45639", - "pull_request:45645" + "pull_request:45645", + "pull_request:45778" ], "soft_pairs": [ - "pull_request:45639|pull_request:45645" + "pull_request:45639|pull_request:45645", + "pull_request:45645|pull_request:45778", + "pull_request:45639|pull_request:45778" ], - "prepared_review_unit_hash": "d03f85e9194c61794d0f0bd1f6cd16c7f1b2537c16740dd04f65fcebfb20afaa", + "prepared_review_unit_hash": "0e6802102eca3044722f28bf7c0c1ac2815887d2f6f30ab0f23d6410c4412efb", "component_budget": { - "node_count": 2, - "item_count": 2, - "soft_pair_count": 1, - "serialized_chars": 4589, - "estimated_input_tokens": 1148, - "estimated_eval_tokens": 2552 + "node_count": 3, + "item_count": 3, + "soft_pair_count": 3, + "serialized_chars": 7470, + "estimated_input_tokens": 1868, + "estimated_eval_tokens": 3992 }, "budget": { - "node_count": 2, - "item_count": 2, - "soft_pair_count": 1, - "serialized_chars": 4589, - "estimated_input_tokens": 1148, - "estimated_eval_tokens": 2552 + "node_count": 3, + "item_count": 3, + "soft_pair_count": 3, + "serialized_chars": 7470, + "estimated_input_tokens": 1868, + "estimated_eval_tokens": 3992 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "Both PRs address the same xdist-related captured_info collision problem, with 45645 effectively extending the core fix from 45639 into CI log collection and related cleanup.", - "confidence": 0.86, + "summary": "All three PRs address the same pytest-xdist/captured_info collision in testing_utils, with PR 45645 providing the broadest end-to-end fix because it also updates CI artifact collection. The relationships are best treated as overlapping fixes for the same bug, not as a strict superset claim.", + "confidence": 0.89, "canonical_issue_reason": null, - "canonical_pr_reason": "45645 is the more complete representative: it includes the xdist-safe output-file change plus the CI workflow update to collect multiple captured_info artifacts.", + "canonical_pr_reason": "PR 45645 is the best canonical PR because it fixes the captured_info collision in the test helper and also updates the CI workflow that consumes those logs, making it the most complete standalone change in the cluster.", "best_issue_reason": null, - "best_pr_reason": "45645 is the best standalone representative because it covers the underlying bug fix and the downstream CI handling needed to preserve those logs.", + "best_pr_reason": "PR 45645 is the strongest representative of the cluster because it covers the core xdist-safe logging fix plus the downstream artifact handling needed for CI, whereas the others are narrower variants of the same bugfix.", "soft_edge_verdicts": [ { "left": "pull_request:45639", "right": "pull_request:45645", "accept": true, - "reason": "Both PRs fix the same concrete captured_info/xdist collision path in testing_utils, and 45645 appears to build on the same fix rather than introduce a separate change." + "reason": "Same underlying xdist captured_info collision in the same testing code path; 45645 is an overlapping, more complete fix that also adjusts CI log collection." + }, + { + "left": "pull_request:45645", + "right": "pull_request:45778", + "accept": true, + "reason": "Both target the same captured_info xdist race in testing_utils; they use different implementation details, but they are still overlapping fixes for the same bug." + }, + { + "left": "pull_request:45639", + "right": "pull_request:45778", + "accept": true, + "reason": "Both PRs fix the same concrete captured_info collision under pytest-xdist in src/transformers/testing_utils.py, so they look like alternative implementations for the same bug." } ] }, - "evaluator_result": null, - "evaluator_used": false, - "retried": false, - "accepted_nontrivial_soft_edge": false, + "evaluator_result": { + "accept": false, + "feedback": "The cluster is plausible, but one soft-edge verdict is too strong for the evidence. In particular, describing PR 45645 as a \"superset\" of PR 45639 is not directly supported by the packet: both touch the same xdist/captured_info area, but they appear to use different implementation details, so the relationship should be stated more conservatively as \"overlapping fix for the same bug.\"" + }, + "evaluator_used": true, + "retried": true, + "accepted_nontrivial_soft_edge": true, "error_kind": null, "error_message": null }, @@ -18543,8 +18422,8 @@ "review_unit_count": 1, "status": "reviewed", "reason": null, - "source": "cache", - "cache_hit": true, + "source": "llm", + "cache_hit": false, "model": "gpt-5.4-mini?service_tier=flex", "cluster_id": "cluster-45683-2", "nodes": [ @@ -18554,40 +18433,40 @@ "soft_pairs": [ "pull_request:45683|pull_request:45734" ], - "prepared_review_unit_hash": "5bf556c48e00ed86e4c03ceb7e01f11cf4b18f29c45b12a017e905cb44e456f3", + "prepared_review_unit_hash": "eba4ca0f26b2d3be0c00e316270a9341d31af2208b44dbab31b1d3a8e81e7538", "component_budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 3891, - "estimated_input_tokens": 973, - "estimated_eval_tokens": 2202 + "serialized_chars": 4469, + "estimated_input_tokens": 1118, + "estimated_eval_tokens": 2492 }, "budget": { "node_count": 2, "item_count": 2, "soft_pair_count": 1, - "serialized_chars": 3891, - "estimated_input_tokens": 973, - "estimated_eval_tokens": 2202 + "serialized_chars": 4469, + "estimated_input_tokens": 1118, + "estimated_eval_tokens": 2492 }, "overflow_policy": "truncate_then_skip", "trimmed": false, "aggressively_trimmed": false, "split": false, "analyst_result": { - "summary": "These PRs are not duplicates: 45683 adds audio-module skips for multimodal quantization crashes, while 45734 fixes skip_modules semantics so auto-detected defaults remain skipped and adds tests.", - "confidence": 0.96, + "summary": "Two open PRs touch the same quantization helper, but they fix different bugs: one special-cases audio submodules, the other changes how default skip lists are combined with user-supplied skips. They should not be clustered as duplicates.", + "confidence": 0.93, "canonical_issue_reason": null, - "canonical_pr_reason": null, + "canonical_pr_reason": "45734 is the better representative PR because it fixes the core skip-module composition logic in `get_modules_to_not_convert`, which is broader and more central than the narrow audio-module special case in 45683.", "best_issue_reason": null, - "best_pr_reason": "45734 is the stronger standalone fix: it patches the skip_modules regression, includes tests, and has the explicit issue link; 45683 addresses a different audio-specific crash.", + "best_pr_reason": "45734 is the strongest standalone PR in the set: it addresses a concrete quantization failure mode in the main code path and its fix is directly about the helper\u2019s behavior rather than a model-specific workaround.", "soft_edge_verdicts": [ { "left": "pull_request:45683", "right": "pull_request:45734", "accept": false, - "reason": "Shared file and issue target are incidental; 45683 skips audio_tower/embed_audio to avoid a uint8 crash, whereas 45734 changes the skip_modules/default-skip control flow to keep lm_head and tied embeddings excluded and prevent a bitsandbytes AssertionError. Different bugs and not mergeable into one PR." + "reason": "Both touch quantization skip logic, but 45683 adds audio-tower/embed_audio exclusions while 45734 changes additive handling of `skip_modules`; these are different underlying bugs and would not naturally merge into one PR." } ] }, diff --git a/analysis/current/manifest.json b/analysis/current/manifest.json index 3d300bdd6047f1f718841421515cdebb35d8c3f5..2ac77f6ccca999f208d0ce8cf74d5cbe38bccc5c 100644 --- a/analysis/current/manifest.json +++ b/analysis/current/manifest.json @@ -1,8 +1,8 @@ { - "analysis_id": "hybrid-model-20260505t000047z", + "analysis_id": "hybrid-model-20260505t060044z", "archived_artifacts": { - "hybrid": "snapshots/20260505T000047Z/analysis-runs/hybrid-model-20260505t000047z/analysis-report-hybrid.json", - "hybrid_reviews": "snapshots/20260505T000047Z/analysis-runs/hybrid-model-20260505t000047z/analysis-report-hybrid.llm-reviews.json" + "hybrid": "snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/analysis-report-hybrid.json", + "hybrid_reviews": "snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/analysis-report-hybrid.llm-reviews.json" }, "artifacts": { "hybrid": "analysis/current/analysis-report-hybrid.json", @@ -10,9 +10,9 @@ }, "channel": "canonical", "model": null, - "published_at": "2026-05-05T00:04:58Z", + "published_at": "2026-05-05T06:09:30Z", "repo": "huggingface/transformers", "schema_version": 1, - "snapshot_id": "20260505T000047Z", + "snapshot_id": "20260505T060044Z", "variant": "hybrid" } diff --git a/snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/analysis-report-hybrid.json b/snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/analysis-report-hybrid.json new file mode 100644 index 0000000000000000000000000000000000000000..245cc71d535fa164c0a085a1ef36871a30b53320 --- /dev/null +++ b/snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/analysis-report-hybrid.json @@ -0,0 +1,2281 @@ +{ + "schema_version": "1.0", + "repo": "huggingface/transformers", + "snapshot_id": "20260505T060044Z", + "generated_at": "2026-05-05T06:09:27Z", + "evidence_quality": "full", + "llm_enrichment": true, + "meta_bugs": [ + { + "cluster_id": "cluster-43979-11", + "summary": "Cluster of 1 issues and 10 PRs centered on issue #43979.", + "status": "open", + "confidence": 0.8, + "canonical_issue_number": 43979, + "canonical_pr_number": 44007, + "issue_numbers": [ + 43979 + ], + "pr_numbers": [ + 43996, + 44007, + 44013, + 44044, + 44066, + 44072, + 44085, + 44129, + 44154, + 44722 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target", + "soft_similarity" + ], + "pr_comparisons": [ + { + "left_pr_number": 43996, + "right_pr_number": 44007, + "code_similarity": 0.179, + "size_similarity": 0.576, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.429, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44013, + "code_similarity": 0.122, + "size_similarity": 0.318, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.392, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44044, + "code_similarity": 0.245, + "size_similarity": 0.864, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.479, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44066, + "code_similarity": 0.225, + "size_similarity": 0.818, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.408, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44072, + "code_similarity": 0.14, + "size_similarity": 0.303, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.528, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44085, + "code_similarity": 0.216, + "size_similarity": 0.783, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.398, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44129, + "code_similarity": 0.163, + "size_similarity": 0.643, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.229, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44154, + "code_similarity": 0.153, + "size_similarity": 0.535, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.31, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 43996, + "right_pr_number": 44722, + "code_similarity": 0.225, + "size_similarity": 0.848, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.368, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44013, + "code_similarity": 0.19, + "size_similarity": 0.553, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.531, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44044, + "code_similarity": 0.186, + "size_similarity": 0.667, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.354, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44066, + "code_similarity": 0.188, + "size_similarity": 0.704, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.315, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44072, + "code_similarity": 0.212, + "size_similarity": 0.526, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.708, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44085, + "code_similarity": 0.195, + "size_similarity": 0.735, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.318, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44129, + "code_similarity": 0.103, + "size_similarity": 0.37, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.191, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44154, + "code_similarity": 0.238, + "size_similarity": 0.93, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.344, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44007, + "right_pr_number": 44722, + "code_similarity": 0.178, + "size_similarity": 0.679, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.28, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44013, + "right_pr_number": 44044, + "code_similarity": 0.126, + "size_similarity": 0.368, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.351, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44013, + "right_pr_number": 44066, + "code_similarity": 0.127, + "size_similarity": 0.389, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.325, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44013, + "right_pr_number": 44072, + "code_similarity": 0.29, + "size_similarity": 0.952, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.667, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44013, + "right_pr_number": 44085, + "code_similarity": 0.131, + "size_similarity": 0.406, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.329, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44013, + "right_pr_number": 44129, + "code_similarity": 0.07, + "size_similarity": 0.205, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.192, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44013, + "right_pr_number": 44154, + "code_similarity": 0.177, + "size_similarity": 0.594, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.389, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44013, + "right_pr_number": 44722, + "code_similarity": 0.118, + "size_similarity": 0.375, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.287, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44044, + "right_pr_number": 44066, + "code_similarity": 0.25, + "size_similarity": 0.947, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.404, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44044, + "right_pr_number": 44072, + "code_similarity": 0.136, + "size_similarity": 0.351, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.442, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44044, + "right_pr_number": 44085, + "code_similarity": 0.24, + "size_similarity": 0.906, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.394, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44044, + "right_pr_number": 44129, + "code_similarity": 0.147, + "size_similarity": 0.555, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.243, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44044, + "right_pr_number": 44154, + "code_similarity": 0.17, + "size_similarity": 0.62, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.306, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44044, + "right_pr_number": 44722, + "code_similarity": 0.257, + "size_similarity": 0.982, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.402, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44066, + "right_pr_number": 44072, + "code_similarity": 0.133, + "size_similarity": 0.37, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.393, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44066, + "right_pr_number": 44085, + "code_similarity": 0.763, + "size_similarity": 0.957, + "file_overlap": 0.5, + "area_overlap": 0.825, + "patch_similarity": 0.887, + "shared_filenames": [ + "src/transformers/models/gptj/modeling_gptj.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/models/gptj/modeling_gptj.py", + "left_ranges": [ + [ + 33, + 41 + ], + [ + 174, + 181 + ], + [ + 250, + 257 + ], + [ + 398, + 419 + ], + [ + 425, + 434 + ], + [ + 459, + 466 + ], + [ + 471, + 488 + ], + [ + 518, + 540 + ], + [ + 554, + 560 + ], + [ + 566, + 575 + ], + [ + 580, + 598 + ], + [ + 601, + 606 + ], + [ + 634, + 640 + ], + [ + 646, + 653 + ], + [ + 658, + 674 + ], + [ + 717, + 722 + ], + [ + 738, + 744 + ], + [ + 749, + 772 + ], + [ + 790, + 795 + ] + ], + "right_ranges": [ + [ + 33, + 40 + ], + [ + 173, + 180 + ], + [ + 249, + 256 + ], + [ + 397, + 405 + ], + [ + 408, + 419 + ], + [ + 425, + 434 + ], + [ + 459, + 465 + ], + [ + 470, + 489 + ], + [ + 517, + 539 + ], + [ + 553, + 559 + ], + [ + 565, + 574 + ], + [ + 579, + 597 + ], + [ + 600, + 611 + ], + [ + 633, + 639 + ], + [ + 645, + 652 + ], + [ + 657, + 673 + ], + [ + 716, + 728 + ], + [ + 737, + 743 + ], + [ + 748, + 771 + ], + [ + 789, + 794 + ] + ] + } + ] + }, + { + "left_pr_number": 44066, + "right_pr_number": 44129, + "code_similarity": 0.145, + "size_similarity": 0.526, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.263, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44066, + "right_pr_number": 44154, + "code_similarity": 0.174, + "size_similarity": 0.654, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.286, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44066, + "right_pr_number": 44722, + "code_similarity": 0.907, + "size_similarity": 0.964, + "file_overlap": 1.0, + "area_overlap": 0.808, + "patch_similarity": 0.874, + "shared_filenames": [ + "src/transformers/models/codegen/modeling_codegen.py", + "src/transformers/models/gptj/modeling_gptj.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/models/codegen/modeling_codegen.py", + "left_ranges": [ + [ + 245, + 266 + ] + ], + "right_ranges": [ + [ + 228, + 234 + ] + ] + }, + { + "filename": "src/transformers/models/gptj/modeling_gptj.py", + "left_ranges": [ + [ + 33, + 41 + ], + [ + 174, + 181 + ], + [ + 250, + 257 + ], + [ + 398, + 419 + ], + [ + 425, + 434 + ], + [ + 459, + 466 + ], + [ + 471, + 488 + ], + [ + 518, + 540 + ], + [ + 554, + 560 + ], + [ + 566, + 575 + ], + [ + 580, + 598 + ], + [ + 601, + 606 + ], + [ + 634, + 640 + ], + [ + 646, + 653 + ], + [ + 658, + 674 + ], + [ + 717, + 722 + ], + [ + 738, + 744 + ], + [ + 749, + 772 + ], + [ + 790, + 795 + ] + ], + "right_ranges": [ + [ + 33, + 42 + ], + [ + 175, + 182 + ], + [ + 251, + 258 + ], + [ + 399, + 420 + ], + [ + 426, + 435 + ], + [ + 460, + 467 + ], + [ + 472, + 489 + ], + [ + 519, + 540 + ], + [ + 554, + 560 + ], + [ + 566, + 575 + ], + [ + 580, + 598 + ], + [ + 601, + 612 + ], + [ + 634, + 640 + ], + [ + 646, + 653 + ], + [ + 658, + 663 + ], + [ + 666, + 674 + ], + [ + 717, + 722 + ], + [ + 738, + 744 + ], + [ + 749, + 772 + ], + [ + 790, + 795 + ] + ] + } + ] + }, + { + "left_pr_number": 44072, + "right_pr_number": 44085, + "code_similarity": 0.137, + "size_similarity": 0.387, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.398, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44072, + "right_pr_number": 44129, + "code_similarity": 0.074, + "size_similarity": 0.195, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.231, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44072, + "right_pr_number": 44154, + "code_similarity": 0.175, + "size_similarity": 0.566, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.414, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44072, + "right_pr_number": 44722, + "code_similarity": 0.124, + "size_similarity": 0.357, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.347, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44085, + "right_pr_number": 44129, + "code_similarity": 0.141, + "size_similarity": 0.503, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.272, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44085, + "right_pr_number": 44154, + "code_similarity": 0.18, + "size_similarity": 0.684, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.289, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44085, + "right_pr_number": 44722, + "code_similarity": 0.728, + "size_similarity": 0.923, + "file_overlap": 0.5, + "area_overlap": 0.791, + "patch_similarity": 0.78, + "shared_filenames": [ + "src/transformers/models/gptj/modeling_gptj.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/models/gptj/modeling_gptj.py", + "left_ranges": [ + [ + 33, + 40 + ], + [ + 173, + 180 + ], + [ + 249, + 256 + ], + [ + 397, + 405 + ], + [ + 408, + 419 + ], + [ + 425, + 434 + ], + [ + 459, + 465 + ], + [ + 470, + 489 + ], + [ + 517, + 539 + ], + [ + 553, + 559 + ], + [ + 565, + 574 + ], + [ + 579, + 597 + ], + [ + 600, + 611 + ], + [ + 633, + 639 + ], + [ + 645, + 652 + ], + [ + 657, + 673 + ], + [ + 716, + 728 + ], + [ + 737, + 743 + ], + [ + 748, + 771 + ], + [ + 789, + 794 + ] + ], + "right_ranges": [ + [ + 33, + 42 + ], + [ + 175, + 182 + ], + [ + 251, + 258 + ], + [ + 399, + 420 + ], + [ + 426, + 435 + ], + [ + 460, + 467 + ], + [ + 472, + 489 + ], + [ + 519, + 540 + ], + [ + 554, + 560 + ], + [ + 566, + 575 + ], + [ + 580, + 598 + ], + [ + 601, + 612 + ], + [ + 634, + 640 + ], + [ + 646, + 653 + ], + [ + 658, + 663 + ], + [ + 666, + 674 + ], + [ + 717, + 722 + ], + [ + 738, + 744 + ], + [ + 749, + 772 + ], + [ + 790, + 795 + ] + ] + } + ] + }, + { + "left_pr_number": 44129, + "right_pr_number": 44154, + "code_similarity": 0.099, + "size_similarity": 0.344, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.199, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44129, + "right_pr_number": 44722, + "code_similarity": 0.146, + "size_similarity": 0.545, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.247, + "shared_filenames": [], + "shared_file_areas": [] + }, + { + "left_pr_number": 44154, + "right_pr_number": 44722, + "code_similarity": 0.164, + "size_similarity": 0.631, + "file_overlap": 0.0, + "area_overlap": 0.0, + "patch_similarity": 0.255, + "shared_filenames": [], + "shared_file_areas": [] + } + ] + }, + { + "cluster_id": "cluster-41211-3", + "summary": "Cluster of 1 issues and 2 PRs centered on issue #41211.", + "status": "open", + "confidence": 0.55, + "canonical_issue_number": 41211, + "canonical_pr_number": 44339, + "issue_numbers": [ + 41211 + ], + "pr_numbers": [ + 41356, + 44339 + ], + "evidence_types": [ + "closing_reference" + ], + "pr_comparisons": [ + { + "left_pr_number": 41356, + "right_pr_number": 44339, + "code_similarity": 0.155, + "size_similarity": 0.094, + "file_overlap": 0.273, + "area_overlap": 0.078, + "patch_similarity": 0.183, + "shared_filenames": [ + "docs/source/en/model_doc/deimv2.md", + "src/transformers/models/deimv2/__init__.py", + "src/transformers/models/deimv2/configuration_deimv2.py", + "src/transformers/models/deimv2/modeling_deimv2.py", + "tests/models/deimv2/__init__.py", + "tests/models/deimv2/test_modeling_deimv2.py" + ], + "shared_file_areas": [ + { + "filename": "docs/source/en/model_doc/deimv2.md", + "left_ranges": [ + [ + 1, + 132 + ] + ], + "right_ranges": [ + [ + 1, + 65 + ] + ] + }, + { + "filename": "src/transformers/models/deimv2/__init__.py", + "left_ranges": [ + [ + 1, + 15 + ] + ], + "right_ranges": [ + [ + 1, + 29 + ] + ] + }, + { + "filename": "src/transformers/models/deimv2/configuration_deimv2.py", + "left_ranges": [ + [ + 1, + 74 + ] + ], + "right_ranges": [ + [ + 1, + 266 + ] + ] + }, + { + "filename": "tests/models/deimv2/test_modeling_deimv2.py", + "left_ranges": [ + [ + 1, + 15 + ] + ], + "right_ranges": [ + [ + 1, + 1734 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-43656-4", + "summary": "Cluster of 1 issues and 3 PRs centered on issue #43824.", + "status": "open", + "confidence": 0.8, + "canonical_issue_number": 43824, + "canonical_pr_number": 43656, + "issue_numbers": [ + 43824 + ], + "pr_numbers": [ + 43656, + 43836, + 43842 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target", + "soft_similarity" + ], + "pr_comparisons": [ + { + "left_pr_number": 43656, + "right_pr_number": 43836, + "code_similarity": 0.36, + "size_similarity": 0.176, + "file_overlap": 1.0, + "area_overlap": 0.051, + "patch_similarity": 0.048, + "shared_filenames": [ + "src/transformers/cli/serve.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/cli/serve.py", + "left_ranges": [ + [ + 11, + 18 + ], + [ + 30, + 36 + ], + [ + 315, + 323 + ], + [ + 665, + 671 + ], + [ + 931, + 937 + ], + [ + 1843, + 1849 + ], + [ + 1868, + 1874 + ] + ], + "right_ranges": [ + [ + 11, + 18 + ], + [ + 359, + 429 + ], + [ + 584, + 590 + ], + [ + 1892, + 1910 + ], + [ + 1917, + 1923 + ] + ] + } + ] + }, + { + "left_pr_number": 43656, + "right_pr_number": 43842, + "code_similarity": 0.405, + "size_similarity": 0.5, + "file_overlap": 1.0, + "area_overlap": 0.0, + "patch_similarity": 0.036, + "shared_filenames": [ + "src/transformers/cli/serve.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/cli/serve.py", + "left_ranges": [ + [ + 11, + 18 + ], + [ + 30, + 36 + ], + [ + 315, + 323 + ], + [ + 665, + 671 + ], + [ + 931, + 937 + ], + [ + 1843, + 1849 + ], + [ + 1868, + 1874 + ] + ], + "right_ranges": [ + [ + 54, + 61 + ], + [ + 587, + 593 + ] + ] + } + ] + }, + { + "left_pr_number": 43836, + "right_pr_number": 43842, + "code_similarity": 0.332, + "size_similarity": 0.088, + "file_overlap": 1.0, + "area_overlap": 0.033, + "patch_similarity": 0.017, + "shared_filenames": [ + "src/transformers/cli/serve.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/cli/serve.py", + "left_ranges": [ + [ + 11, + 18 + ], + [ + 359, + 429 + ], + [ + 584, + 590 + ], + [ + 1892, + 1910 + ], + [ + 1917, + 1923 + ] + ], + "right_ranges": [ + [ + 54, + 61 + ], + [ + 587, + 593 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-45561-4", + "summary": "Cluster of 1 issues and 3 PRs centered on issue #45561.", + "status": "open", + "confidence": 0.8, + "canonical_issue_number": 45561, + "canonical_pr_number": 45645, + "issue_numbers": [ + 45561 + ], + "pr_numbers": [ + 45639, + 45645, + 45778 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target", + "soft_similarity" + ], + "pr_comparisons": [ + { + "left_pr_number": 45639, + "right_pr_number": 45645, + "code_similarity": 0.532, + "size_similarity": 0.619, + "file_overlap": 0.5, + "area_overlap": 0.543, + "patch_similarity": 0.451, + "shared_filenames": [ + "src/transformers/testing_utils.py", + "tests/utils/test_testing_utils.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/testing_utils.py", + "left_ranges": [ + [ + 3529, + 3536 + ], + [ + 3753, + 3787 + ] + ], + "right_ranges": [ + [ + 3525, + 3558 + ], + [ + 3782, + 3788 + ] + ] + }, + { + "filename": "tests/utils/test_testing_utils.py", + "left_ranges": [ + [ + 1, + 86 + ] + ], + "right_ranges": [ + [ + 1, + 114 + ] + ] + } + ] + }, + { + "left_pr_number": 45639, + "right_pr_number": 45778, + "code_similarity": 0.341, + "size_similarity": 0.141, + "file_overlap": 0.5, + "area_overlap": 0.386, + "patch_similarity": 0.181, + "shared_filenames": [ + "src/transformers/testing_utils.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/testing_utils.py", + "left_ranges": [ + [ + 3529, + 3536 + ], + [ + 3753, + 3787 + ] + ], + "right_ranges": [ + [ + 3529, + 3537 + ], + [ + 3762, + 3770 + ] + ] + } + ] + }, + { + "left_pr_number": 45645, + "right_pr_number": 45778, + "code_similarity": 0.175, + "size_similarity": 0.087, + "file_overlap": 0.25, + "area_overlap": 0.18, + "patch_similarity": 0.128, + "shared_filenames": [ + "src/transformers/testing_utils.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/testing_utils.py", + "left_ranges": [ + [ + 3525, + 3558 + ], + [ + 3782, + 3788 + ] + ], + "right_ranges": [ + [ + 3529, + 3537 + ], + [ + 3762, + 3770 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-43240-3", + "summary": "Cluster of 1 issues and 2 PRs centered on issue #43240.", + "status": "open", + "confidence": 0.75, + "canonical_issue_number": 43240, + "canonical_pr_number": 43251, + "issue_numbers": [ + 43240 + ], + "pr_numbers": [ + 43251, + 43254 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target" + ], + "pr_comparisons": [ + { + "left_pr_number": 43251, + "right_pr_number": 43254, + "code_similarity": 0.794, + "size_similarity": 0.64, + "file_overlap": 1.0, + "area_overlap": 0.667, + "patch_similarity": 0.882, + "shared_filenames": [ + "src/transformers/loss/loss_utils.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/loss/loss_utils.py", + "left_ranges": [ + [ + 30, + 50 + ] + ], + "right_ranges": [ + [ + 30, + 43 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-41115-3", + "summary": "Cluster of 1 issues and 2 PRs centered on issue #41115.", + "status": "open", + "confidence": 0.55, + "canonical_issue_number": 41115, + "canonical_pr_number": 45613, + "issue_numbers": [ + 41115 + ], + "pr_numbers": [ + 41116, + 45613 + ], + "evidence_types": [ + "closing_reference" + ], + "pr_comparisons": [ + { + "left_pr_number": 41116, + "right_pr_number": 45613, + "code_similarity": 0.805, + "size_similarity": 0.885, + "file_overlap": 0.75, + "area_overlap": 0.854, + "patch_similarity": 0.693, + "shared_filenames": [ + "docs/source/en/_toctree.yml", + "docs/source/en/model_doc/minicpm3.md", + "src/transformers/models/auto/modeling_auto.py", + "src/transformers/models/minicpm3/__init__.py", + "src/transformers/models/minicpm3/configuration_minicpm3.py", + "src/transformers/models/minicpm3/modeling_minicpm3.py", + "src/transformers/models/minicpm3/modular_minicpm3.py", + "tests/models/minicpm3/__init__.py", + "tests/models/minicpm3/test_modeling_minicpm3.py" + ], + "shared_file_areas": [ + { + "filename": "docs/source/en/_toctree.yml", + "left_ranges": [ + [ + 713, + 720 + ] + ], + "right_ranges": [ + [ + 713, + 720 + ] + ] + }, + { + "filename": "docs/source/en/model_doc/minicpm3.md", + "left_ranges": [ + [ + 1, + 67 + ] + ], + "right_ranges": [ + [ + 1, + 45 + ] + ] + }, + { + "filename": "src/transformers/models/auto/modeling_auto.py", + "left_ranges": [ + [ + 285, + 291 + ], + [ + 700, + 706 + ], + [ + 1298, + 1304 + ] + ], + "right_ranges": [ + [ + 699, + 705 + ], + [ + 1301, + 1307 + ] + ] + }, + { + "filename": "src/transformers/models/minicpm3/__init__.py", + "left_ranges": [ + [ + 1, + 29 + ] + ], + "right_ranges": [ + [ + 1, + 29 + ] + ] + }, + { + "filename": "src/transformers/models/minicpm3/configuration_minicpm3.py", + "left_ranges": [ + [ + 1, + 141 + ] + ], + "right_ranges": [ + [ + 1, + 126 + ] + ] + }, + { + "filename": "src/transformers/models/minicpm3/modeling_minicpm3.py", + "left_ranges": [ + [ + 1, + 544 + ] + ], + "right_ranges": [ + [ + 1, + 522 + ] + ] + }, + { + "filename": "src/transformers/models/minicpm3/modular_minicpm3.py", + "left_ranges": [ + [ + 1, + 444 + ] + ], + "right_ranges": [ + [ + 1, + 342 + ] + ] + }, + { + "filename": "tests/models/minicpm3/test_modeling_minicpm3.py", + "left_ranges": [ + [ + 1, + 109 + ] + ], + "right_ranges": [ + [ + 1, + 136 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-45647-3", + "summary": "Cluster of 1 issues and 2 PRs centered on issue #45647.", + "status": "open", + "confidence": 0.75, + "canonical_issue_number": 45647, + "canonical_pr_number": 45738, + "issue_numbers": [ + 45647 + ], + "pr_numbers": [ + 45737, + 45738 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target" + ], + "pr_comparisons": [ + { + "left_pr_number": 45737, + "right_pr_number": 45738, + "code_similarity": 0.526, + "size_similarity": 0.11, + "file_overlap": 0.5, + "area_overlap": 1.0, + "patch_similarity": 0.024, + "shared_filenames": [ + "src/transformers/models/musicgen_melody/modeling_musicgen_melody.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/models/musicgen_melody/modeling_musicgen_melody.py", + "left_ranges": [ + [ + 491, + 497 + ] + ], + "right_ranges": [ + [ + 491, + 497 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-45081-3", + "summary": "Cluster of 1 issues and 2 PRs centered on issue #45081.", + "status": "open", + "confidence": 0.8, + "canonical_issue_number": 45081, + "canonical_pr_number": 45317, + "issue_numbers": [ + 45081 + ], + "pr_numbers": [ + 45086, + 45317 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target", + "soft_similarity" + ], + "pr_comparisons": [ + { + "left_pr_number": 45086, + "right_pr_number": 45317, + "code_similarity": 0.589, + "size_similarity": 0.34, + "file_overlap": 0.5, + "area_overlap": 1.0, + "patch_similarity": 0.136, + "shared_filenames": [ + "src/transformers/tokenization_utils_tokenizers.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/tokenization_utils_tokenizers.py", + "left_ranges": [ + [ + 1360, + 1370 + ], + [ + 1374, + 1380 + ] + ], + "right_ranges": [ + [ + 1360, + 1370 + ], + [ + 1374, + 1380 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-43698-3", + "summary": "Cluster of 1 issues and 2 PRs centered on issue #43698.", + "status": "open", + "confidence": 0.75, + "canonical_issue_number": 43698, + "canonical_pr_number": 43779, + "issue_numbers": [ + 43698 + ], + "pr_numbers": [ + 43779, + 43816 + ], + "evidence_types": [ + "closing_reference", + "shared_issue_target" + ], + "pr_comparisons": [ + { + "left_pr_number": 43779, + "right_pr_number": 43816, + "code_similarity": 0.418, + "size_similarity": 0.538, + "file_overlap": 1.0, + "area_overlap": 0.02, + "patch_similarity": 0.02, + "shared_filenames": [ + "src/transformers/integrations/integration_utils.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/integrations/integration_utils.py", + "left_ranges": [ + [ + 2227, + 2241 + ], + [ + 2303, + 2309 + ] + ], + "right_ranges": [ + [ + 2278, + 2291 + ], + [ + 2309, + 2322 + ] + ] + } + ] + } + ] + }, + { + "cluster_id": "cluster-44018-2", + "summary": "Cluster of 2 related pull requests linked by soft_similarity.", + "status": "open", + "confidence": 0.5, + "canonical_issue_number": null, + "canonical_pr_number": 44068, + "issue_numbers": [], + "pr_numbers": [ + 44018, + 44068 + ], + "evidence_types": [ + "soft_similarity" + ], + "pr_comparisons": [ + { + "left_pr_number": 44018, + "right_pr_number": 44068, + "code_similarity": 0.766, + "size_similarity": 0.939, + "file_overlap": 1.0, + "area_overlap": 0.425, + "patch_similarity": 0.866, + "shared_filenames": [ + "src/transformers/models/gpt_neo/modeling_gpt_neo.py", + "tests/models/gpt_neo/test_modeling_gpt_neo.py" + ], + "shared_file_areas": [ + { + "filename": "src/transformers/models/gpt_neo/modeling_gpt_neo.py", + "left_ranges": [ + [ + 26, + 46 + ], + [ + 138, + 143 + ], + [ + 182, + 187 + ], + [ + 283, + 295 + ], + [ + 327, + 341 + ], + [ + 348, + 354 + ], + [ + 360, + 369 + ], + [ + 399, + 406 + ], + [ + 411, + 419 + ], + [ + 428, + 433 + ], + [ + 472, + 492 + ], + [ + 507, + 513 + ], + [ + 519, + 528 + ], + [ + 541, + 559 + ], + [ + 562, + 567 + ], + [ + 595, + 601 + ], + [ + 607, + 614 + ], + [ + 627, + 643 + ], + [ + 685, + 690 + ], + [ + 708, + 714 + ], + [ + 720, + 727 + ], + [ + 740, + 757 + ], + [ + 761, + 766 + ], + [ + 780, + 786 + ], + [ + 791, + 798 + ], + [ + 807, + 822 + ], + [ + 840, + 845 + ] + ], + "right_ranges": [ + [ + 26, + 31 + ], + [ + 34, + 44 + ], + [ + 136, + 141 + ], + [ + 180, + 185 + ], + [ + 281, + 293 + ], + [ + 325, + 339 + ], + [ + 346, + 352 + ], + [ + 358, + 367 + ], + [ + 397, + 404 + ], + [ + 409, + 417 + ], + [ + 426, + 434 + ], + [ + 463, + 483 + ], + [ + 498, + 504 + ], + [ + 510, + 519 + ], + [ + 532, + 550 + ], + [ + 553, + 564 + ], + [ + 586, + 592 + ], + [ + 598, + 605 + ], + [ + 618, + 634 + ], + [ + 676, + 688 + ], + [ + 699, + 705 + ], + [ + 711, + 718 + ], + [ + 731, + 748 + ], + [ + 752, + 762 + ], + [ + 771, + 777 + ], + [ + 782, + 789 + ], + [ + 798, + 813 + ], + [ + 831, + 836 + ] + ] + }, + { + "filename": "tests/models/gpt_neo/test_modeling_gpt_neo.py", + "left_ranges": [ + [ + 458, + 464 + ] + ], + "right_ranges": [ + [ + 458, + 464 + ] + ] + } + ] + } + ] + } + ], + "duplicate_issues": [], + "duplicate_prs": [ + { + "cluster_id": "cluster-41115-3", + "canonical_pr_number": 45613, + "duplicate_pr_numbers": [ + 41116 + ], + "target_issue_number": 41115, + "reason": "PRs in cluster-41115-3 are treated as duplicates because they converge on issue #41115 with closing_reference evidence." + }, + { + "cluster_id": "cluster-41211-3", + "canonical_pr_number": 44339, + "duplicate_pr_numbers": [ + 41356 + ], + "target_issue_number": 41211, + "reason": "PRs in cluster-41211-3 are treated as duplicates because they converge on issue #41211 with closing_reference evidence." + }, + { + "cluster_id": "cluster-43240-3", + "canonical_pr_number": 43251, + "duplicate_pr_numbers": [ + 43254 + ], + "target_issue_number": 43240, + "reason": "PRs in cluster-43240-3 are treated as duplicates because they converge on issue #43240 with closing_reference, shared_issue_target evidence." + }, + { + "cluster_id": "cluster-43656-4", + "canonical_pr_number": 43656, + "duplicate_pr_numbers": [ + 43836, + 43842 + ], + "target_issue_number": 43824, + "reason": "PRs in cluster-43656-4 are treated as duplicates because they converge on issue #43824 with closing_reference, shared_issue_target, soft_similarity evidence." + }, + { + "cluster_id": "cluster-43698-3", + "canonical_pr_number": 43779, + "duplicate_pr_numbers": [ + 43816 + ], + "target_issue_number": 43698, + "reason": "PRs in cluster-43698-3 are treated as duplicates because they converge on issue #43698 with closing_reference, shared_issue_target evidence." + }, + { + "cluster_id": "cluster-43979-11", + "canonical_pr_number": 44007, + "duplicate_pr_numbers": [ + 43996, + 44013, + 44044, + 44066, + 44072, + 44085, + 44129, + 44154, + 44722 + ], + "target_issue_number": 43979, + "reason": "PRs in cluster-43979-11 are treated as duplicates because they converge on issue #43979 with closing_reference, shared_issue_target, soft_similarity evidence." + }, + { + "cluster_id": "cluster-44018-2", + "canonical_pr_number": 44068, + "duplicate_pr_numbers": [ + 44018 + ], + "target_issue_number": null, + "reason": "PRs in cluster-44018-2 are treated as duplicates because they share soft_similarity evidence." + }, + { + "cluster_id": "cluster-45081-3", + "canonical_pr_number": 45317, + "duplicate_pr_numbers": [ + 45086 + ], + "target_issue_number": 45081, + "reason": "PRs in cluster-45081-3 are treated as duplicates because they converge on issue #45081 with closing_reference, shared_issue_target, soft_similarity evidence." + }, + { + "cluster_id": "cluster-45561-4", + "canonical_pr_number": 45645, + "duplicate_pr_numbers": [ + 45639, + 45778 + ], + "target_issue_number": 45561, + "reason": "PRs in cluster-45561-4 are treated as duplicates because they converge on issue #45561 with closing_reference, shared_issue_target, soft_similarity evidence." + }, + { + "cluster_id": "cluster-45647-3", + "canonical_pr_number": 45738, + "duplicate_pr_numbers": [ + 45737 + ], + "target_issue_number": 45647, + "reason": "PRs in cluster-45647-3 are treated as duplicates because they converge on issue #45647 with closing_reference, shared_issue_target evidence." + } + ], + "best_issue": { + "cluster_id": "cluster-45561-4", + "issue_number": 45561, + "reason": "Issue #45561 is the strongest global issue candidate because it is open, belongs to a cluster with 4 artifacts, and carries 2 discussion comments plus 8 inbound references.", + "score": 152.0 + }, + "best_pr": { + "cluster_id": "cluster-43979-11", + "pr_number": 44007, + "reason": "PR #44007 is the strongest global PR candidate because it is open, belongs to a cluster with 11 artifacts, links to 1 issues, and carries 2 review/discussion events.", + "score": 238.886 + } +} diff --git a/snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/analysis-report-hybrid.llm-reviews.json b/snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/analysis-report-hybrid.llm-reviews.json new file mode 100644 index 0000000000000000000000000000000000000000..0dd584408f8186d1cb732c474be4fe725b9ffaeb --- /dev/null +++ b/snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/analysis-report-hybrid.llm-reviews.json @@ -0,0 +1,18547 @@ +{ + "schema_version": "1.0", + "repo": "huggingface/transformers", + "snapshot_id": "20260505T060044Z", + "generated_at": "2026-05-05T06:09:27Z", + "model": "gpt-5.4-mini?service_tier=flex", + "reviews": [ + { + "label": "issue", + "component_index": 1, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-13244-3", + "nodes": [ + "issue:13244", + "issue:15980", + "issue:16618" + ], + "soft_pairs": [ + "issue:13244|issue:16618", + "issue:15980|issue:16618" + ], + "prepared_review_unit_hash": "d56c5d5a3f11345f515d7f21dbf2c83ed36190b41540e76cc58fa67ffd9340ba", + "component_budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 2, + "serialized_chars": 2478, + "estimated_input_tokens": 620, + "estimated_eval_tokens": 1496 + }, + "budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 2, + "serialized_chars": 2478, + "estimated_input_tokens": 620, + "estimated_eval_tokens": 1496 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These issues do not appear to be duplicates: one is about TAPAS tokenization parity, one about an unauthenticated private-model download error, and one about failing to load bert-base-uncased. The cluster is heterogeneous rather than a single underlying bug.", + "confidence": 0.97, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:13244", + "right": "issue:16618", + "accept": false, + "reason": "Different problems: TAPAS tokenization mismatch vs. model loading failure for bert-base-uncased." + }, + { + "left": "issue:15980", + "right": "issue:16618", + "accept": false, + "reason": "Different code paths: private-model auth/download error vs. a general model-load failure." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:24643", + "issue:30277", + "issue:34634", + "issue:34928", + "issue:35545", + "issue:36010", + "issue:36331", + "issue:39290", + "issue:41093", + "issue:41720", + "issue:41762", + "issue:42915", + "issue:43039", + "issue:43404", + "issue:43425", + "issue:43531", + "issue:43541", + "issue:43596", + "issue:43638", + "issue:43716", + "issue:43792", + "issue:43827", + "issue:43828", + "issue:43854", + "issue:43866", + "issue:43901", + "issue:43975", + "issue:44292", + "issue:44322", + "issue:44457", + "issue:44509", + "issue:44512", + "issue:44560", + "issue:44661", + "issue:44805", + "issue:44841", + "issue:44863", + "issue:44918", + "issue:45070", + "issue:45081", + "issue:45084", + "issue:45161", + "issue:45237", + "issue:45362", + "issue:45464", + "issue:45507", + "issue:45685" + ], + "soft_pairs": [ + "issue:44322|issue:45464", + "issue:45237|issue:45685", + "issue:43716|issue:45237", + "issue:43828|issue:45237", + "issue:45237|issue:45507", + "issue:44292|issue:45237", + "issue:30277|issue:35545", + "issue:35545|issue:42915", + "issue:35545|issue:41720", + "issue:44509|issue:44512", + "issue:35545|issue:36010", + "issue:43638|issue:44805", + "issue:43596|issue:44805", + "issue:44805|issue:45161", + "issue:44805|issue:44918", + "issue:44661|issue:44805", + "issue:34634|issue:35545", + "issue:30277|issue:43638", + "issue:43596|issue:43638", + "issue:43975|issue:44457", + "issue:43039|issue:45161", + "issue:44292|issue:44457", + "issue:34928|issue:44805", + "issue:24643|issue:30277", + "issue:43638|issue:45161", + "issue:34928|issue:36331", + "issue:43596|issue:45161", + "issue:43716|issue:45507", + "issue:39290|issue:45081", + "issue:43828|issue:45161", + "issue:43425|issue:44292", + "issue:44918|issue:45161", + "issue:34928|issue:41720", + "issue:43716|issue:45685", + "issue:43541|issue:43716", + "issue:43828|issue:44292", + "issue:45507|issue:45685", + "issue:43716|issue:44292", + "issue:44292|issue:44918", + "issue:43039|issue:44918", + "issue:41762|issue:43638", + "issue:43827|issue:44512", + "issue:43716|issue:44560", + "issue:43828|issue:45685", + "issue:44560|issue:45685", + "issue:43039|issue:43828", + "issue:30277|issue:36331", + "issue:43039|issue:44805", + "issue:34928|issue:36010", + "issue:43901|issue:44512", + "issue:41762|issue:43596", + "issue:39290|issue:41720", + "issue:43039|issue:44560", + "issue:43866|issue:44863", + "issue:43854|issue:43866", + "issue:39290|issue:43531", + "issue:41093|issue:43638", + "issue:43425|issue:45070", + "issue:45362|issue:45464", + "issue:43792|issue:45070", + "issue:43404|issue:43866", + "issue:45464|issue:45685", + "issue:43541|issue:45685", + "issue:43541|issue:43828", + "issue:36010|issue:39290", + "issue:44841|issue:45084", + "issue:41093|issue:41762" + ], + "prepared_review_unit_hash": "afb0273794e45991c23982a6b35e3342e60e0aa4b344c67203d5ae6ad741fae4", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 67, + "serialized_chars": 47120, + "estimated_input_tokens": 11780, + "estimated_eval_tokens": 23816 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This set is mostly a grab bag of unrelated model/runtime/docs issues. The only clear duplicate-looking pair is the ZeRO-3/BERT size-0 IndexError reports.", + "confidence": 0.79, + "canonical_issue_reason": "issue:43596 is the clearest representative of the one plausible duplicate thread here (ZeRO-3 + BERT + size-0 IndexError), and it is the more general/earlier report.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43596 best captures the only strong duplicate cluster in this batch and is the cleanest canonical issue candidate.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44322", + "right": "issue:45464", + "accept": false, + "reason": "Different Qwen3.5 failures: config attribute error vs streaming chat/completions failure." + }, + { + "left": "issue:45237", + "right": "issue:45685", + "accept": false, + "reason": "Different hardware/backend problems on different models; not the same bug." + }, + { + "left": "issue:43716", + "right": "issue:45237", + "accept": false, + "reason": "Mistral-3 dtype mismatch with image preprocessing is unrelated to GPT-OSS on AMD GPUs." + }, + { + "left": "issue:43828", + "right": "issue:45237", + "accept": false, + "reason": "Phi-tiny-MoE autocast dtype mismatch is a different code path from GPT-OSS AMD GPU failures." + }, + { + "left": "issue:45237", + "right": "issue:45507", + "accept": false, + "reason": "AMD GPU runtime issue vs GraniteMoEHybrid invalid method; no shared underlying defect." + }, + { + "left": "issue:44292", + "right": "issue:45237", + "accept": false, + "reason": "Qwen-3-8B-NVFP4 error and GPT-OSS AMD GPU error are unrelated model/runtime issues." + }, + { + "left": "issue:30277", + "right": "issue:35545", + "accept": false, + "reason": "DeepSpeed collective mismatch during training vs ModernBERT ONNX export are unrelated." + }, + { + "left": "issue:35545", + "right": "issue:42915", + "accept": false, + "reason": "ONNX export failure and Qwen3Moe FP8 failure do not look like the same bug." + }, + { + "left": "issue:35545", + "right": "issue:41720", + "accept": false, + "reason": "ModernBERT export problem and Qwen3 auto-device-map cuda assert are different failures." + }, + { + "left": "issue:44509", + "right": "issue:44512", + "accept": false, + "reason": "Both are v5 docs cleanup items, but they reference different removed docs content and are not the same change." + }, + { + "left": "issue:35545", + "right": "issue:36010", + "accept": false, + "reason": "ModernBERT export and GenerationMixin import breakage are unrelated." + }, + { + "left": "issue:43638", + "right": "issue:44805", + "accept": false, + "reason": "Both are IndexError reports, but the titles point to different concrete failures and models." + }, + { + "left": "issue:43596", + "right": "issue:44805", + "accept": false, + "reason": "Same broad ZeRO-3/indexing area, but the reported error and symptom differ too much to merge." + }, + { + "left": "issue:44805", + "right": "issue:45161", + "accept": false, + "reason": "Mask shape mismatch is unrelated to GPT-OSS tensor-parallel failure." + }, + { + "left": "issue:44805", + "right": "issue:44918", + "accept": false, + "reason": "Different model/trainer failures; no evidence of a shared code path." + }, + { + "left": "issue:44661", + "right": "issue:44805", + "accept": false, + "reason": "Tokenizer-mapping setup bug is unrelated to the mask/indexing error." + }, + { + "left": "issue:34634", + "right": "issue:35545", + "accept": false, + "reason": "BarkProcessor voice_preset issue is unrelated to ModernBERT ONNX export." + }, + { + "left": "issue:30277", + "right": "issue:43638", + "accept": false, + "reason": "Both involve DeepSpeed, but one is collective mismatch and the other is Bert zero-3 indexing; different bugs." + }, + { + "left": "issue:43596", + "right": "issue:43638", + "accept": true, + "reason": "Near-identical ZeRO-3 + BERT + size-0 IndexError reports; same underlying bug and plausible single fix." + }, + { + "left": "issue:43975", + "right": "issue:44457", + "accept": false, + "reason": "DeepSeek detokenization bug and LoRA save/reload inconsistency are different problems." + }, + { + "left": "issue:43039", + "right": "issue:45161", + "accept": false, + "reason": "Liger Kernel cross-entropy routing and GPT-OSS tensor-parallel failure are unrelated." + }, + { + "left": "issue:44292", + "right": "issue:44457", + "accept": false, + "reason": "Qwen NVFP4 runtime issue and LoRA merge/save mismatch are not the same defect." + }, + { + "left": "issue:34928", + "right": "issue:44805", + "accept": false, + "reason": "Activation checkpointing/FSDP recompute mismatch is unrelated to the mask shape IndexError." + }, + { + "left": "issue:24643", + "right": "issue:30277", + "accept": false, + "reason": "DeepSpeed weight-dimension error and collective mismatch are different training failures." + }, + { + "left": "issue:43638", + "right": "issue:45161", + "accept": false, + "reason": "ZeRO-3/BERT index error is unrelated to GPT-OSS tensor-parallel not working." + }, + { + "left": "issue:43716", + "right": "issue:45507", + "accept": false, + "reason": "Different model families and different failure modes." + }, + { + "left": "issue:39290", + "right": "issue:45081", + "accept": false, + "reason": "Gemma3 sliding_window_pattern and Mistral tokenizer regex patch are unrelated." + }, + { + "left": "issue:43828", + "right": "issue:45161", + "accept": false, + "reason": "Autocast dtype mismatch is not the same as GPT-OSS TP failure." + }, + { + "left": "issue:43425", + "right": "issue:44292", + "accept": false, + "reason": "Torch version compatibility and Qwen NVFP4 runtime error are different issues." + }, + { + "left": "issue:44918", + "right": "issue:45161", + "accept": false, + "reason": "Qwen3.5 embedding unpacking under TRL is unrelated to GPT-OSS TP behavior." + }, + { + "left": "issue:34928", + "right": "issue:41720", + "accept": false, + "reason": "FSDP checkpointing mismatch and Qwen3 auto-device-map assert are unrelated." + }, + { + "left": "issue:43716", + "right": "issue:45685", + "accept": false, + "reason": "Mistral-3 dtype mismatch and MPS histogram kernel error are different backend bugs." + }, + { + "left": "issue:43541", + "right": "issue:43716", + "accept": false, + "reason": "Mixtral grouped_mm tracing error is unrelated to Mistral-3 image preprocessing dtype mismatch." + }, + { + "left": "issue:43828", + "right": "issue:44292", + "accept": false, + "reason": "Both are model runtime errors, but the model and failure mechanism differ." + }, + { + "left": "issue:43039", + "right": "issue:44918", + "accept": false, + "reason": "Liger Kernel cross-entropy routing and Qwen3.5 embedding unpacking are unrelated." + }, + { + "left": "issue:41762", + "right": "issue:43638", + "accept": false, + "reason": "Both involve ZeRO-3, but Gemma3 loading and Bert loading failures are not the same concrete bug." + }, + { + "left": "issue:43827", + "right": "issue:44512", + "accept": false, + "reason": "Both are docs issues, but they concern different removed v5 references and should stay separate." + }, + { + "left": "issue:43716", + "right": "issue:44560", + "accept": false, + "reason": "Mistral-3 dtype mismatch and Qwen3-vl StopIteration are unrelated." + }, + { + "left": "issue:43828", + "right": "issue:45685", + "accept": false, + "reason": "Autocast dtype mismatch and MPS histogram kernel error are different backend issues." + }, + { + "left": "issue:44560", + "right": "issue:45685", + "accept": false, + "reason": "Video embedding StopIteration and MPS Int histogram support are unrelated." + }, + { + "left": "issue:43039", + "right": "issue:43828", + "accept": false, + "reason": "Different model/backend failures with no shared code-path evidence." + }, + { + "left": "issue:30277", + "right": "issue:36331", + "accept": false, + "reason": "DeepSpeed collective mismatch and CustomTrainer signature breakage are unrelated." + }, + { + "left": "issue:43039", + "right": "issue:44805", + "accept": false, + "reason": "Liger Kernel cross-entropy issue is unrelated to the mask/indexing IndexError." + }, + { + "left": "issue:34928", + "right": "issue:36010", + "accept": false, + "reason": "Activation checkpointing/FSDP mismatch and GenerationMixin import error are unrelated." + }, + { + "left": "issue:43901", + "right": "issue:44512", + "accept": false, + "reason": "Different documentation topics and different removed behaviors." + }, + { + "left": "issue:41762", + "right": "issue:43596", + "accept": false, + "reason": "Same DeepSpeed/size-0 theme, but Gemma3 zero-3 loading and Bert zero-3 loading are not the same bug." + }, + { + "left": "issue:39290", + "right": "issue:41720", + "accept": false, + "reason": "Gemma3 sliding_window_pattern and Qwen3 auto device mapping are different defects." + }, + { + "left": "issue:43039", + "right": "issue:44560", + "accept": false, + "reason": "Liger Kernel loss routing and Qwen3-vl video StopIteration are unrelated." + }, + { + "left": "issue:43866", + "right": "issue:44863", + "accept": false, + "reason": "Checkpoint corruption and NemotronH checkpoint loading are not the same underlying issue." + }, + { + "left": "issue:43854", + "right": "issue:43866", + "accept": false, + "reason": "Unit-test load failure and checkpoint corruption are different problems." + }, + { + "left": "issue:39290", + "right": "issue:43531", + "accept": false, + "reason": "Both mention Qwen3-related sliding-window behavior, but the titles point to different concrete failures." + }, + { + "left": "issue:41093", + "right": "issue:43638", + "accept": false, + "reason": "Mask/tensor shape mismatch and Bert zero-3 size-0 index error are not clearly the same bug." + }, + { + "left": "issue:43425", + "right": "issue:45070", + "accept": false, + "reason": "Torch version incompatibility and pydantic PretrainedConfig field breakage are unrelated." + }, + { + "left": "issue:45362", + "right": "issue:45464", + "accept": false, + "reason": "Qwen3.5 chat crash and Qwen3.5 streaming inference failure are different symptoms with no evidence of one root cause." + }, + { + "left": "issue:43792", + "right": "issue:45070", + "accept": false, + "reason": "Whisper load failure and pydantic config breakage are unrelated." + }, + { + "left": "issue:43404", + "right": "issue:43866", + "accept": false, + "reason": "Mistral3 untied lm_head weight and Ovis2 checkpoint corruption are different defects." + }, + { + "left": "issue:45464", + "right": "issue:45685", + "accept": false, + "reason": "Qwen3.5 streaming inference failure is unrelated to the MPS histogram kernel issue." + }, + { + "left": "issue:43541", + "right": "issue:45685", + "accept": false, + "reason": "Mixtral dynamo-tracing grouped_mm error and MPS Int histogram error are unrelated." + }, + { + "left": "issue:43541", + "right": "issue:43828", + "accept": false, + "reason": "Grouped_mm tracing failure and autocast dtype mismatch are different bugs." + }, + { + "left": "issue:36010", + "right": "issue:39290", + "accept": false, + "reason": "GenerationMixin import breakage and Gemma3 sliding_window_pattern issue are unrelated." + }, + { + "left": "issue:44841", + "right": "issue:45084", + "accept": false, + "reason": "Processor load failure and template compilation error are different code paths." + }, + { + "left": "issue:41093", + "right": "issue:41762", + "accept": false, + "reason": "Both are indexing/loading errors, but the specific symptoms and models differ too much." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The assessment is grounded in the packet and stays conservative. The one accepted duplicate pair, issue:43596 and issue:43638, is well supported by the near-identical ZeRO-3/BERT size-0 IndexError titles and timestamps, and the rejection of the rest avoids overclaiming from broad thematic similarity alone. The summary is appropriately cautious and does not overstate evidence." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 2, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:39186", + "issue:39290", + "issue:39692", + "issue:40990", + "issue:41093", + "issue:42915", + "issue:43054", + "issue:43296", + "issue:43366", + "issue:43404", + "issue:43531", + "issue:43541", + "issue:43572", + "issue:43596", + "issue:43644", + "issue:43645", + "issue:43716", + "issue:43742", + "issue:43828", + "issue:43866", + "issue:43873", + "issue:43950", + "issue:43975", + "issue:43994", + "issue:44291", + "issue:44292", + "issue:44360", + "issue:44387", + "issue:44403", + "issue:44492", + "issue:44512", + "issue:44530", + "issue:44534", + "issue:44589", + "issue:44863", + "issue:44918", + "issue:44928", + "issue:44960", + "issue:45005", + "issue:45070", + "issue:45161", + "issue:45399", + "issue:45464", + "issue:45507", + "issue:45632", + "issue:45685", + "issue:45698" + ], + "soft_pairs": [ + "issue:43296|issue:43366", + "issue:43572|issue:45070", + "issue:44387|issue:45005", + "issue:43541|issue:45161", + "issue:43828|issue:45464", + "issue:44291|issue:45070", + "issue:43366|issue:45464", + "issue:43531|issue:45070", + "issue:39290|issue:43742", + "issue:43541|issue:44589", + "issue:41093|issue:43596", + "issue:43054|issue:43994", + "issue:43716|issue:44918", + "issue:44918|issue:45685", + "issue:45632|issue:45698", + "issue:43404|issue:45005", + "issue:43645|issue:45399", + "issue:40990|issue:44960", + "issue:44530|issue:45005", + "issue:42915|issue:45399", + "issue:44360|issue:44512", + "issue:44863|issue:45005", + "issue:43873|issue:45005", + "issue:39186|issue:44918", + "issue:44492|issue:44512", + "issue:44918|issue:45507", + "issue:43975|issue:44292", + "issue:43950|issue:44534", + "issue:43644|issue:43950", + "issue:44292|issue:45507", + "issue:39692|issue:43054", + "issue:43828|issue:44928", + "issue:43866|issue:44403" + ], + "prepared_review_unit_hash": "b1b1d3bd3df7e53a75860fbf5af49e93061f9dce33bf3cfd6aa2bc4aa83daf90", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 33, + "serialized_chars": 36618, + "estimated_input_tokens": 9155, + "estimated_eval_tokens": 18566 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Mostly a mixed-topic cluster rather than one duplicate set. Two clear duplicate pockets stand out: SigLIP2 bad outputs, and the v5 non-persistent-buffer corruption regression. Most other soft pairs are only broadly related by subsystem or symptom and should stay separate.", + "confidence": 0.88, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": "issue:43950 is the strongest standalone representative: it describes the concrete from_pretrained/non-persistent-buffer corruption regression clearly and is corroborated by near-duplicate wording in issue:43644 and issue:44534.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43296", + "right": "issue:43366", + "accept": false, + "reason": "Different models and different failures: PaddleOCR-VL/vLLM load error vs. gpt-oss GGUF support." + }, + { + "left": "issue:43572", + "right": "issue:45070", + "accept": false, + "reason": "One is a missing StableLmConfig field after v5 update; the other is a broader PretrainedConfig/pydantic-field regression." + }, + { + "left": "issue:44387", + "right": "issue:45005", + "accept": false, + "reason": "Int4 quantization OOM and tied-weights bugs are different code paths and fixes." + }, + { + "left": "issue:43541", + "right": "issue:45161", + "accept": false, + "reason": "Mixtral torch-dynamo grouped_mm tracing failure is unrelated to GPT-OSS tensor-parallel behavior." + }, + { + "left": "issue:43828", + "right": "issue:45464", + "accept": false, + "reason": "Autocast dtype mismatch in Phi-tiny-MoE is not the same as Qwen3.5 streaming chat/completions failure." + }, + { + "left": "issue:44291", + "right": "issue:45070", + "accept": false, + "reason": "Unexpected _is_hf_initialized argument during init_empty_weights is a different bug than the pydantic PretrainedConfig field issue." + }, + { + "left": "issue:43366", + "right": "issue:45464", + "accept": false, + "reason": "Both mention modern models, but GPT-OSS GGUF support and Qwen3.5 streaming inference are separate problems." + }, + { + "left": "issue:43531", + "right": "issue:45070", + "accept": false, + "reason": "Qwen3-MoE sliding_window handling is not the same as a PretrainedConfig field regression." + }, + { + "left": "issue:39290", + "right": "issue:43742", + "accept": false, + "reason": "Gemma3TextConfig missing attribute in vLLM and MobileLLM key-error loading are distinct model/load bugs." + }, + { + "left": "issue:43541", + "right": "issue:44589", + "accept": false, + "reason": "Torch-dynamo grouped_mm tracing failure is unrelated to the missing Float8 storage type error." + }, + { + "left": "issue:41093", + "right": "issue:43596", + "accept": false, + "reason": "Mask length mismatch during indexing is a different failure from the zero-size tensor index error with ZeRO-3/BERT." + }, + { + "left": "issue:43054", + "right": "issue:43994", + "accept": true, + "reason": "Both describe the same SigLIP2 regression: degraded/nonsensical text embeddings or outputs from the same model family, likely the same processor/model mismatch." + }, + { + "left": "issue:43716", + "right": "issue:44918", + "accept": false, + "reason": "Mistral-3 image-preprocessor dtype mismatch and Qwen3.5 SFT embedding unpacking are unrelated bugs." + }, + { + "left": "issue:44918", + "right": "issue:45685", + "accept": false, + "reason": "TRL SFT embedding unpacking and the MPS histogram_mps Int error are different failures in different stacks." + }, + { + "left": "issue:45632", + "right": "issue:45698", + "accept": false, + "reason": "Both involve custom loading, but one is a trust_remote_code cache-path collision and the other is wrong module selection after save_pretrained; not the same concrete bug." + }, + { + "left": "issue:43404", + "right": "issue:45005", + "accept": false, + "reason": "Both mention tied weights, but one is a Mistral3 lm_head tying bug and the other is a translation-models v5 regression." + }, + { + "left": "issue:43645", + "right": "issue:45399", + "accept": false, + "reason": "Notebook custom-model initialization and flash-attn fallback gating are unrelated." + }, + { + "left": "issue:40990", + "right": "issue:44960", + "accept": false, + "reason": "High perplexity on gpt-oss-20b and the GLM5 issue do not share a concrete fix path." + }, + { + "left": "issue:44530", + "right": "issue:45005", + "accept": false, + "reason": "Qwen3.5 PagedAttentionCache linear_attention crash is not the same as tied-weights behavior in translation models." + }, + { + "left": "issue:42915", + "right": "issue:45399", + "accept": false, + "reason": "FineGrainedFP8Config failure and flash-attn fallback checks are separate issues." + }, + { + "left": "issue:44360", + "right": "issue:44512", + "accept": false, + "reason": "DSA indexer ReLU discussion and a docs typo about the removed run command are unrelated." + }, + { + "left": "issue:44863", + "right": "issue:45005", + "accept": false, + "reason": "NemotronH checkpoint-loading failure is different from the tied-weights translation-model regression." + }, + { + "left": "issue:43873", + "right": "issue:45005", + "accept": false, + "reason": "Quantization/offloading behavior and tied-weights handling are not the same bug." + }, + { + "left": "issue:39186", + "right": "issue:44918", + "accept": false, + "reason": "FSDP weight dimensionality error and Qwen3.5 embedding unpacking are unrelated." + }, + { + "left": "issue:44492", + "right": "issue:44512", + "accept": false, + "reason": "A cache-strategy typo is not the same as outdated docs about the removed transformers run command." + }, + { + "left": "issue:44918", + "right": "issue:45507", + "accept": false, + "reason": "TRL SFT embedding unpacking and GraniteMoEHybrid invalid method calls are different code paths." + }, + { + "left": "issue:43975", + "right": "issue:44292", + "accept": false, + "reason": "Incorrect detokenization for deepseek-coder is unrelated to the Qwen-3-8B-NVFP4 runtime error." + }, + { + "left": "issue:43950", + "right": "issue:44534", + "accept": true, + "reason": "Same v5 regression: non-persistent buffers are being overwritten/corrupted on load, just phrased differently." + }, + { + "left": "issue:43644", + "right": "issue:43950", + "accept": true, + "reason": "Same underlying non-persistent-buffer corruption bug in Transformers 5.x." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet and stays conservative overall. The accepted soft edges are well supported by the issue titles, especially the non-persistent-buffer regression trio and the SigLIP2 output pair. The rejected pairs are mostly appropriately kept separate based on distinct models, failure modes, or subsystems, with no clear overstatement of evidence." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 3, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:36331", + "issue:38175", + "issue:41553", + "issue:41762", + "issue:42491", + "issue:42915", + "issue:42947", + "issue:43054", + "issue:43257", + "issue:43404", + "issue:43493", + "issue:43643", + "issue:43742", + "issue:43782", + "issue:43792", + "issue:43824", + "issue:43827", + "issue:43854", + "issue:43856", + "issue:43866", + "issue:43872", + "issue:43881", + "issue:43883", + "issue:43940", + "issue:43950", + "issue:43975", + "issue:44220", + "issue:44291", + "issue:44368", + "issue:44387", + "issue:44403", + "issue:44451", + "issue:44488", + "issue:44509", + "issue:44589", + "issue:44661", + "issue:44863", + "issue:44960", + "issue:44991", + "issue:45020", + "issue:45081", + "issue:45237", + "issue:45313", + "issue:45356", + "issue:45399", + "issue:45520", + "issue:45701" + ], + "soft_pairs": [ + "issue:36331|issue:41762", + "issue:41553|issue:43883", + "issue:43872|issue:44589", + "issue:38175|issue:43054", + "issue:41553|issue:43881", + "issue:43054|issue:45701", + "issue:42947|issue:44387", + "issue:44403|issue:44451", + "issue:43643|issue:45020", + "issue:43881|issue:43883", + "issue:43975|issue:45520", + "issue:43054|issue:43493", + "issue:43975|issue:45237", + "issue:44403|issue:44488", + "issue:42947|issue:43856", + "issue:42491|issue:44368", + "issue:41762|issue:42915", + "issue:43975|issue:44451", + "issue:44661|issue:45020", + "issue:43824|issue:43872", + "issue:44991|issue:45399", + "issue:45020|issue:45356", + "issue:43782|issue:44960", + "issue:43827|issue:44509", + "issue:45081|issue:45399", + "issue:43792|issue:44220", + "issue:43404|issue:43950", + "issue:43742|issue:45399", + "issue:42915|issue:45020", + "issue:43257|issue:43866", + "issue:43742|issue:45020", + "issue:41553|issue:44991", + "issue:43940|issue:45313", + "issue:43872|issue:44291", + "issue:43854|issue:44863", + "issue:42915|issue:43872" + ], + "prepared_review_unit_hash": "5e37c8dfa8b5574eace33cd366420468aeab63c6c0c49ff005fcd56b808a5210", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 36, + "serialized_chars": 37866, + "estimated_input_tokens": 9467, + "estimated_eval_tokens": 19190 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This is a mixed cluster of mostly unrelated model-loading, tokenizer, training, and docs issues. Only two soft edges look like true duplicates: the v5 docs pipeline cleanup pair and the `_is_hf_initialized` loading regression pair.", + "confidence": 0.77, + "canonical_issue_reason": "issue:43872 is the clearest exact-duplicate anchor here: issue:44291 describes the same `_is_hf_initialized` kwarg failure in the init_empty_weights / quantized-loading path.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45020 is the broadest cluster-level umbrella for recent remote_code/model-loading regressions, even though most items are not exact duplicates of it.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:36331", + "right": "issue:41762", + "accept": false, + "reason": "Different failure modes: custom trainer signature mismatch vs Gemma3 ZeRO-3 index error." + }, + { + "left": "issue:41553", + "right": "issue:43883", + "accept": false, + "reason": "Unrelated models and code paths: Voxtral AutoTokenizer error message vs Molmo tied-weights attribute error." + }, + { + "left": "issue:43872", + "right": "issue:44589", + "accept": false, + "reason": "Both are loading-related TypeErrors, but they point to different incompatibilities (_is_hf_initialized vs Float8 storage lookup)." + }, + { + "left": "issue:38175", + "right": "issue:43054", + "accept": false, + "reason": "Both mention SigLIP2, but one is zero probabilities and the other is degraded text embeddings; not clearly the same bug." + }, + { + "left": "issue:41553", + "right": "issue:43881", + "accept": false, + "reason": "Different models and unrelated loading failures." + }, + { + "left": "issue:43054", + "right": "issue:45701", + "accept": false, + "reason": "SigLIP2 embedding quality issue vs generic tokenization change regression; not the same underlying bug." + }, + { + "left": "issue:42947", + "right": "issue:44387", + "accept": false, + "reason": "Both involve memory, but one is LoRA gradient checkpointing and the other is int4 reserved-memory growth; different regressions." + }, + { + "left": "issue:44403", + "right": "issue:44451", + "accept": false, + "reason": "Generic loading noise vs a concrete ScandiBERT loading failure; not a duplicate." + }, + { + "left": "issue:43643", + "right": "issue:45020", + "accept": false, + "reason": "401remote_code issues are broad, but this is specifically missing fields from AutoConfig; not enough evidence of the same bug." + }, + { + "left": "issue:43881", + "right": "issue:43883", + "accept": false, + "reason": "Different model families and different failures." + }, + { + "left": "issue:43975", + "right": "issue:45520", + "accept": false, + "reason": "Tokenizer detokenization regression vs flash_attn import key error on Python 3.13." + }, + { + "left": "issue:43054", + "right": "issue:43493", + "accept": false, + "reason": "Related SigLIP2 area, but this is a performance/embedding-quality report and the other is a broader implementation discrepancy claim." + }, + { + "left": "issue:43975", + "right": "issue:45237", + "accept": false, + "reason": "Tokenizer regression vs AMD GPU runtime support issue." + }, + { + "left": "issue:44403", + "right": "issue:44488", + "accept": false, + "reason": "Noise/warnings on load vs a specific model load failure; different problems." + }, + { + "left": "issue:42947", + "right": "issue:43856", + "accept": false, + "reason": "Both are training/memory reports, but they describe different symptoms and likely different code paths." + }, + { + "left": "issue:42491", + "right": "issue:44368", + "accept": false, + "reason": "Qwen3 MoE version-compatibility/load issue vs a tie_word_embeddings warning in Qwen3.5 LoRA fine-tuning." + }, + { + "left": "issue:41762", + "right": "issue:42915", + "accept": false, + "reason": "Gemma3 ZeRO-3 index error vs Qwen3 MoE FP8 failure; unrelated." + }, + { + "left": "issue:43975", + "right": "issue:44451", + "accept": false, + "reason": "Different model/tokenizer regressions with no clear shared root cause." + }, + { + "left": "issue:44661", + "right": "issue:45020", + "accept": false, + "reason": "add-new-model-like mapping bug is unrelated to the broader remote_code loading regressions." + }, + { + "left": "issue:43824", + "right": "issue:43872", + "accept": false, + "reason": "Qwen2.5-VL import error vs bitsandbytes _is_hf_initialized TypeError; unrelated." + }, + { + "left": "issue:44991", + "right": "issue:45399", + "accept": false, + "reason": "Tokenizer loading regression vs flash-attn fallback gating; different bugs." + }, + { + "left": "issue:45020", + "right": "issue:45356", + "accept": false, + "reason": "Broad remote_code breakage vs a specific Kimi-K2.5 tokenizer regression; not the same concrete defect." + }, + { + "left": "issue:43782", + "right": "issue:44960", + "accept": false, + "reason": "Different model families and distinct loading errors." + }, + { + "left": "issue:43827", + "right": "issue:44509", + "accept": true, + "reason": "Both report the same stale docs problem: pipeline task names removed in v5 are still referenced." + }, + { + "left": "issue:45081", + "right": "issue:45399", + "accept": false, + "reason": "Tokenizer regex patch crash vs flash-attn fallback gating; unrelated." + }, + { + "left": "issue:43792", + "right": "issue:44220", + "accept": false, + "reason": "Whisper runtime failure and a feature-extraction-function issue may be related, but the evidence is not strong enough to call them the same bug." + }, + { + "left": "issue:43404", + "right": "issue:43950", + "accept": false, + "reason": "Weight tying bug in Mistral3 vs silent non-persistent buffer corruption; different regressions." + }, + { + "left": "issue:43742", + "right": "issue:45399", + "accept": false, + "reason": "MobileLLM loading key error vs flash-attn fallback behavior; unrelated." + }, + { + "left": "issue:42915", + "right": "issue:45020", + "accept": false, + "reason": "Qwen3 MoE FP8 failure is much narrower than the broad remote_code regression bucket." + }, + { + "left": "issue:43257", + "right": "issue:43866", + "accept": false, + "reason": "Qwen3 MoE weight-conversion/load issue vs Ovis2 checkpoint corruption; unrelated." + }, + { + "left": "issue:43742", + "right": "issue:45020", + "accept": false, + "reason": "Specific MobileLLM key error does not match the broad remote_code regression." + }, + { + "left": "issue:41553", + "right": "issue:44991", + "accept": false, + "reason": "Voxtral tokenizer error message vs tokenizer-loading failure for a different model; unrelated." + }, + { + "left": "issue:43940", + "right": "issue:45313", + "accept": false, + "reason": "Both are ZeRO-3 weight-loading issues, but they affect different Qwen variants and describe different missing-weight patterns." + }, + { + "left": "issue:43872", + "right": "issue:44291", + "accept": true, + "reason": "Same underlying `_is_hf_initialized` TypeError during init_empty_weights / quantized model loading." + }, + { + "left": "issue:43854", + "right": "issue:44863", + "accept": false, + "reason": "GLM-4.7-Flash test loading problem vs NemotronH checkpoint-loading failure; unrelated." + }, + { + "left": "issue:42915", + "right": "issue:43872", + "accept": false, + "reason": "Different failures: FP8 Qwen3 MoE loading vs `_is_hf_initialized` quantized-loading TypeError." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet, and the two accepted soft edges are well-supported by the issue titles. The remaining soft-edge rejections stay conservative and do not overclaim duplicate status." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 4, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:24643", + "issue:30064", + "issue:36010", + "issue:36331", + "issue:38175", + "issue:39692", + "issue:41093", + "issue:41553", + "issue:42617", + "issue:42915", + "issue:43257", + "issue:43317", + "issue:43366", + "issue:43531", + "issue:43550", + "issue:43596", + "issue:43638", + "issue:43645", + "issue:43646", + "issue:43824", + "issue:43828", + "issue:43872", + "issue:43950", + "issue:43957", + "issue:43994", + "issue:44155", + "issue:44479", + "issue:44488", + "issue:44530", + "issue:44560", + "issue:44589", + "issue:44661", + "issue:44756", + "issue:44805", + "issue:44863", + "issue:44877", + "issue:44910", + "issue:44918", + "issue:44928", + "issue:45030", + "issue:45305", + "issue:45313", + "issue:45325", + "issue:45357", + "issue:45538", + "issue:45584", + "issue:45588" + ], + "soft_pairs": [ + "issue:44928|issue:45588", + "issue:41553|issue:44488", + "issue:43257|issue:45305", + "issue:36331|issue:41093", + "issue:43872|issue:45538", + "issue:44530|issue:44863", + "issue:41093|issue:44805", + "issue:44479|issue:44560", + "issue:24643|issue:30064", + "issue:43366|issue:43828", + "issue:43366|issue:43531", + "issue:43550|issue:44910", + "issue:43645|issue:43646", + "issue:43366|issue:44589", + "issue:43950|issue:45325", + "issue:44560|issue:44918", + "issue:42617|issue:43366", + "issue:42915|issue:43957", + "issue:43824|issue:43957", + "issue:36010|issue:42915", + "issue:38175|issue:43994", + "issue:43646|issue:43950", + "issue:44155|issue:45584", + "issue:44877|issue:45030", + "issue:44560|issue:44805", + "issue:43638|issue:44661", + "issue:43596|issue:44661", + "issue:43317|issue:44756", + "issue:38175|issue:39692", + "issue:45313|issue:45357" + ], + "prepared_review_unit_hash": "8bf8f40d4c157d33c86d4fe87cd01e8ca5272531b6a58533cdc480a1bd89acd5", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 30, + "serialized_chars": 36297, + "estimated_input_tokens": 9075, + "estimated_eval_tokens": 18406 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Mostly a heterogeneous cluster of unrelated issues, with only a few clear duplicate pockets. The strongest overlaps are the mask/tensor shape IndexError reports, the Transformers 5.0 custom model initialization regression, and the SigLIP2 output regression; the rest look distinct.", + "confidence": 0.82, + "canonical_issue_reason": "No single canonical issue for the whole cluster; it splits into multiple unrelated bug families. The closest duplicate anchor is the mask-shape IndexError pair.", + "canonical_pr_reason": null, + "best_issue_reason": "No global best issue for the entire cluster for the same reason; if forced to anchor one duplicate pocket, the earliest mask-shape IndexError report is the cleanest representative.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44928", + "right": "issue:45588", + "accept": false, + "reason": "Different Qwen3.5 bugs: SDPA/3D position_ids NaN vs flash_attention s_aux=None crash." + }, + { + "left": "issue:41553", + "right": "issue:44488", + "accept": false, + "reason": "Different models and failures; Voxtral tokenizer error message vs sleng-bert load failure." + }, + { + "left": "issue:43257", + "right": "issue:45305", + "accept": false, + "reason": "Both DeepSpeed-related, but one is weight conversion and the other gradient averaging; not the same bug." + }, + { + "left": "issue:36331", + "right": "issue:41093", + "accept": false, + "reason": "Different code paths: custom trainer signature mismatch vs mask/tensor shape IndexError." + }, + { + "left": "issue:43872", + "right": "issue:45538", + "accept": false, + "reason": "bitsandbytes init incompatibility vs CLIPTokenizer max_length bug." + }, + { + "left": "issue:44530", + "right": "issue:44863", + "accept": false, + "reason": "Qwen3.5 PagedAttention cache crash vs NemotronH checkpoint loading." + }, + { + "left": "issue:41093", + "right": "issue:44805", + "accept": true, + "reason": "Same generic mask/tensor shape mismatch IndexError; 44805 looks like a later duplicate report." + }, + { + "left": "issue:44479", + "right": "issue:44560", + "accept": false, + "reason": "Both are video/Qwen issues, but the failures differ: broad video-input regression vs StopIteration in Qwen3-vl-embedding." + }, + { + "left": "issue:24643", + "right": "issue:30064", + "accept": false, + "reason": "Training/DeepSpeed weight shape error vs image processor void-segmentation issue." + }, + { + "left": "issue:43366", + "right": "issue:43828", + "accept": false, + "reason": "GGUF/gpt-oss support vs Phi-tiny-MoE autocast dtype mismatch." + }, + { + "left": "issue:43366", + "right": "issue:43531", + "accept": false, + "reason": "Unrelated Qwen3-MoE sliding_window bug vs GGUF gpt-oss support." + }, + { + "left": "issue:43550", + "right": "issue:44910", + "accept": false, + "reason": "Different models and kernels: Bamba torch.compile/SDPA vs Qwen3.5 flash-attention position_ids crash." + }, + { + "left": "issue:43645", + "right": "issue:43646", + "accept": true, + "reason": "Same Transformers 5.0 custom model initialization regression; one is just a notebook-specific reproduction." + }, + { + "left": "issue:43366", + "right": "issue:44589", + "accept": false, + "reason": "GGUF architecture support vs missing Float8 storage type." + }, + { + "left": "issue:43950", + "right": "issue:45325", + "accept": false, + "reason": "Buffer loading regression vs Qwen2.5-VL rope/temporal position scaling." + }, + { + "left": "issue:44560", + "right": "issue:44918", + "accept": false, + "reason": "Video StopIteration in Qwen3-vl-embedding vs Qwen3.5 TRL unpacking failure." + }, + { + "left": "issue:42617", + "right": "issue:43366", + "accept": false, + "reason": "3d_parallel.py runtime failure unrelated to GGUF support." + }, + { + "left": "issue:42915", + "right": "issue:43957", + "accept": false, + "reason": "Qwen3Moe FineGrainedFP8Config failure vs meta-device model loading." + }, + { + "left": "issue:43824", + "right": "issue:43957", + "accept": false, + "reason": "Import/export of Qwen2_5_VL class vs meta-device loading regression." + }, + { + "left": "issue:36010", + "right": "issue:42915", + "accept": false, + "reason": "GenerationMixin import error vs Qwen3Moe FP8 config." + }, + { + "left": "issue:38175", + "right": "issue:43994", + "accept": true, + "reason": "Same SigLIP2 model/output regression; both report broken predictions for google/siglip2-base-patch16-224." + }, + { + "left": "issue:43646", + "right": "issue:43950", + "accept": false, + "reason": "Custom model init regression vs non-persistent buffer corruption." + }, + { + "left": "issue:44155", + "right": "issue:45584", + "accept": false, + "reason": "AudioFlamingo3 track leak vs Whisper empty transcription bug." + }, + { + "left": "issue:44877", + "right": "issue:45030", + "accept": false, + "reason": "Different config-validation failures on different models." + }, + { + "left": "issue:44560", + "right": "issue:44805", + "accept": false, + "reason": "Different error classes: video StopIteration vs generic mask shape mismatch." + }, + { + "left": "issue:43638", + "right": "issue:44661", + "accept": false, + "reason": "Zero3/Bert index error vs tokenizer-mapping restriction in add-new-model-like." + }, + { + "left": "issue:43596", + "right": "issue:44661", + "accept": false, + "reason": "Zero3/Bert index error vs tokenizer-mapping restriction in add-new-model-like." + }, + { + "left": "issue:43317", + "right": "issue:44756", + "accept": false, + "reason": "Dequantized model offload failure vs Strix Halo mmap OOM." + }, + { + "left": "issue:38175", + "right": "issue:39692", + "accept": false, + "reason": "Same model family but different issues: doc-example validation/quantization vs output probabilities." + }, + { + "left": "issue:45313", + "right": "issue:45357", + "accept": false, + "reason": "Load-time ZeRO-3 weight issue vs save_pretrained visual encoder key regression." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet: the cluster is mostly heterogeneous, and the three accepted duplicate pockets are tightly supported by the titles. The rejected soft edges are conservative and do not overclaim evidence." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 5, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:36010", + "issue:39404", + "issue:39692", + "issue:41720", + "issue:42491", + "issue:42915", + "issue:42947", + "issue:43257", + "issue:43335", + "issue:43404", + "issue:43425", + "issue:43454", + "issue:43644", + "issue:43645", + "issue:43708", + "issue:43720", + "issue:43854", + "issue:43856", + "issue:43873", + "issue:43931", + "issue:43957", + "issue:43994", + "issue:44075", + "issue:44164", + "issue:44291", + "issue:44387", + "issue:44402", + "issue:44451", + "issue:44488", + "issue:44534", + "issue:44661", + "issue:44683", + "issue:44756", + "issue:44910", + "issue:44960", + "issue:44977", + "issue:44991", + "issue:45216", + "issue:45276", + "issue:45292", + "issue:45335", + "issue:45362", + "issue:45406", + "issue:45588", + "issue:45663", + "issue:45704", + "issue:45759" + ], + "soft_pairs": [ + "issue:43644|issue:44534", + "issue:44960|issue:44977", + "issue:44451|issue:44488", + "issue:42491|issue:43720", + "issue:43708|issue:44075", + "issue:43854|issue:43931", + "issue:42491|issue:45276", + "issue:41720|issue:42915", + "issue:43856|issue:43873", + "issue:43257|issue:43854", + "issue:39692|issue:43994", + "issue:42915|issue:44756", + "issue:42491|issue:43931", + "issue:43856|issue:44387", + "issue:44661|issue:44683", + "issue:43644|issue:43645", + "issue:44910|issue:45588", + "issue:45362|issue:45406", + "issue:42491|issue:45292", + "issue:44488|issue:44991", + "issue:43404|issue:43454", + "issue:36010|issue:39404", + "issue:45276|issue:45335", + "issue:44164|issue:44291", + "issue:44402|issue:44488", + "issue:39692|issue:44960", + "issue:43425|issue:43957", + "issue:42947|issue:45704", + "issue:43335|issue:45663", + "issue:45663|issue:45759", + "issue:43873|issue:44387", + "issue:44387|issue:44756", + "issue:43854|issue:45216", + "issue:44291|issue:44756" + ], + "prepared_review_unit_hash": "8637dd97acadfb7b439285db8292b5acf7620bdea49ad9e3fa2adf2155f78a6e", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 34, + "serialized_chars": 37316, + "estimated_input_tokens": 9329, + "estimated_eval_tokens": 18914 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This set is mostly a grab bag of distinct Transformers bugs; only two soft edges look like true duplicates: the Transformers 5 non-persistent-buffer junk issue and the resize_token_embeddings propagation bug. The rest are too model- or code-path-specific to merge.", + "confidence": 0.81, + "canonical_issue_reason": "issue:43644 is the strongest representative because it has an almost exact duplicate in 44534 and the bug statement is very specific and reusable as a canonical report.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43644 is the best overall issue candidate: clear repro class, strong duplicate evidence, and a concrete Transformers 5 regression with broad impact.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43644", + "right": "issue:44534", + "accept": true, + "reason": "Same bug: Transformers v5 filling non-persistent buffers with junk; titles describe the same regression nearly verbatim." + }, + { + "left": "issue:44960", + "right": "issue:44977", + "accept": false, + "reason": "Both involve generation/flash-attention, but they target different models and failure modes; not the same concrete bug." + }, + { + "left": "issue:44451", + "right": "issue:44488", + "accept": false, + "reason": "Both are model-loading complaints, but the titles point to different models and no shared root cause is evident." + }, + { + "left": "issue:42491", + "right": "issue:43720", + "accept": false, + "reason": "Unrelated: Qwen3 MoE LoRA compatibility vs BitNet packed-weight loading in CI." + }, + { + "left": "issue:43708", + "right": "issue:44075", + "accept": false, + "reason": "Different problems: checkpoint resume step counting vs SGD optimizer arguments not being used." + }, + { + "left": "issue:43854", + "right": "issue:43931", + "accept": false, + "reason": "Different model-loading failures for different models and different symptoms." + }, + { + "left": "issue:42491", + "right": "issue:45276", + "accept": false, + "reason": "Different subsystems: Qwen3 MoE LoRA loading vs token embedding resize propagation." + }, + { + "left": "issue:41720", + "right": "issue:42915", + "accept": false, + "reason": "Both are Qwen3-related, but one is a device-mapping CUDA assert and the other is a FineGrainedFP8Config failure." + }, + { + "left": "issue:43856", + "right": "issue:43873", + "accept": false, + "reason": "Both mention memory/quantization themes, but the concrete bugs are different." + }, + { + "left": "issue:43257", + "right": "issue:43854", + "accept": false, + "reason": "Different model families and unrelated loading paths." + }, + { + "left": "issue:39692", + "right": "issue:43994", + "accept": false, + "reason": "Both are SigLIP2-related, but one is a docs example with model/processor mismatch and the other is a nonsensical-output complaint." + }, + { + "left": "issue:42915", + "right": "issue:44756", + "accept": false, + "reason": "Unrelated: Qwen3 MoE FP8 failure vs Strix Halo mmap/OOM behavior." + }, + { + "left": "issue:42491", + "right": "issue:43931", + "accept": false, + "reason": "Different issues: Qwen3 MoE LoRA loading vs Qwen3-VL weight-shape mismatch." + }, + { + "left": "issue:43856", + "right": "issue:44387", + "accept": false, + "reason": "Both are memory-related, but one is training inefficiency and the other is int4 quantization OOM; not the same bug." + }, + { + "left": "issue:44661", + "right": "issue:44683", + "accept": false, + "reason": "Different areas: tokenizer-mapping registration vs compiled flex_attention on new torch." + }, + { + "left": "issue:43644", + "right": "issue:43645", + "accept": false, + "reason": "Different regressions: junk non-persistent buffers vs custom-model initialization in notebooks." + }, + { + "left": "issue:44910", + "right": "issue:45588", + "accept": false, + "reason": "Both are flash-attention crashes, but one is a packed-sequence position_ids bug and the other is an s_aux=None AttributeError." + }, + { + "left": "issue:45362", + "right": "issue:45406", + "accept": false, + "reason": "Different model/application paths: chat crash in Qwen3.5 vs serve crash in Gemma4Processor." + }, + { + "left": "issue:42491", + "right": "issue:45292", + "accept": false, + "reason": "Unrelated: Qwen3 MoE LoRA loading vs resize_token_embeddings not updating output embeddings." + }, + { + "left": "issue:44488", + "right": "issue:44991", + "accept": false, + "reason": "Both are tokenizer/loading problems, but they concern different models and different failure symptoms." + }, + { + "left": "issue:43404", + "right": "issue:43454", + "accept": false, + "reason": "Similar wording about tied lm_head weights, but these are separate model implementations and not clearly one duplicate bug." + }, + { + "left": "issue:36010", + "right": "issue:39404", + "accept": false, + "reason": "Unrelated: missing GenerationMixin import vs Whisper pipeline return_language regression." + }, + { + "left": "issue:45276", + "right": "issue:45335", + "accept": true, + "reason": "Same underlying bug: resize_token_embeddings is not propagating to the relevant tied output/decoder embeddings in related model variants." + }, + { + "left": "issue:44164", + "right": "issue:44291", + "accept": false, + "reason": "Different serialization/loading failures: extra_state handling vs init_empty_weights/_is_hf_initialized TypeError." + }, + { + "left": "issue:44402", + "right": "issue:44488", + "accept": false, + "reason": "Both are model loading issues, but they involve different models and different failure descriptions." + }, + { + "left": "issue:39692", + "right": "issue:44960", + "accept": false, + "reason": "Different model families and unrelated failure modes." + }, + { + "left": "issue:43425", + "right": "issue:43957", + "accept": false, + "reason": "Torch version incompatibility is unrelated to meta-device loading breakage." + }, + { + "left": "issue:42947", + "right": "issue:45704", + "accept": false, + "reason": "Different concerns: LoRA gradient checkpointing vs an apex RMSNorm memory leak." + }, + { + "left": "issue:43335", + "right": "issue:45663", + "accept": false, + "reason": "Unrelated model-specific bugs: sparse-layer config creation vs FSDP2 shared_kv_states handling." + }, + { + "left": "issue:45663", + "right": "issue:45759", + "accept": false, + "reason": "Different bugs in different model/config paths; no shared concrete code-path problem." + }, + { + "left": "issue:43873", + "right": "issue:44387", + "accept": false, + "reason": "Both are memory-related, but one is quantization offloading behavior and the other is int4 CUDA reserved-memory growth." + }, + { + "left": "issue:44387", + "right": "issue:44756", + "accept": false, + "reason": "Different memory/OOM issues with no shared root cause." + }, + { + "left": "issue:43854", + "right": "issue:45216", + "accept": false, + "reason": "Different model-loading/save regressions for different models." + }, + { + "left": "issue:44291", + "right": "issue:44756", + "accept": false, + "reason": "Different issues: init_empty_weights argument mismatch vs mmap-induced OOM." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary stays grounded in the packet and the duplicate calls are conservative. The two accepted soft edges are strongly supported by the titles/body-level similarity, and the rejected pairs are appropriately cautious about model- or failure-specific differences. The only mild stretch is the phrasing about \"broad impact\" for issue:43644, but it does not materially overstate the evidence." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 6, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:39404", + "issue:41720", + "issue:42947", + "issue:43257", + "issue:43299", + "issue:43381", + "issue:43425", + "issue:43611", + "issue:43645", + "issue:43646", + "issue:43708", + "issue:43716", + "issue:43828", + "issue:43854", + "issue:43856", + "issue:43901", + "issue:43906", + "issue:43931", + "issue:43957", + "issue:44155", + "issue:44164", + "issue:44265", + "issue:44292", + "issue:44402", + "issue:44451", + "issue:44479", + "issue:44509", + "issue:44530", + "issue:44589", + "issue:44756", + "issue:44805", + "issue:44863", + "issue:44877", + "issue:44898", + "issue:44912", + "issue:44918", + "issue:44928", + "issue:44960", + "issue:45084", + "issue:45276", + "issue:45292", + "issue:45335", + "issue:45588", + "issue:45663", + "issue:45684", + "issue:45698", + "issue:45701" + ], + "soft_pairs": [ + "issue:44292|issue:44912", + "issue:42947|issue:43381", + "issue:45684|issue:45698", + "issue:45276|issue:45292", + "issue:43381|issue:44928", + "issue:44164|issue:44756", + "issue:41720|issue:44155", + "issue:45588|issue:45663", + "issue:43646|issue:43708", + "issue:44479|issue:44877", + "issue:43854|issue:44960", + "issue:43957|issue:44756", + "issue:45292|issue:45335", + "issue:44402|issue:44451", + "issue:43645|issue:44164", + "issue:43828|issue:43957", + "issue:43257|issue:44863", + "issue:43425|issue:43828", + "issue:44589|issue:44756", + "issue:44479|issue:45663", + "issue:43906|issue:44530", + "issue:39404|issue:45701", + "issue:43611|issue:43645", + "issue:44530|issue:45084", + "issue:43425|issue:43716", + "issue:43931|issue:44863", + "issue:43901|issue:44509", + "issue:39404|issue:44912", + "issue:44863|issue:45663", + "issue:44265|issue:44898", + "issue:43299|issue:43931", + "issue:44479|issue:44805", + "issue:43425|issue:44918", + "issue:43856|issue:44756", + "issue:44589|issue:45084" + ], + "prepared_review_unit_hash": "1b38e4851ea3336280727f64fbf49915264e4aa012bb64d2496b1de1c7b7ce10", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 35, + "serialized_chars": 37526, + "estimated_input_tokens": 9382, + "estimated_eval_tokens": 19020 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly heterogeneous: most items are separate regressions across loading, quantization, training, and docs. The only clearly duplicate-worthy subcluster is the resize_token_embeddings regression around output embeddings/decoder embeddings (45276, 45292, 45335).", + "confidence": 0.69, + "canonical_issue_reason": "issue:45292 is the broadest, most central description of the resize_token_embeddings regression affecting output embeddings; issue:45276 and issue:45335 look like model-specific manifestations of the same bug.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45292 is the best anchor for deduping the resize_token_embeddings bug because it is generic rather than model-specific and matches the shared underlying failure mode.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44292", + "right": "issue:44912", + "accept": false, + "reason": "Both are quantization-related load failures, but they involve different models and symptoms (NVFP4 run error vs MXFP4 fallback to bf16), so they do not look like the same bug." + }, + { + "left": "issue:42947", + "right": "issue:43381", + "accept": false, + "reason": "Both mention gradient checkpointing, but one is about PEFT/LoRA effectiveness and the other is an eval-mode restriction; different code paths and failure modes." + }, + { + "left": "issue:45684", + "right": "issue:45698", + "accept": false, + "reason": "One is about read-only permissions during save_pretrained, the other about loading the wrong custom module after save_pretrained; related area, but distinct bugs." + }, + { + "left": "issue:45276", + "right": "issue:45292", + "accept": true, + "reason": "Both report resize_token_embeddings not propagating to output embeddings; issue:45276 is a Gemma4-specific manifestation of the broader bug in issue:45292." + }, + { + "left": "issue:43381", + "right": "issue:44928", + "accept": false, + "reason": "These are unrelated: eval-mode gradient checkpointing restriction versus RLHF NaN/gradient explosion from 3D position_ids and SDPA fallback." + }, + { + "left": "issue:44164", + "right": "issue:44756", + "accept": false, + "reason": "Save/from_pretrained extra_state handling and mmap OOM on Strix Halo are unrelated problems." + }, + { + "left": "issue:41720", + "right": "issue:44155", + "accept": false, + "reason": "Different subsystems and symptoms: device-mapping cuda assert on Qwen3 versus batched AudioFlamingo3 inference leakage." + }, + { + "left": "issue:45588", + "right": "issue:45663", + "accept": false, + "reason": "Flash-attention s_aux=None crash and Gemma4 FSDP2 KeyError are different concrete bugs despite both touching model internals." + }, + { + "left": "issue:43646", + "right": "issue:43708", + "accept": false, + "reason": "Custom model initialization breakage and Trainer max_steps miscalculation on checkpoint resume are unrelated." + }, + { + "left": "issue:44479", + "right": "issue:44877", + "accept": false, + "reason": "Video-input regression for Qwen models and strict config blocking granite_speech loading are different failures." + }, + { + "left": "issue:43854", + "right": "issue:44960", + "accept": false, + "reason": "GLM-4.7-Flash loading issues and GLM5 are related only by family name; the reports describe different model/version problems." + }, + { + "left": "issue:43957", + "right": "issue:44756", + "accept": false, + "reason": "Meta-device loading failures and Strix Halo mmap OOM are unrelated." + }, + { + "left": "issue:45292", + "right": "issue:45335", + "accept": true, + "reason": "Both describe the same resize_token_embeddings propagation bug: issue:45292 mentions output_embeddings generally, while issue:45335 is the decoder.embed_tokens manifestation in t5gemma." + }, + { + "left": "issue:44402", + "right": "issue:44451", + "accept": false, + "reason": "A tokenizer vocab mismatch and a generic load failure for ScandiBERT are not clearly the same underlying bug." + }, + { + "left": "issue:43645", + "right": "issue:44164", + "accept": false, + "reason": "Notebook custom-model initialization and extra_state save/load handling are different issues." + }, + { + "left": "issue:43828", + "right": "issue:43957", + "accept": false, + "reason": "Autocast dtype mismatch in Phi-tiny-MoE and meta-device loading breakage are separate bugs." + }, + { + "left": "issue:43257", + "right": "issue:44863", + "accept": false, + "reason": "Both are loading/checkpoint problems, but one is Qwen3 MoE conversion with accelerate+deepspeed and the other is NemotronH checkpoint compatibility; not the same bug." + }, + { + "left": "issue:43425", + "right": "issue:43828", + "accept": false, + "reason": "Torch version incompatibility and autocast dtype mismatch are different failure modes." + }, + { + "left": "issue:44589", + "right": "issue:44756", + "accept": false, + "reason": "Float8 storage deserialization and mmap OOM are unrelated." + }, + { + "left": "issue:44479", + "right": "issue:45663", + "accept": false, + "reason": "Qwen video regression and Gemma4 FSDP2 shared_kv_states KeyError are unrelated." + }, + { + "left": "issue:43906", + "right": "issue:44530", + "accept": false, + "reason": "An isolated reproduction of another issue and a PagedAttentionCache linear_attention crash are not clearly the same bug." + }, + { + "left": "issue:39404", + "right": "issue:45701", + "accept": false, + "reason": "Whisper pipeline return_language and version-dependent tokenization changes are unrelated." + }, + { + "left": "issue:43611", + "right": "issue:43645", + "accept": false, + "reason": "Both are Transformers 5.0 custom-model regressions, but one is about base_model_prefix loading and the other about notebook-based initialization; too different to merge as duplicates." + }, + { + "left": "issue:44530", + "right": "issue:45084", + "accept": false, + "reason": "PagedAttentionCache linear_attention crash and compile non-template-nodes TypeError are unrelated." + }, + { + "left": "issue:43425", + "right": "issue:43716", + "accept": false, + "reason": "Torch 2.10 incompatibility and Mistral-3 image-preprocessor/model dtype mismatch are different issues." + }, + { + "left": "issue:43931", + "right": "issue:44863", + "accept": false, + "reason": "Qwen3-VL weight-shape mismatch and NemotronH checkpoint loading are different model-specific load failures." + }, + { + "left": "issue:43901", + "right": "issue:44509", + "accept": false, + "reason": "Both are docs issues, but they cover different pipeline/task deprecations and do not describe the same change." + }, + { + "left": "issue:39404", + "right": "issue:44912", + "accept": false, + "reason": "Whisper pipeline language return and quantization loading/fallback issues are unrelated." + }, + { + "left": "issue:44863", + "right": "issue:45663", + "accept": false, + "reason": "NemotronH checkpoint loading and Gemma4 FSDP2 KeyError are distinct model-loading/training bugs." + }, + { + "left": "issue:44265", + "right": "issue:44898", + "accept": false, + "reason": "torch.export.export with torch_compilable_check and Perceiver resize/interpolation failure are different problems." + }, + { + "left": "issue:43299", + "right": "issue:43931", + "accept": false, + "reason": "Both involve Qwen3-VL, but one is a dev-version MoE loading regression and the other is a checkpoint shape mismatch; not clearly the same underlying bug." + }, + { + "left": "issue:44479", + "right": "issue:44805", + "accept": false, + "reason": "Video-input regression and mask/index shape mismatch are not the same issue." + }, + { + "left": "issue:43856", + "right": "issue:44756", + "accept": false, + "reason": "Qwen3 MoE training memory inefficiency and Strix Halo mmap OOM are unrelated performance/memory issues." + }, + { + "left": "issue:44589", + "right": "issue:45084", + "accept": false, + "reason": "Float8 storage lookup failure and compile-on-template-nodes error are unrelated." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is well grounded in the packet and stays mostly conservative. The resize_token_embeddings trio (45276, 45292, 45335) is the only clearly duplicate-looking subcluster from the titles, and the other pairwise rejections are cautious rather than overstated." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 7, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:32090", + "issue:34567", + "issue:36032", + "issue:39404", + "issue:40444", + "issue:40990", + "issue:42915", + "issue:42947", + "issue:43317", + "issue:43404", + "issue:43421", + "issue:43425", + "issue:43452", + "issue:43482", + "issue:43541", + "issue:43582", + "issue:43596", + "issue:43611", + "issue:43632", + "issue:43643", + "issue:43644", + "issue:43645", + "issue:43646", + "issue:43716", + "issue:43883", + "issue:43994", + "issue:44079", + "issue:44164", + "issue:44206", + "issue:44220", + "issue:44451", + "issue:44466", + "issue:44479", + "issue:44488", + "issue:44530", + "issue:44671", + "issue:44683", + "issue:44811", + "issue:44849", + "issue:44987", + "issue:44991", + "issue:45081", + "issue:45084", + "issue:45325", + "issue:45507", + "issue:45584", + "issue:45685", + "issue:45759" + ], + "soft_pairs": [ + "issue:43596|issue:44683", + "issue:40990|issue:42947", + "issue:43425|issue:45685", + "issue:39404|issue:44671", + "issue:43994|issue:44079", + "issue:44671|issue:44991", + "issue:44164|issue:45759", + "issue:36032|issue:43452", + "issue:44991|issue:45081", + "issue:32090|issue:39404", + "issue:44164|issue:44479", + "issue:44488|issue:45081", + "issue:40990|issue:43421", + "issue:43582|issue:43632", + "issue:43883|issue:44488", + "issue:43611|issue:43644", + "issue:44206|issue:44220", + "issue:43716|issue:45084", + "issue:44811|issue:45584", + "issue:44488|issue:44987", + "issue:34567|issue:40444", + "issue:43643|issue:44530", + "issue:43644|issue:43646", + "issue:43883|issue:44991", + "issue:44451|issue:44987", + "issue:36032|issue:42915", + "issue:43541|issue:45084", + "issue:43452|issue:43482", + "issue:42915|issue:44530", + "issue:43317|issue:43482", + "issue:44849|issue:45325", + "issue:43404|issue:44466", + "issue:36032|issue:43645", + "issue:45084|issue:45507" + ], + "prepared_review_unit_hash": "9e21ceda20d06098cbb57ef658b06518e23060997fc7ba19d0709546217658a3", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 34, + "serialized_chars": 37891, + "estimated_input_tokens": 9473, + "estimated_eval_tokens": 19202 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly a grab-bag of unrelated Transformers regression reports, with one clear duplicate-like GGUF loading pair. The most representative issue is the generic v5 model-loading regression around `base_model_prefix`/core loading behavior.", + "confidence": 0.69, + "canonical_issue_reason": "Issue 43611 is the broadest, most central regression report in the set: a Transformers v5 model-loading failure affecting the core loading path, which matches the overall cluster theme better than the more model-specific reports.", + "canonical_pr_reason": null, + "best_issue_reason": "43611 is the best single issue to anchor this cluster because it is generic, high-signal, and closest to the shared v5 loading-regression pattern seen across several nearby issues.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43596", + "right": "issue:44683", + "accept": false, + "reason": "Different failures in different stacks: DeepSpeed ZeRO-3 initialization vs compiled flex_attention on torch>=2.9." + }, + { + "left": "issue:40990", + "right": "issue:42947", + "accept": false, + "reason": "Unrelated bugs: perplexity on a specific model/dataset vs gradient-checkpointing behavior with PEFT LoRA." + }, + { + "left": "issue:43425", + "right": "issue:45685", + "accept": false, + "reason": "No shared code path; one is a Torch version incompatibility, the other is an MPS-specific unsupported op in a MoE path." + }, + { + "left": "issue:39404", + "right": "issue:44671", + "accept": false, + "reason": "Different model families and symptoms: Whisper pipeline return_language vs CamemBERT masked-LM predictions." + }, + { + "left": "issue:43994", + "right": "issue:44079", + "accept": false, + "reason": "Both mention incorrect outputs, but the concrete bugs differ and there is no clear shared code-path evidence." + }, + { + "left": "issue:44671", + "right": "issue:44991", + "accept": false, + "reason": "Separate tokenizer/model loading issues for different models; no indication they share the same underlying regression." + }, + { + "left": "issue:44164", + "right": "issue:45759", + "accept": false, + "reason": "save/from_pretrained extra_state handling is unrelated to `from_config` unwrapping of `text_config` in composite configs." + }, + { + "left": "issue:36032", + "right": "issue:43452", + "accept": false, + "reason": "T5 tokenizer special-token method conflict is unrelated to GGUF loading failures." + }, + { + "left": "issue:44991", + "right": "issue:45081", + "accept": false, + "reason": "Both are tokenizer-loading problems, but for different models and different failure mechanisms." + }, + { + "left": "issue:32090", + "right": "issue:39404", + "accept": false, + "reason": "Trainer GPU broadcast TypeError is unrelated to Whisper pipeline return_language behavior." + }, + { + "left": "issue:44164", + "right": "issue:44479", + "accept": false, + "reason": "Save/load extra_state handling and Qwen VL video input regression are unrelated." + }, + { + "left": "issue:44488", + "right": "issue:45081", + "accept": false, + "reason": "Different model loaders and different root causes; only a broad tokenizer-loading theme is shared." + }, + { + "left": "issue:40990", + "right": "issue:43421", + "accept": false, + "reason": "Perplexity quality issue vs runtime post-processor update bug; not the same bug or fix." + }, + { + "left": "issue:43582", + "right": "issue:43632", + "accept": false, + "reason": "Apple Silicon allocator TypeError and `_is_hf_initialized` regression are unrelated." + }, + { + "left": "issue:43883", + "right": "issue:44488", + "accept": false, + "reason": "Missing tied-weights key attribute in Molmo is unrelated to a BERT model loading failure." + }, + { + "left": "issue:43611", + "right": "issue:43644", + "accept": false, + "reason": "Both are v5 regressions, but one is model-loading/base_model_prefix and the other is non-persistent buffer initialization; different bugs." + }, + { + "left": "issue:44206", + "right": "issue:44220", + "accept": false, + "reason": "The titles suggest the same audio area, but the concrete failures are not clearly the same code-path problem." + }, + { + "left": "issue:43716", + "right": "issue:45084", + "accept": false, + "reason": "Image preprocessor dtype mismatch is unrelated to a template-node compilation error." + }, + { + "left": "issue:44811", + "right": "issue:45584", + "accept": false, + "reason": "Both involve Whisper and special tokens, but one is batch_decode skipping tokens and the other is empty-transcription generation after alignment; too different to merge." + }, + { + "left": "issue:44488", + "right": "issue:44987", + "accept": false, + "reason": "Different model-loading failures with no shared concrete regression evidence." + }, + { + "left": "issue:34567", + "right": "issue:40444", + "accept": false, + "reason": "Trainer token-count tracking is unrelated to multi-image IterableDataset fine-tuning failure." + }, + { + "left": "issue:43643", + "right": "issue:44530", + "accept": false, + "reason": "AutoConfig trust_remote_code field loss is unrelated to PagedAttentionCache crashing on Qwen3.5 linear_attention." + }, + { + "left": "issue:43644", + "right": "issue:43646", + "accept": false, + "reason": "A non-persistent buffer initialization regression is not the same as custom model initialization in Jupyter notebooks." + }, + { + "left": "issue:43883", + "right": "issue:44991", + "accept": false, + "reason": "Molmo tied-weights attribute error and tokenizer loading for EMBEDDIA/est-roberta are unrelated." + }, + { + "left": "issue:44451", + "right": "issue:44987", + "accept": false, + "reason": "Different model-specific loading failures with no clear shared implementation bug." + }, + { + "left": "issue:36032", + "right": "issue:42915", + "accept": false, + "reason": "Tokenizer method-name conflict is unrelated to Qwen3Moe FineGrainedFP8Config failure." + }, + { + "left": "issue:43541", + "right": "issue:45084", + "accept": false, + "reason": "TorchDynamo grouped_mm tracing error is unrelated to template-node compilation." + }, + { + "left": "issue:43452", + "right": "issue:43482", + "accept": true, + "reason": "Both are GGUF loading regressions in Transformers v5 affecting `from_pretrained`; they plausibly come from the same GGUF loader-path bug." + }, + { + "left": "issue:42915", + "right": "issue:44530", + "accept": false, + "reason": "Different issues in different subsystems: FineGrainedFP8Config vs PagedAttentionCache linear_attention crash." + }, + { + "left": "issue:43317", + "right": "issue:43482", + "accept": false, + "reason": "GPU+CPU offload dequantized-model loading and GGUF loading in v5 are distinct failures." + }, + { + "left": "issue:44849", + "right": "issue:45325", + "accept": false, + "reason": "Qwen hidden-states behavior and Qwen2.5-VL rope-index scaling are different bugs." + }, + { + "left": "issue:43404", + "right": "issue:44466", + "accept": false, + "reason": "Both mention tied weights, but one is a Mistral3 model definition bug and the other is device-dependent serialization; not clearly the same fix." + }, + { + "left": "issue:36032", + "right": "issue:43645", + "accept": false, + "reason": "Tokenizer special-token conflict is unrelated to custom model initialization in notebooks." + }, + { + "left": "issue:45084", + "right": "issue:45507", + "accept": false, + "reason": "Template compilation failure and GraniteMoEHybrid invalid method call are unrelated." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The cluster read is mostly grounded in the packet: the chosen anchor issue 43611 is a reasonable broad v5 loading-regression representative, and the rejected soft pairs are generally conservative and tied to the issue titles. The only accepted GGUF pair is phrased as plausible rather than certain, which keeps it within bounds. No mergeability concerns apply here because these are issues, not PR pairs." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 8, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:36032", + "issue:40990", + "issue:41553", + "issue:42175", + "issue:42915", + "issue:43262", + "issue:43296", + "issue:43317", + "issue:43322", + "issue:43377", + "issue:43404", + "issue:43425", + "issue:43531", + "issue:43540", + "issue:43582", + "issue:43611", + "issue:43632", + "issue:43645", + "issue:43646", + "issue:43742", + "issue:43818", + "issue:43819", + "issue:43856", + "issue:43994", + "issue:44062", + "issue:44220", + "issue:44265", + "issue:44291", + "issue:44403", + "issue:44451", + "issue:44479", + "issue:44534", + "issue:44589", + "issue:44610", + "issue:44683", + "issue:44756", + "issue:44811", + "issue:44877", + "issue:44987", + "issue:44991", + "issue:45081", + "issue:45084", + "issue:45127", + "issue:45216", + "issue:45325", + "issue:45357", + "issue:45542" + ], + "soft_pairs": [ + "issue:43262|issue:43377", + "issue:43646|issue:44534", + "issue:43994|issue:44610", + "issue:44987|issue:44991", + "issue:44403|issue:45081", + "issue:43322|issue:43540", + "issue:43377|issue:43819", + "issue:44220|issue:44683", + "issue:45216|issue:45357", + "issue:44451|issue:45081", + "issue:42915|issue:45084", + "issue:43645|issue:44534", + "issue:44265|issue:44610", + "issue:44987|issue:45081", + "issue:43404|issue:45127", + "issue:41553|issue:43582", + "issue:43632|issue:44756", + "issue:43296|issue:43531", + "issue:44062|issue:44987", + "issue:42915|issue:44291", + "issue:36032|issue:43742", + "issue:43531|issue:44877", + "issue:43611|issue:44534", + "issue:42915|issue:44589", + "issue:40990|issue:44811", + "issue:44479|issue:45325", + "issue:43317|issue:43856", + "issue:42175|issue:45542", + "issue:43425|issue:44403", + "issue:43818|issue:45357" + ], + "prepared_review_unit_hash": "4e64e49d7f90b6fe2920451b49ddb20f97a5039dd2df1390722f78b44cef57d1", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 30, + "serialized_chars": 36043, + "estimated_input_tokens": 9011, + "estimated_eval_tokens": 18278 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly unrelated issue reports with broad Transformers/v5 overlap. The Qwen3.5 save_pretrained pair is the only close match, but it should be treated as likely related rather than definitively the same bug from titles alone.", + "confidence": 0.69, + "canonical_issue_reason": "issue:45216 is the broadest Qwen3.5 save_pretrained regression report and the best anchor for that mini-cluster, but 45357 should only be considered likely related without stronger body-level evidence.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45216 is the best issue anchor because it names the core Qwen3.5 save_pretrained regression and is broader than the later, more specific report.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43262", + "right": "issue:43377", + "accept": false, + "reason": "Different audio bugs: sampling-rate defaulting in apply_chat_template vs batched/single mismatch from missing padding-mask support." + }, + { + "left": "issue:43646", + "right": "issue:44534", + "accept": false, + "reason": "Different v5 regressions: custom model initialization vs non-persistent buffers being filled with junk." + }, + { + "left": "issue:43994", + "right": "issue:44610", + "accept": false, + "reason": "Different models and failure modes: SigLIP2 bad outputs vs OmDet-Turbo processor/model input-size mismatch." + }, + { + "left": "issue:44987", + "right": "issue:44991", + "accept": false, + "reason": "Different affected models and code paths: loading physical-intelligence/fast vs tokenizer loading for EMBEDDIA/est-roberta." + }, + { + "left": "issue:44403", + "right": "issue:45081", + "accept": false, + "reason": "Unrelated problems: generic loading noise vs a Mistral tokenizer regex patch crash." + }, + { + "left": "issue:43322", + "right": "issue:43540", + "accept": false, + "reason": "Different multimodal models and bugs: Llava Next segfault vs Qwen3OmniMoe video-processing ValueError." + }, + { + "left": "issue:43377", + "right": "issue:43819", + "accept": false, + "reason": "Both are audio/model correctness issues, but they affect different models and distinct mechanisms: padding masks vs missing STE in DAC." + }, + { + "left": "issue:44220", + "right": "issue:44683", + "accept": false, + "reason": "Different subsystems: audio fbank extraction vs compiled flex_attention on newer torch." + }, + { + "left": "issue:45216", + "right": "issue:45357", + "accept": false, + "reason": "Likely related Qwen3.5 save_pretrained regressions, but the titles alone do not prove the same root cause or that 45357 is a direct manifestation of 45216." + }, + { + "left": "issue:44451", + "right": "issue:45081", + "accept": false, + "reason": "Different models and failures: ScandiBERT load failure vs Mistral regex-patching crash." + }, + { + "left": "issue:42915", + "right": "issue:45084", + "accept": false, + "reason": "Different bug classes: FineGrainedFP8Config load failure vs template-node compilation TypeError." + }, + { + "left": "issue:43645", + "right": "issue:44534", + "accept": false, + "reason": "Different regressions: notebook/custom-model initialization issues vs corrupted non-persistent buffers." + }, + { + "left": "issue:44265", + "right": "issue:44610", + "accept": false, + "reason": "Different code paths: torch.export with torch_compilable_check vs processor output size mismatch." + }, + { + "left": "issue:44987", + "right": "issue:45081", + "accept": false, + "reason": "Different affected components: model loading for physical-intelligence/fast vs Mistral tokenizer regex patching." + }, + { + "left": "issue:43404", + "right": "issue:45127", + "accept": false, + "reason": "Related to weight tying in broad terms, but one is Mistral3 lm_head tying and the other is LoRA merge with extended vocab." + }, + { + "left": "issue:41553", + "right": "issue:43582", + "accept": false, + "reason": "Different domains: AutoTokenizer error messaging for Voxtral vs Apple Silicon warmup TypeError." + }, + { + "left": "issue:43632", + "right": "issue:44756", + "accept": false, + "reason": "Different regressions: _is_hf_initialized handling vs mmap OOM behavior on Strix Halo." + }, + { + "left": "issue:43296", + "right": "issue:43531", + "accept": false, + "reason": "Different models and bugs: PaddleOCR-VL load failure in vLLM vs Qwen3-MoE sliding_window issue." + }, + { + "left": "issue:44062", + "right": "issue:44987", + "accept": false, + "reason": "Tokenizer AddedToken keyword duplication is unrelated to loading physical-intelligence/fast." + }, + { + "left": "issue:42915", + "right": "issue:44291", + "accept": false, + "reason": "Different failures in different paths: Qwen3Moe FP8 config vs unexpected init_empty_weights argument." + }, + { + "left": "issue:36032", + "right": "issue:43742", + "accept": false, + "reason": "Different model/tokenizer failures: T5Tokenizer special-token conflict vs MobileLLM key error." + }, + { + "left": "issue:43531", + "right": "issue:44877", + "accept": false, + "reason": "Different areas and symptoms: Qwen3-MoE sliding-window behavior vs granite_speech strict-config loading." + }, + { + "left": "issue:43611", + "right": "issue:44534", + "accept": false, + "reason": "Different v5 regressions: base_model_prefix loading breakage vs junk values in non-persistent buffers." + }, + { + "left": "issue:42915", + "right": "issue:44589", + "accept": false, + "reason": "Different failures in different paths: Qwen3Moe FP8 config vs missing Float8 storage type." + }, + { + "left": "issue:40990", + "right": "issue:44811", + "accept": false, + "reason": "Unrelated issues: perplexity regression on gpt-oss-20b vs Whisper batch_decode special-token handling." + }, + { + "left": "issue:44479", + "right": "issue:45325", + "accept": false, + "reason": "Both involve video/VL behavior, but the affected models and concrete bugs differ: video-input regression vs rope-index scaling." + }, + { + "left": "issue:43317", + "right": "issue:43856", + "accept": false, + "reason": "Different problems: device_map/offload loading failure vs Qwen3 MoE training memory usage." + }, + { + "left": "issue:42175", + "right": "issue:45542", + "accept": false, + "reason": "Related TensorFlow packaging themes, but these are different packaging/backend failures and not the same bug." + }, + { + "left": "issue:43425", + "right": "issue:44403", + "accept": false, + "reason": "Torch version incompatibility is unrelated to generic transformer-loading noise." + }, + { + "left": "issue:43818", + "right": "issue:45357", + "accept": false, + "reason": "Different models and code paths: Video-LLaVA weight/temporal-attention issues vs Qwen3.5 save_pretrained serialization." + } + ] + }, + "evaluator_result": { + "accept": false, + "feedback": "The cluster summary is mostly grounded, but the accepted 45216\u219445357 linkage is stated too strongly for the packet. The titles support that both are Qwen3.5 save_pretrained regressions, but not that they are definitively the same underlying bug or that 45357 is proven to be a more specific manifestation of 45216. Please soften that verdict to something like 'likely related' unless you have body-level evidence." + }, + "evaluator_used": true, + "retried": true, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 9, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:33666", + "issue:35707", + "issue:36032", + "issue:40444", + "issue:41720", + "issue:42175", + "issue:42915", + "issue:43296", + "issue:43377", + "issue:43421", + "issue:43454", + "issue:43482", + "issue:43525", + "issue:43531", + "issue:43611", + "issue:43632", + "issue:43644", + "issue:43646", + "issue:43792", + "issue:43828", + "issue:43881", + "issue:43957", + "issue:43994", + "issue:44079", + "issue:44265", + "issue:44291", + "issue:44387", + "issue:44403", + "issue:44466", + "issue:44488", + "issue:44534", + "issue:44568", + "issue:44589", + "issue:44617", + "issue:44683", + "issue:44829", + "issue:44877", + "issue:44898", + "issue:44991", + "issue:45020", + "issue:45072", + "issue:45081", + "issue:45325", + "issue:45381", + "issue:45479", + "issue:45491", + "issue:45538", + "issue:45663" + ], + "soft_pairs": [ + "issue:43421|issue:44568", + "issue:44079|issue:45663", + "issue:44683|issue:45020", + "issue:36032|issue:41720", + "issue:43632|issue:44534", + "issue:43792|issue:44589", + "issue:44403|issue:44991", + "issue:43482|issue:45538", + "issue:33666|issue:40444", + "issue:44829|issue:45479", + "issue:40444|issue:45325", + "issue:43482|issue:43525", + "issue:44291|issue:44589", + "issue:43377|issue:45491", + "issue:41720|issue:45081", + "issue:44265|issue:45072", + "issue:42175|issue:43296", + "issue:43632|issue:45072", + "issue:43611|issue:43646", + "issue:44079|issue:44534", + "issue:43531|issue:44291", + "issue:43994|issue:44898", + "issue:43454|issue:44466", + "issue:43881|issue:44488", + "issue:43828|issue:44683", + "issue:42915|issue:43482", + "issue:44079|issue:45072", + "issue:43792|issue:43957", + "issue:43644|issue:44079", + "issue:40444|issue:45381", + "issue:45325|issue:45381", + "issue:44387|issue:44617", + "issue:35707|issue:40444", + "issue:44291|issue:44877" + ], + "prepared_review_unit_hash": "d9e3f7ba65062e2ca027a88be24c1586969e49e92cc5f3e2ac8ee32fa5d53156", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 34, + "serialized_chars": 38070, + "estimated_input_tokens": 9518, + "estimated_eval_tokens": 19292 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly a collection of separate Transformers v5 / model-loading regressions, not one duplicate set. The closest center is a broad initialization/loading regression, but most pairs are only thematically similar.", + "confidence": 0.38, + "canonical_issue_reason": "issue:43632 is the strongest concrete anchor for the cluster: a core Transformers v5 initialization regression (`_is_hf_initialized`) that fits the many nearby model-loading breakages better than the narrower, model-specific reports.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45020 is the best cluster representative overall because it is the broadest umbrella for the shared theme of recent Transformers versions breaking model loading, especially for remote-code models.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43421", + "right": "issue:44568", + "accept": false, + "reason": "Both are tokenizer/special-token regressions, but they describe different failure modes and code paths: runtime post-processor updates vs BOS/EOS insertion." + }, + { + "left": "issue:44079", + "right": "issue:45663", + "accept": false, + "reason": "One is a ModelOutput key assignment bug; the other is a Gemma-4 FSDP2 attention state KeyError. They are unrelated." + }, + { + "left": "issue:44683", + "right": "issue:45020", + "accept": false, + "reason": "torch >= 2.9 flex_attention compilation failure is a different problem from remote_code model loading regressions." + }, + { + "left": "issue:36032", + "right": "issue:41720", + "accept": false, + "reason": "Tokenizer method-name conflict and Qwen3 auto device mapping CUDA assert are unrelated." + }, + { + "left": "issue:43632", + "right": "issue:44534", + "accept": false, + "reason": "Both are v5 regressions, but one concerns `_is_hf_initialized` during init and the other concerns non-persistent buffer contents; different bugs." + }, + { + "left": "issue:43792", + "right": "issue:44589", + "accept": false, + "reason": "Whisper loading failure and missing Float8 storage are distinct load/runtime issues." + }, + { + "left": "issue:44403", + "right": "issue:44991", + "accept": false, + "reason": "Generic load noise and a specific tokenizer load failure for EMBEDDIA/est-roberta are not the same bug." + }, + { + "left": "issue:43482", + "right": "issue:45538", + "accept": false, + "reason": "GGUF loading under v5 and CLIPTokenizer max length behavior are unrelated." + }, + { + "left": "issue:33666", + "right": "issue:40444", + "accept": false, + "reason": "Both involve Qwen VL training, but one is multi-GPU training and the other is iterable-dataset/multi-image prompt handling; not the same issue." + }, + { + "left": "issue:44829", + "right": "issue:45479", + "accept": false, + "reason": "Degenerate training with flash_attention_3 is a different defect from the single-label-classification zero-loss bug." + }, + { + "left": "issue:40444", + "right": "issue:45325", + "accept": false, + "reason": "Both are Qwen2.5-VL problems, but they affect different paths: dataset/multi-image finetuning vs still-image temporal position-id scaling." + }, + { + "left": "issue:43482", + "right": "issue:43525", + "accept": false, + "reason": "Qwen2.5-GGUF loading failure and missing pad_token_id on Llama4Config are unrelated loading errors." + }, + { + "left": "issue:44291", + "right": "issue:44589", + "accept": false, + "reason": "Unexpected `_is_hf_initialized` during init-empty-weights is not the same as the Float8 storage lookup failure." + }, + { + "left": "issue:43377", + "right": "issue:45491", + "accept": false, + "reason": "Missing padding-mask support in MIMI encoder and Gemma3 sliding-window NaNs are separate batching/attention bugs." + }, + { + "left": "issue:41720", + "right": "issue:45081", + "accept": false, + "reason": "Qwen3 device-mapping CUDA asserts and a Mistral tokenizer regex crash are unrelated." + }, + { + "left": "issue:44265", + "right": "issue:45072", + "accept": false, + "reason": "torch.export/torch_compilable_check failure is distinct from dtype mismatches in SwitchTransformers/TimmWrapperModel." + }, + { + "left": "issue:42175", + "right": "issue:43296", + "accept": false, + "reason": "Packaging/backend inclusion issue and PaddleOCR-VL loading in vLLM are unrelated." + }, + { + "left": "issue:43632", + "right": "issue:45072", + "accept": false, + "reason": "Initialization-flag regression and bf16 dtype mismatch are different code-path problems." + }, + { + "left": "issue:43611", + "right": "issue:43646", + "accept": false, + "reason": "Both are broad v5 loading regressions, but one is about `base_model_prefix` and the other about custom model initialization; not the same concrete bug." + }, + { + "left": "issue:44079", + "right": "issue:44534", + "accept": false, + "reason": "ModelOutput key handling and junk-filled non-persistent buffers are unrelated." + }, + { + "left": "issue:43531", + "right": "issue:44291", + "accept": false, + "reason": "Qwen3-MoE sliding_window behavior and init-empty-weights `_is_hf_initialized` are separate issues." + }, + { + "left": "issue:43994", + "right": "issue:44898", + "accept": false, + "reason": "SigLIP2 wrong outputs and Perceiver resolution/interpolation failure are different model-specific bugs." + }, + { + "left": "issue:43454", + "right": "issue:44466", + "accept": false, + "reason": "Tied lm_head weights not being set and serialization differences of lm_head.weight are related only thematically, not the same failure." + }, + { + "left": "issue:43881", + "right": "issue:44488", + "accept": false, + "reason": "glm-4v-9b loading failure and cjvt/sleng-bert tokenizer loading failure are unrelated." + }, + { + "left": "issue:43828", + "right": "issue:44683", + "accept": false, + "reason": "Autocast dtype mismatch and compiled flex_attention failure are different problems." + }, + { + "left": "issue:42915", + "right": "issue:43482", + "accept": false, + "reason": "FineGrainedFP8Config failure and GGUF loading under v5 are unrelated." + }, + { + "left": "issue:44079", + "right": "issue:45072", + "accept": false, + "reason": "ModelOutput key assignment bug and bf16 dtype mismatch are not the same underlying issue." + }, + { + "left": "issue:43792", + "right": "issue:43957", + "accept": false, + "reason": "Whisper load failure and meta-device loading regression are distinct." + }, + { + "left": "issue:43644", + "right": "issue:44079", + "accept": false, + "reason": "Non-persistent buffer serialization junk and ModelOutput key assignment are different bugs." + }, + { + "left": "issue:40444", + "right": "issue:45381", + "accept": false, + "reason": "Qwen2.5-VL multi-image iterable training failure and video vision_position_ids bug affect different paths." + }, + { + "left": "issue:45325", + "right": "issue:45381", + "accept": false, + "reason": "Both are Qwen2.5-VL position-id complaints, but one is still-image temporal scaling and the other is video vision_position_ids; close, but not clearly the same concrete bug." + }, + { + "left": "issue:44387", + "right": "issue:44617", + "accept": false, + "reason": "v5 int4 memory growth/OOM and Sam3Video OOM are different resource issues." + }, + { + "left": "issue:35707", + "right": "issue:40444", + "accept": false, + "reason": "Progressive generation with inputs_embeds/past_key_values is unrelated to Qwen2.5-VL multi-image iterable finetuning." + }, + { + "left": "issue:44291", + "right": "issue:44877", + "accept": false, + "reason": "Unexpected `_is_hf_initialized` argument and strict config blocking granite_speech are different loading failures." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 10, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:34634", + "issue:35707", + "issue:39692", + "issue:40444", + "issue:40990", + "issue:41720", + "issue:43262", + "issue:43296", + "issue:43317", + "issue:43377", + "issue:43404", + "issue:43454", + "issue:43526", + "issue:43550", + "issue:43572", + "issue:43577", + "issue:43606", + "issue:43632", + "issue:43701", + "issue:43708", + "issue:43720", + "issue:43746", + "issue:43792", + "issue:43819", + "issue:43873", + "issue:43881", + "issue:43883", + "issue:44186", + "issue:44265", + "issue:44423", + "issue:44466", + "issue:44610", + "issue:44683", + "issue:44756", + "issue:44877", + "issue:44898", + "issue:44977", + "issue:44991", + "issue:45072", + "issue:45325", + "issue:45357", + "issue:45406", + "issue:45563", + "issue:45584", + "issue:45632", + "issue:45684", + "issue:45701", + "issue:45759" + ], + "soft_pairs": [ + "issue:43526|issue:43883", + "issue:43317|issue:43873", + "issue:43606|issue:44265", + "issue:43881|issue:44991", + "issue:43701|issue:45325", + "issue:44186|issue:44265", + "issue:43632|issue:45563", + "issue:43262|issue:43746", + "issue:43606|issue:43819", + "issue:44610|issue:45325", + "issue:43454|issue:43550", + "issue:43377|issue:44610", + "issue:43454|issue:45072", + "issue:40990|issue:45584", + "issue:43572|issue:45759", + "issue:43377|issue:44898", + "issue:39692|issue:40444", + "issue:44466|issue:45357", + "issue:43792|issue:44683", + "issue:43296|issue:44991", + "issue:43296|issue:44877", + "issue:43454|issue:43720", + "issue:43873|issue:45701", + "issue:41720|issue:44977", + "issue:45632|issue:45684", + "issue:43404|issue:45357", + "issue:44423|issue:45406", + "issue:44186|issue:45072", + "issue:43577|issue:43708", + "issue:34634|issue:35707", + "issue:35707|issue:41720", + "issue:43873|issue:44756" + ], + "prepared_review_unit_hash": "83a9dcad506494e60c7e5ecedfa7d64dda1259d04ebb768fdcd1a5015bc50775", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 32, + "serialized_chars": 37539, + "estimated_input_tokens": 9385, + "estimated_eval_tokens": 19026 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Mostly unrelated closed issues with only a loose shared theme around loading/offloading, config, and multimodal runtime regressions. None of the soft pairs look like true duplicates, so I reject them all and center the cluster on the broad offloading/quantization report.", + "confidence": 0.71, + "canonical_issue_reason": "issue:43873 is the broadest and most central report in the small offloading/quantization subset, with additional discussion and inbound references.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43873 is the best representative issue because it captures the widest symptom area among the related loading/offloading reports.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43526", + "right": "issue:43883", + "accept": false, + "reason": "Different components and failures: BeitImageProcessorFast label reduction vs Molmo missing attribute during loading." + }, + { + "left": "issue:43317", + "right": "issue:43873", + "accept": false, + "reason": "Both mention offloading/quantization, but one is a dequantized-model load failure and the other is a generic offloading complaint; not the same bug." + }, + { + "left": "issue:43606", + "right": "issue:44265", + "accept": false, + "reason": "CPU offload device mismatch is unrelated to torch.export failures around torch_compilable_check." + }, + { + "left": "issue:43881", + "right": "issue:44991", + "accept": false, + "reason": "PaddleOCR-VL model loading in vLLM and tokenizer loading for est-roberta are different failure modes." + }, + { + "left": "issue:43701", + "right": "issue:45325", + "accept": false, + "reason": "Checkpoint-resume key mismatch is unrelated to Qwen2.5-VL rope-index scaling." + }, + { + "left": "issue:44186", + "right": "issue:44265", + "accept": false, + "reason": "Tokenizer crashes on NER/padding are unrelated to torch.export compatibility issues." + }, + { + "left": "issue:43632", + "right": "issue:45563", + "accept": false, + "reason": "_is_hf_initialized flag breakage is unrelated to a stale generate warning." + }, + { + "left": "issue:43262", + "right": "issue:43746", + "accept": false, + "reason": "Audio sampling-rate defaulting and PEFT adapter local-checkpoint loading are different code paths." + }, + { + "left": "issue:43606", + "right": "issue:43819", + "accept": false, + "reason": "CPU offload device mismatch and DAC.from_latents/STE mismatch are unrelated model bugs." + }, + { + "left": "issue:44610", + "right": "issue:45325", + "accept": false, + "reason": "Processor image size mismatch is unrelated to temporal position-id scaling in Qwen2.5-VL." + }, + { + "left": "issue:43454", + "right": "issue:43550", + "accept": false, + "reason": "lm_head weight tying and torch.compile SDPA failures are not the same underlying issue." + }, + { + "left": "issue:43377", + "right": "issue:44610", + "accept": false, + "reason": "MIMI batching/padding-mask behavior and OmDet processor resolution mismatch are unrelated." + }, + { + "left": "issue:43454", + "right": "issue:45072", + "accept": false, + "reason": "Weight-tying bugs are unrelated to dtype mismatch issues in bfloat16 inference." + }, + { + "left": "issue:40990", + "right": "issue:45584", + "accept": false, + "reason": "High perplexity on GPT-OSS and Whisper empty-transcription failure are unrelated generation problems." + }, + { + "left": "issue:43572", + "right": "issue:45759", + "accept": false, + "reason": "Missing pad_token_idx in StableLmConfig is unrelated to unwrapping text_config for Qwen multimodal configs." + }, + { + "left": "issue:43377", + "right": "issue:44898", + "accept": false, + "reason": "Missing padding-mask support in MIMI encoder is unrelated to Perceiver positional encoding interpolation." + }, + { + "left": "issue:39692", + "right": "issue:40444", + "accept": false, + "reason": "SigLIP2 doc/example errors and Qwen2.5-VL iterable-dataset finetuning failures are different multimodal issues." + }, + { + "left": "issue:44466", + "right": "issue:45357", + "accept": false, + "reason": "lm_head.weight serialization inconsistency and incorrect visual encoder keys are both save_pretrained-related but not the same bug." + }, + { + "left": "issue:43792", + "right": "issue:44683", + "accept": false, + "reason": "Whisper runtime failure and compiled flex_attention failure are unrelated." + }, + { + "left": "issue:43296", + "right": "issue:44991", + "accept": false, + "reason": "PaddleOCR-VL load failure in vLLM is unrelated to tokenizer loading for est-roberta." + }, + { + "left": "issue:43296", + "right": "issue:44877", + "accept": false, + "reason": "Strict config rejection for granite_speech is a different loading/config bug than PaddleOCR-VL model loading." + }, + { + "left": "issue:43454", + "right": "issue:43720", + "accept": false, + "reason": "lm_head weight tying is unrelated to BitNet packed-weight unpacking during accelerate loading." + }, + { + "left": "issue:43873", + "right": "issue:45701", + "accept": false, + "reason": "Generic offloading/quantization issues do not match a tokenization-version regression." + }, + { + "left": "issue:41720", + "right": "issue:44977", + "accept": false, + "reason": "Both are Qwen-family runtime bugs, but auto device mapping cuda asserts and flash-attention generation failures are different code paths." + }, + { + "left": "issue:45632", + "right": "issue:45684", + "accept": false, + "reason": "Cache-path collisions for trust_remote_code are unrelated to permission propagation in save_pretrained." + }, + { + "left": "issue:43404", + "right": "issue:45357", + "accept": false, + "reason": "A missing tie_weights call and incorrect visual encoder keys are different serialization/model-init bugs." + }, + { + "left": "issue:44423", + "right": "issue:45406", + "accept": false, + "reason": "Continuous-batching multimodal crash and Gemma4Processor missing _tokenizer are related to serve, but not the same concrete defect." + }, + { + "left": "issue:44186", + "right": "issue:45072", + "accept": false, + "reason": "Tokenizer padding/truncation crashes are unrelated to bfloat16 dtype mismatches." + }, + { + "left": "issue:43577", + "right": "issue:43708", + "accept": false, + "reason": "dtype propagation for BLIP2 is unrelated to Trainer resume max_steps calculation." + }, + { + "left": "issue:34634", + "right": "issue:35707", + "accept": false, + "reason": "Bark voice_preset failure and progressive generation with inputs_embeds/past_key_values are unrelated." + }, + { + "left": "issue:35707", + "right": "issue:41720", + "accept": false, + "reason": "Progressive generation state handling is unrelated to Qwen3 auto-device-map cuda asserts." + }, + { + "left": "issue:43873", + "right": "issue:44756", + "accept": false, + "reason": "Generic offloading/quantization problems are not the same as disabling mmap to avoid OOM on Strix Halo." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 11, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:30064", + "issue:34634", + "issue:36010", + "issue:36331", + "issue:40990", + "issue:41720", + "issue:42617", + "issue:42915", + "issue:43295", + "issue:43316", + "issue:43317", + "issue:43322", + "issue:43388", + "issue:43421", + "issue:43450", + "issue:43526", + "issue:43531", + "issue:43540", + "issue:43550", + "issue:43572", + "issue:43606", + "issue:43720", + "issue:43749", + "issue:43782", + "issue:43881", + "issue:44186", + "issue:44292", + "issue:44361", + "issue:44387", + "issue:44479", + "issue:44492", + "issue:44509", + "issue:44534", + "issue:44610", + "issue:44671", + "issue:44743", + "issue:44877", + "issue:44898", + "issue:44913", + "issue:44918", + "issue:44991", + "issue:45059", + "issue:45161", + "issue:45406", + "issue:45464", + "issue:45701", + "issue:45704" + ], + "soft_pairs": [ + "issue:43749|issue:44534", + "issue:43526|issue:43881", + "issue:44877|issue:44991", + "issue:42617|issue:42915", + "issue:43450|issue:45701", + "issue:43550|issue:43720", + "issue:36331|issue:42915", + "issue:44610|issue:44898", + "issue:44186|issue:44610", + "issue:44387|issue:45704", + "issue:44479|issue:44918", + "issue:36331|issue:42617", + "issue:43316|issue:44913", + "issue:42915|issue:43317", + "issue:40990|issue:45701", + "issue:41720|issue:43322", + "issue:43295|issue:45701", + "issue:43550|issue:43606", + "issue:41720|issue:45161", + "issue:44671|issue:44743", + "issue:43531|issue:43572", + "issue:44292|issue:44479", + "issue:41720|issue:42617", + "issue:36010|issue:42617", + "issue:44186|issue:44361", + "issue:43540|issue:43782", + "issue:45406|issue:45464", + "issue:41720|issue:44292", + "issue:43421|issue:45701", + "issue:30064|issue:34634", + "issue:44492|issue:44509", + "issue:43388|issue:45059", + "issue:42915|issue:43322", + "issue:43606|issue:43720" + ], + "prepared_review_unit_hash": "8604d3d40b1a2b07cac2e63c172aeb3200af909eea5af0ad884b7905620ed748", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 34, + "serialized_chars": 36948, + "estimated_input_tokens": 9237, + "estimated_eval_tokens": 18730 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The soft pairs are only superficially similar; they span unrelated models, APIs, and failure modes. I would not merge any of these pairs as duplicates.", + "confidence": 0.94, + "canonical_issue_reason": "No clear canonical issue: this set is heterogeneous, with distinct bugs across loading, tokenization, multimodal processing, quantization, and serving paths.", + "canonical_pr_reason": null, + "best_issue_reason": "No global best issue candidate; none of the items subsumes the others or represents the shared underlying defect.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43749", + "right": "issue:44534", + "accept": false, + "reason": "FSDP CPU/RAM-efficient loading vs non-persistent buffer initialization are different code paths and symptoms." + }, + { + "left": "issue:43526", + "right": "issue:43881", + "accept": false, + "reason": "A BEiT image-processor label bug and a GLM-4V loading failure are unrelated model-specific issues." + }, + { + "left": "issue:44877", + "right": "issue:44991", + "accept": false, + "reason": "Both involve loading regressions, but one is strict config handling and the other is tokenizer loading; not the same defect." + }, + { + "left": "issue:42617", + "right": "issue:42915", + "accept": false, + "reason": "Different failures: a 3d_parallel.py runtime problem versus a Qwen3MoE FP8 config issue." + }, + { + "left": "issue:43450", + "right": "issue:45701", + "accept": false, + "reason": "A batched video-shape bug is unrelated to a general tokenization-version regression." + }, + { + "left": "issue:43550", + "right": "issue:43720", + "accept": false, + "reason": "These are different model/runtime bugs: torch.compile+SDPA for Bamba versus packed-weight loading for BitNet." + }, + { + "left": "issue:36331", + "right": "issue:42915", + "accept": false, + "reason": "Trainer API keyword mismatch and Qwen3MoE FP8 loading are distinct problems." + }, + { + "left": "issue:44610", + "right": "issue:44898", + "accept": false, + "reason": "Both mention image sizes, but they affect different models and different resize/interpolation logic." + }, + { + "left": "issue:44186", + "right": "issue:44610", + "accept": false, + "reason": "LayoutLMv2 tokenizer crashes and OmDet-Turbo image-size mismatch are unrelated." + }, + { + "left": "issue:44387", + "right": "issue:45704", + "accept": false, + "reason": "CUDA reserved-memory growth under int4 quantization is not the same as an apex RMSNorm memory leak." + }, + { + "left": "issue:44479", + "right": "issue:44918", + "accept": false, + "reason": "One is a video-input regression in Qwen VL models; the other is embedding unpacking with TRL SFT." + }, + { + "left": "issue:36331", + "right": "issue:42617", + "accept": false, + "reason": "Trainer compute_loss signature breakage is unrelated to the 3d_parallel.py failure." + }, + { + "left": "issue:43316", + "right": "issue:44913", + "accept": false, + "reason": "Config API discrepancy in Gemma3TextConfig is not the same as GPTNeoX rotary_pct reload behavior." + }, + { + "left": "issue:42915", + "right": "issue:43317", + "accept": false, + "reason": "Both are loading-related, but one is FP8 Qwen3MoE and the other is dequantized model offload with device_map=auto." + }, + { + "left": "issue:40990", + "right": "issue:45701", + "accept": false, + "reason": "High perplexity on a model/dataset pair is not the same as a tokenizer behavior change." + }, + { + "left": "issue:41720", + "right": "issue:43322", + "accept": false, + "reason": "Qwen3 auto device-map assertions and Llava Next loading segfaults are different failures." + }, + { + "left": "issue:43295", + "right": "issue:45701", + "accept": false, + "reason": "Processor/tokenizer API regression and general tokenization-version changes are not the same code path." + }, + { + "left": "issue:43550", + "right": "issue:43606", + "accept": false, + "reason": "Different models and different offload/compile issues; only the broad 'device mismatch' theme overlaps." + }, + { + "left": "issue:41720", + "right": "issue:45161", + "accept": false, + "reason": "Auto device mapping on Qwen3 is not the same as tensor-parallel GPT-OSS MoE behavior." + }, + { + "left": "issue:44671", + "right": "issue:44743", + "accept": false, + "reason": "CamemBERT MLM prediction error and Qwen3.5 recurrent-state reset are unrelated." + }, + { + "left": "issue:43531", + "right": "issue:43572", + "accept": false, + "reason": "Sliding-window behavior in Qwen3-MoE and missing pad_token_idx in StableLmConfig are different bugs." + }, + { + "left": "issue:44292", + "right": "issue:44479", + "accept": false, + "reason": "NVFP4 run failure and Qwen VL video regression are unrelated model-specific issues." + }, + { + "left": "issue:41720", + "right": "issue:42617", + "accept": false, + "reason": "Auto device mapping assert and 3d_parallel.py runtime failure are distinct problems." + }, + { + "left": "issue:36010", + "right": "issue:42617", + "accept": false, + "reason": "GenerationMixin import failure is unrelated to the 3d_parallel.py issue." + }, + { + "left": "issue:44186", + "right": "issue:44361", + "accept": false, + "reason": "Two different tokenizer bugs affecting different models and tasks." + }, + { + "left": "issue:43540", + "right": "issue:43782", + "accept": false, + "reason": "Qwen3OmniMoe video-processing ValueError and Qwen3VL weight_only load error are different code paths." + }, + { + "left": "issue:45406", + "right": "issue:45464", + "accept": false, + "reason": "Gemma4Processor missing _tokenizer and Qwen3.5 streaming inference failure are separate serving bugs." + }, + { + "left": "issue:41720", + "right": "issue:44292", + "accept": false, + "reason": "Auto device mapping on Qwen3 and Qwen-3-8B-NVFP4 runtime failure are not the same defect." + }, + { + "left": "issue:43421", + "right": "issue:45701", + "accept": false, + "reason": "Runtime special-token post-processor updates and general tokenization-version changes overlap only loosely; not a duplicate." + }, + { + "left": "issue:30064", + "right": "issue:34634", + "accept": false, + "reason": "Void segmentation-map processing and Bark voice_preset issues are unrelated." + }, + { + "left": "issue:44492", + "right": "issue:44509", + "accept": false, + "reason": "A cache-strategy typo and stale docs for removed pipeline tasks are different documentation issues." + }, + { + "left": "issue:43388", + "right": "issue:45059", + "accept": false, + "reason": "Metric label truncation and SAM3 text/box behavior are unrelated." + }, + { + "left": "issue:42915", + "right": "issue:43322", + "accept": false, + "reason": "Qwen3MoE FP8 loading and Llava Next segfaults are different model-loading failures." + }, + { + "left": "issue:43606", + "right": "issue:43720", + "accept": false, + "reason": "CPU offload device mismatch for Bark and packed-weight unpacking for BitNet are distinct bugs." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 12, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:33666", + "issue:34567", + "issue:34634", + "issue:35707", + "issue:39692", + "issue:41720", + "issue:42886", + "issue:43381", + "issue:43421", + "issue:43525", + "issue:43531", + "issue:43550", + "issue:43596", + "issue:43606", + "issue:43653", + "issue:43746", + "issue:43782", + "issue:43844", + "issue:43873", + "issue:43927", + "issue:43994", + "issue:44062", + "issue:44186", + "issue:44220", + "issue:44368", + "issue:44451", + "issue:44514", + "issue:44545", + "issue:44589", + "issue:44610", + "issue:44625", + "issue:44671", + "issue:44843", + "issue:44898", + "issue:44918", + "issue:44936", + "issue:44987", + "issue:44991", + "issue:45042", + "issue:45072", + "issue:45081", + "issue:45356", + "issue:45357", + "issue:45381", + "issue:45479", + "issue:45588", + "issue:45704", + "issue:45759" + ], + "soft_pairs": [ + "issue:44514|issue:44545", + "issue:43550|issue:43653", + "issue:43531|issue:45759", + "issue:43606|issue:43653", + "issue:43421|issue:43927", + "issue:34567|issue:35707", + "issue:44368|issue:45357", + "issue:44898|issue:45072", + "issue:44451|issue:44991", + "issue:43873|issue:45704", + "issue:41720|issue:43746", + "issue:44062|issue:44451", + "issue:43421|issue:43653", + "issue:42886|issue:44843", + "issue:44987|issue:45042", + "issue:45588|issue:45759", + "issue:33666|issue:35707", + "issue:43531|issue:44589", + "issue:43421|issue:44186", + "issue:43381|issue:43844", + "issue:39692|issue:44610", + "issue:44918|issue:44936", + "issue:45081|issue:45356", + "issue:43525|issue:43531", + "issue:44625|issue:45479", + "issue:43782|issue:45381", + "issue:43994|issue:44671", + "issue:34634|issue:41720", + "issue:43381|issue:45704", + "issue:43596|issue:44220", + "issue:39692|issue:43873" + ], + "prepared_review_unit_hash": "fa6969c618e3bee768fcbf33b464f80887d58bb27f95a72ed22e7c1411fa5b99", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 31, + "serialized_chars": 37274, + "estimated_input_tokens": 9319, + "estimated_eval_tokens": 18894 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The cluster is mostly heterogeneous; only the two Qwen2_5_VLProcessor `apply_chat_template` reports are the same underlying bug. All other suggested pairs appear to be unrelated issues in different models/code paths.", + "confidence": 0.94, + "canonical_issue_reason": "issue:44514 is the earlier, more specific report of the Qwen2_5_VLProcessor batched `apply_chat_template` crash; issue:44545 looks like the later duplicate of the same failure.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:44514 is the clearest representative of the only true duplicate pair in this set.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44514", + "right": "issue:44545", + "accept": true, + "reason": "Same component, same batched `apply_chat_template` crash when `padding=False`; titles describe the same bug." + }, + { + "left": "issue:43550", + "right": "issue:43653", + "accept": false, + "reason": "Different subsystems: torch.compile/SDPA model bug vs tokenizer special-token registration." + }, + { + "left": "issue:43531", + "right": "issue:45759", + "accept": false, + "reason": "Qwen3-MoE sliding-window behavior is unrelated to composite Qwen multimodal config unwrapping." + }, + { + "left": "issue:43606", + "right": "issue:43653", + "accept": false, + "reason": "CPU offload device mismatch for Bark is unrelated to BigBirdTokenizer special-token decoding." + }, + { + "left": "issue:43421", + "right": "issue:43927", + "accept": false, + "reason": "Both touch tokenizer state, but one is runtime post-processor refresh and the other is config persistence/load; not the same bug." + }, + { + "left": "issue:34567", + "right": "issue:35707", + "accept": false, + "reason": "Trainer token counting and progressive generation with embeddings/past key values are unrelated failures." + }, + { + "left": "issue:44368", + "right": "issue:45357", + "accept": false, + "reason": "A warning about `tie_word_embeddings` is not the same as incorrect visual-encoder keys being saved." + }, + { + "left": "issue:44898", + "right": "issue:45072", + "accept": false, + "reason": "Perceiver resolution handling and bfloat16 dtype mismatch in different models are unrelated." + }, + { + "left": "issue:44451", + "right": "issue:44991", + "accept": false, + "reason": "Both are tokenizer-loading regressions, but they affect different models and likely different code paths." + }, + { + "left": "issue:43873", + "right": "issue:45704", + "accept": false, + "reason": "Quantization/offloading behavior is unrelated to a T5 apex RMSNorm memory leak." + }, + { + "left": "issue:41720", + "right": "issue:43746", + "accept": false, + "reason": "Qwen3 auto device-map CUDA assert is unrelated to GraniteSpeech PEFT checkpoint loading." + }, + { + "left": "issue:44062", + "right": "issue:44451", + "accept": false, + "reason": "AddedToken construction error is unrelated to loading ScandiBERT." + }, + { + "left": "issue:43421", + "right": "issue:43653", + "accept": false, + "reason": "Special-token post-processing updates are different from BigBirdTokenizer mask-token registration/decoding." + }, + { + "left": "issue:42886", + "right": "issue:44843", + "accept": false, + "reason": "Both involve offline mode, but one is generic cache loading while the other is a specific `_patch_mistral_regex` hub call." + }, + { + "left": "issue:44987", + "right": "issue:45042", + "accept": false, + "reason": "Loading a specific model in 5.1.0 is unrelated to PIL image processors incorrectly requiring torchvision." + }, + { + "left": "issue:45588", + "right": "issue:45759", + "accept": false, + "reason": "Flash-attention sinkless model crash is unrelated to Qwen multimodal config unwrapping." + }, + { + "left": "issue:33666", + "right": "issue:35707", + "accept": false, + "reason": "Qwen2-VL multi-GPU training and progressive generation are different problems." + }, + { + "left": "issue:43531", + "right": "issue:44589", + "accept": false, + "reason": "Qwen3-MoE sliding-window behavior is unrelated to the Float8 storage lookup error." + }, + { + "left": "issue:43421", + "right": "issue:44186", + "accept": false, + "reason": "Runtime special-token updates are not the same as LayoutLMv2Tokenizer NER/padding crashes." + }, + { + "left": "issue:43381", + "right": "issue:43844", + "accept": false, + "reason": "Eval-mode gradient checkpointing restriction is unrelated to abnormal gradient growth under ZeRO-3." + }, + { + "left": "issue:39692", + "right": "issue:44610", + "accept": false, + "reason": "Both mention processor/model mismatches, but they involve different models and distinct preprocessing bugs." + }, + { + "left": "issue:44918", + "right": "issue:44936", + "accept": false, + "reason": "TRL SFT embedding unpacking and trainer evaluate-after-train failures are different training issues." + }, + { + "left": "issue:45081", + "right": "issue:45356", + "accept": false, + "reason": "Both reference mistral-regex-related tokenization, but one is an AttributeError crash and the other is a codec-handling regression." + }, + { + "left": "issue:43525", + "right": "issue:43531", + "accept": false, + "reason": "Missing `pad_token_id` on Llama4Config is unrelated to Qwen3-MoE sliding-window behavior." + }, + { + "left": "issue:44625", + "right": "issue:45479", + "accept": false, + "reason": "Qwen3.5 config propagation and degenerate single-label classification loss are separate model-classification issues." + }, + { + "left": "issue:43782", + "right": "issue:45381", + "accept": false, + "reason": "Qwen3VL weight-only loading failure is unrelated to Qwen2.5-VL video vision-position IDs." + }, + { + "left": "issue:43994", + "right": "issue:44671", + "accept": false, + "reason": "Incorrect SigLIP2 outputs and CamemBERT MLM regression are different model-specific failures." + }, + { + "left": "issue:34634", + "right": "issue:41720", + "accept": false, + "reason": "BarkProcessor voice preset handling is unrelated to Qwen3 auto device-mapping CUDA asserts." + }, + { + "left": "issue:43381", + "right": "issue:45704", + "accept": false, + "reason": "Gradient checkpointing in eval mode is unrelated to T5\u2019s apex RMSNorm leak." + }, + { + "left": "issue:43596", + "right": "issue:44220", + "accept": false, + "reason": "DeepSpeed ZeRO-3 init with BertModel and `_torch_extract_fbank_features()` are unrelated errors." + }, + { + "left": "issue:39692", + "right": "issue:43873", + "accept": false, + "reason": "SigLIP2 doc-example errors are unrelated to quantization/offloading behavior." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet: the 44514/44545 pair is the only clearly duplicated issue based on the near-identical titles, and the other pair rejections stay conservative rather than overstating sameness." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 13, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:30064", + "issue:36010", + "issue:42915", + "issue:43335", + "issue:43421", + "issue:43531", + "issue:43540", + "issue:43550", + "issue:43575", + "issue:43577", + "issue:43618", + "issue:43643", + "issue:43653", + "issue:43697", + "issue:43716", + "issue:43720", + "issue:43746", + "issue:43756", + "issue:43782", + "issue:43818", + "issue:43819", + "issue:43873", + "issue:44008", + "issue:44062", + "issue:44077", + "issue:44186", + "issue:44220", + "issue:44448", + "issue:44488", + "issue:44561", + "issue:44625", + "issue:44661", + "issue:44671", + "issue:44756", + "issue:44811", + "issue:44843", + "issue:44877", + "issue:44898", + "issue:44977", + "issue:45020", + "issue:45072", + "issue:45081", + "issue:45183", + "issue:45381", + "issue:45479", + "issue:45704", + "issue:45759" + ], + "soft_pairs": [ + "issue:42915|issue:45759", + "issue:44220|issue:44661", + "issue:45183|issue:45479", + "issue:43618|issue:43697", + "issue:43540|issue:44008", + "issue:43720|issue:43819", + "issue:44077|issue:44625", + "issue:30064|issue:36010", + "issue:43575|issue:45704", + "issue:44756|issue:45759", + "issue:43716|issue:44008", + "issue:43335|issue:43756", + "issue:43653|issue:43720", + "issue:42915|issue:43746", + "issue:43746|issue:45381", + "issue:43746|issue:43873", + "issue:44843|issue:45081", + "issue:43550|issue:45072", + "issue:44186|issue:44898", + "issue:43550|issue:43819", + "issue:43421|issue:44811", + "issue:43577|issue:43818", + "issue:43643|issue:44877", + "issue:44062|issue:44488", + "issue:43531|issue:43643", + "issue:43540|issue:45381", + "issue:44448|issue:44671", + "issue:43653|issue:43819", + "issue:43782|issue:44977", + "issue:44561|issue:45020" + ], + "prepared_review_unit_hash": "d816910637bf2f00fcef59a0745aa91171c2dedc220b02ad3503082186406b1e", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 30, + "serialized_chars": 36372, + "estimated_input_tokens": 9093, + "estimated_eval_tokens": 18442 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The set is a mix of unrelated model-, tokenizer-, config-, and runtime-specific bugs. None of the soft-pair candidates look like true duplicates or the same fixable code-path problem, and there are no PRs in the input.", + "confidence": 0.86, + "canonical_issue_reason": "issue:43873 is the broadest, most active open report here, with multiple inbound references and discussion around a runtime behavior regression.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43873 is the strongest triage anchor because it is active, broadly scoped, and already attracting follow-up references.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:42915", + "right": "issue:45759", + "accept": false, + "reason": "Different problems: Qwen3MoE FP8 config failure vs. AutoModelForCausalLM config unwrapping for composite Qwen multimodal configs." + }, + { + "left": "issue:44220", + "right": "issue:44661", + "accept": false, + "reason": "Different code paths: audio feature extraction failure vs. tokenizer-model mapping logic in add-new-model-like." + }, + { + "left": "issue:45183", + "right": "issue:45479", + "accept": false, + "reason": "One is input validation/error messaging, the other is a classification loss bug with num_labels=1." + }, + { + "left": "issue:43618", + "right": "issue:43697", + "accept": false, + "reason": "Unrelated regressions in different models: CLIP output attentions vs. RTDetrV2 output differences in v5." + }, + { + "left": "issue:43540", + "right": "issue:44008", + "accept": false, + "reason": "Different multimodal failures in different models; no shared concrete bug or fix path." + }, + { + "left": "issue:43720", + "right": "issue:43819", + "accept": false, + "reason": "Different mechanisms: accelerate loading/unpacking packed weights vs. DAC latent conversion/STE mismatch." + }, + { + "left": "issue:44077", + "right": "issue:44625", + "accept": false, + "reason": "One is a model post_init restriction, the other is config propagation for Qwen3.5 labels." + }, + { + "left": "issue:30064", + "right": "issue:36010", + "accept": false, + "reason": "Completely unrelated: image processor segmentation-map handling vs. an import error for GenerationMixin." + }, + { + "left": "issue:43575", + "right": "issue:45704", + "accept": false, + "reason": "Different issues: tensor-parallel OOM on Qwen2-57B vs. T5 using a leaky apex RMSNorm." + }, + { + "left": "issue:44756", + "right": "issue:45759", + "accept": false, + "reason": "Different scope: mmap/OOM on Strix Halo vs. config unwrapping for Qwen multimodal models." + }, + { + "left": "issue:43716", + "right": "issue:44008", + "accept": false, + "reason": "Different models and failures: Mistral-3 dtype mismatch vs. Gemma 3n variable name collision." + }, + { + "left": "issue:43335", + "right": "issue:43756", + "accept": false, + "reason": "Unrelated model architecture bugs: sparse layer creation vs. RoPE layer retention." + }, + { + "left": "issue:43653", + "right": "issue:43720", + "accept": false, + "reason": "Tokenizer special-token registration bug vs. BitNet packed-weight loading bug." + }, + { + "left": "issue:42915", + "right": "issue:43746", + "accept": false, + "reason": "Different models and failure modes: Qwen3MoE FP8 config vs. GraniteSpeech PEFT checkpoint loading." + }, + { + "left": "issue:43746", + "right": "issue:45381", + "accept": false, + "reason": "Checkpoint loading with PEFT adapters is unrelated to Qwen2.5-VL video vision_position_ids." + }, + { + "left": "issue:43746", + "right": "issue:43873", + "accept": false, + "reason": "Unrelated: GraniteSpeech checkpoint loading vs. quantization/offloading behavior." + }, + { + "left": "issue:44843", + "right": "issue:45081", + "accept": false, + "reason": "Both touch _patch_mistral_regex, but they are different bugs: offline model_info call vs. backend_tokenizer AttributeError during regex fixing." + }, + { + "left": "issue:43550", + "right": "issue:45072", + "accept": false, + "reason": "Different subsystems and symptoms: torch.compile+SDPA failure vs. dtype mismatches in bfloat16 inference." + }, + { + "left": "issue:44186", + "right": "issue:44898", + "accept": false, + "reason": "Distinct preprocessing/tokenization bug vs. image classification failure at non-default resolution." + }, + { + "left": "issue:43550", + "right": "issue:43819", + "accept": false, + "reason": "Different model/runtime issues: compile/SDPA failure vs. DAC latent/forward mismatch." + }, + { + "left": "issue:43421", + "right": "issue:44811", + "accept": false, + "reason": "Post-processor updating for special tokens is not the same as Whisper batch_decode skip_special_tokens handling." + }, + { + "left": "issue:43577", + "right": "issue:43818", + "accept": false, + "reason": "Different models and bugs: BLIP2 dtype preservation vs. Video-LLaVA missing temporal attention/weight sharing." + }, + { + "left": "issue:43643", + "right": "issue:44877", + "accept": false, + "reason": "Both involve config loading, but one is missing fields from trust_remote_code AutoConfig and the other is strict granite_speech config rejection." + }, + { + "left": "issue:44062", + "right": "issue:44488", + "accept": false, + "reason": "Different tokenizer/loading failures: AddedToken keyword collision vs. a model load regression for cjvt/sleng-bert." + }, + { + "left": "issue:43531", + "right": "issue:43643", + "accept": false, + "reason": "Qwen3-MoE sliding_window behavior is unrelated to AutoConfig trust_remote_code field loss." + }, + { + "left": "issue:44448", + "right": "issue:44671", + "accept": false, + "reason": "Both are v5 regressions, but they affect different models and concrete outputs; not the same underlying bug." + }, + { + "left": "issue:43653", + "right": "issue:43819", + "accept": false, + "reason": "Tokenizer special-token registration and DAC STE mismatch are unrelated." + }, + { + "left": "issue:43782", + "right": "issue:44977", + "accept": false, + "reason": "Different Qwen variants and failure modes: from_pretrained weight_only error vs. flash-attention generation bug." + }, + { + "left": "issue:44561", + "right": "issue:45020", + "accept": false, + "reason": "44561 is specifically the removal of is_torch_fx_available breaking remote_code models; 45020 is a broader report about recent versions breaking remote_code models, not the same concrete bug." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 14, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:30064", + "issue:34634", + "issue:36010", + "issue:36331", + "issue:42175", + "issue:43262", + "issue:43322", + "issue:43329", + "issue:43404", + "issue:43452", + "issue:43526", + "issue:43577", + "issue:43582", + "issue:43606", + "issue:43644", + "issue:43650", + "issue:43653", + "issue:43697", + "issue:43720", + "issue:43749", + "issue:43782", + "issue:43819", + "issue:43825", + "issue:43827", + "issue:43906", + "issue:43994", + "issue:44060", + "issue:44186", + "issue:44368", + "issue:44442", + "issue:44485", + "issue:44509", + "issue:44610", + "issue:44671", + "issue:44805", + "issue:44821", + "issue:44857", + "issue:44912", + "issue:44936", + "issue:44991", + "issue:45072", + "issue:45081", + "issue:45084", + "issue:45406", + "issue:45538", + "issue:45600", + "issue:45685", + "issue:45701" + ], + "soft_pairs": [ + "issue:43825|issue:43994", + "issue:43322|issue:45685", + "issue:44821|issue:45084", + "issue:43825|issue:43827", + "issue:30064|issue:36331", + "issue:44485|issue:44509", + "issue:43577|issue:43582", + "issue:44186|issue:44442", + "issue:43606|issue:45072", + "issue:43452|issue:44991", + "issue:34634|issue:36010", + "issue:44805|issue:44936", + "issue:42175|issue:45081", + "issue:44610|issue:45072", + "issue:45081|issue:45406", + "issue:44060|issue:44368", + "issue:43650|issue:43906", + "issue:43720|issue:45072", + "issue:44671|issue:45701", + "issue:43404|issue:43782", + "issue:43582|issue:44857", + "issue:43262|issue:43819", + "issue:43526|issue:43697", + "issue:45084|issue:45538", + "issue:43329|issue:45600", + "issue:43644|issue:43749", + "issue:43653|issue:44912" + ], + "prepared_review_unit_hash": "fbff41e56a679b16bda0d3690fce5af578a8dee00dba3fdef0be3728830f6735", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 27, + "serialized_chars": 35989, + "estimated_input_tokens": 8998, + "estimated_eval_tokens": 18252 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is heterogeneous: the items span unrelated bugs in processors, tokenizers, training, loading, and docs. I found no convincing duplicate issue/PR pairings.", + "confidence": 0.9, + "canonical_issue_reason": "No single issue is a good canonical representative because the cluster does not center on one underlying bug; the titles point to many distinct code paths and models.", + "canonical_pr_reason": null, + "best_issue_reason": "No issue is a strong cluster anchor here. The most discussed items are still about different failures, so using any one as canonical would be misleading.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43825", + "right": "issue:43994", + "accept": false, + "reason": "Both mention pipeline behavior, but one is about a wrong v5 error message and the other about SigLIP2 producing bad outputs; different bugs." + }, + { + "left": "issue:43322", + "right": "issue:45685", + "accept": false, + "reason": "Different subsystems and failure modes: Llava Next loading segfault vs an MPS histogram/int dtype issue." + }, + { + "left": "issue:44821", + "right": "issue:45084", + "accept": false, + "reason": "Unrelated: AutoImageProcessor URL loading vs a template compilation TypeError." + }, + { + "left": "issue:43825", + "right": "issue:43827", + "accept": false, + "reason": "Closely related theme (v5 pipeline removals), but one is runtime error messaging and the other is docs cleanup; not the same concrete bug." + }, + { + "left": "issue:30064", + "right": "issue:36331", + "accept": false, + "reason": "Different areas: image segmentation-map processing vs CustomTrainer compute_loss signature mismatch." + }, + { + "left": "issue:44485", + "right": "issue:44509", + "accept": false, + "reason": "GLM-5 RoPE implementation discussion is unrelated to pipeline docs references." + }, + { + "left": "issue:43577", + "right": "issue:43582", + "accept": false, + "reason": "Blip2 dtype loading bug vs Apple Silicon allocator warmup TypeError are unrelated." + }, + { + "left": "issue:44186", + "right": "issue:44442", + "accept": false, + "reason": "Both are tokenizer bugs, but they affect different tokenizers with different behaviors and fixes." + }, + { + "left": "issue:43606", + "right": "issue:45072", + "accept": false, + "reason": "Both involve device/dtype issues, but they are separate models and separate code paths." + }, + { + "left": "issue:43452", + "right": "issue:44991", + "accept": false, + "reason": "Tokenizer/model loading regressions, but not the same concrete failure; one is gguf_file handling, the other is a specific tokenizer load breakage." + }, + { + "left": "issue:34634", + "right": "issue:36010", + "accept": false, + "reason": "BarkProcessor voice_preset bug is unrelated to the GenerationMixin import error." + }, + { + "left": "issue:44805", + "right": "issue:44936", + "accept": false, + "reason": "Mask shape IndexError and trainer.evaluate() after train() are different problems." + }, + { + "left": "issue:42175", + "right": "issue:45081", + "accept": false, + "reason": "Package/backend install issue vs Mistral regex patch AttributeError; no shared underlying bug." + }, + { + "left": "issue:44610", + "right": "issue:45072", + "accept": false, + "reason": "OmDet-Turbo input-size mismatch is unrelated to dtype mismatch inference issues." + }, + { + "left": "issue:45081", + "right": "issue:45406", + "accept": false, + "reason": "Both are AttributeErrors in tokenizer/processor code, but on different attributes and different components." + }, + { + "left": "issue:44060", + "right": "issue:44368", + "accept": false, + "reason": "Related to Qwen tied-weights warnings, but the reported failures are not the same concrete bug or fix." + }, + { + "left": "issue:43650", + "right": "issue:43906", + "accept": false, + "reason": "Placeholder issue title vs an isolated reproduction of a different issue; no substantive overlap." + }, + { + "left": "issue:43720", + "right": "issue:45072", + "accept": false, + "reason": "Packed-weight loading bug vs bfloat16 dtype mismatch; unrelated." + }, + { + "left": "issue:44671", + "right": "issue:45701", + "accept": false, + "reason": "CamemBERT MLM regression in v5 is not the same as the broad statement that version changes tokenization." + }, + { + "left": "issue:43404", + "right": "issue:43782", + "accept": false, + "reason": "Different model families and different loading/tied-weight failures." + }, + { + "left": "issue:43582", + "right": "issue:44857", + "accept": false, + "reason": "Apple Silicon allocator warmup TypeError is unrelated to a CUDA AMP loss crash." + }, + { + "left": "issue:43262", + "right": "issue:43819", + "accept": false, + "reason": "Audio chat-template sampling-rate default bug vs DAC.from_latents mismatch; different functionality." + }, + { + "left": "issue:43526", + "right": "issue:43697", + "accept": false, + "reason": "Different models and different bugs: label reduction in BeitImageProcessorFast vs RTDetrV2 output drift." + }, + { + "left": "issue:45084", + "right": "issue:45538", + "accept": false, + "reason": "Template compilation error is unrelated to CLIPTokenizer model_max_length behavior." + }, + { + "left": "issue:43329", + "right": "issue:45600", + "accept": false, + "reason": "Multimodal token counting bug vs auto_mappings referencing removed configs; no shared underlying defect." + }, + { + "left": "issue:43644", + "right": "issue:43749", + "accept": false, + "reason": "Non-persistent buffer junk in v5 and broken FSDP CPU RAM efficient loading are different loading problems." + }, + { + "left": "issue:43653", + "right": "issue:44912", + "accept": false, + "reason": "BigBird special-token decoding bug is unrelated to MXFP4 quantization fallback." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 15, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:34567", + "issue:34634", + "issue:36010", + "issue:36331", + "issue:38175", + "issue:39692", + "issue:40990", + "issue:41720", + "issue:43257", + "issue:43262", + "issue:43329", + "issue:43450", + "issue:43452", + "issue:43454", + "issue:43526", + "issue:43531", + "issue:43540", + "issue:43550", + "issue:43577", + "issue:43582", + "issue:43638", + "issue:43643", + "issue:43645", + "issue:43697", + "issue:43720", + "issue:43746", + "issue:43749", + "issue:43819", + "issue:43854", + "issue:43927", + "issue:43931", + "issue:44291", + "issue:44479", + "issue:44554", + "issue:44568", + "issue:44589", + "issue:44610", + "issue:44811", + "issue:44849", + "issue:44912", + "issue:44936", + "issue:44977", + "issue:45042", + "issue:45356", + "issue:45357", + "issue:45488", + "issue:45701", + "issue:45759" + ], + "soft_pairs": [ + "issue:43638|issue:44936", + "issue:43746|issue:44811", + "issue:38175|issue:44610", + "issue:44589|issue:45042", + "issue:43454|issue:43577", + "issue:43854|issue:44977", + "issue:43526|issue:44554", + "issue:36010|issue:43645", + "issue:44568|issue:45488", + "issue:34634|issue:36331", + "issue:39692|issue:44912", + "issue:43257|issue:43329", + "issue:43819|issue:43927", + "issue:43720|issue:44610", + "issue:43749|issue:44291", + "issue:43550|issue:44912", + "issue:43262|issue:44811", + "issue:44568|issue:45356", + "issue:43531|issue:45042", + "issue:43643|issue:45759", + "issue:36010|issue:36331", + "issue:34567|issue:40990", + "issue:43329|issue:43450", + "issue:43329|issue:44479", + "issue:44849|issue:45701", + "issue:43257|issue:43931", + "issue:43697|issue:45357", + "issue:43452|issue:43582", + "issue:41720|issue:43540", + "issue:44291|issue:45042" + ], + "prepared_review_unit_hash": "dba4fec7fd2248aaac6fba3f39fc3f599bb0fd097a37cdcb185898679487d297", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 30, + "serialized_chars": 37051, + "estimated_input_tokens": 9263, + "estimated_eval_tokens": 18782 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is a grab bag of mostly unrelated Transformers regressions. The soft pairs share broad themes like loading, tokenization, video, or multimodal processing, but they do not look like the same concrete bug/change.", + "confidence": 0.88, + "canonical_issue_reason": "issue:43749 is the broadest, highest-impact loading/initialization regression in the set and has the strongest discussion/reference signal; the rest are mostly narrower model- or feature-specific bugs.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43749 best represents the cluster because it is a central infrastructure-loading failure rather than a model-specific edge case.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43638", + "right": "issue:44936", + "accept": false, + "reason": "Different trainer failures: Deepspeed ZeRO-3 index error vs trainer.evaluate() after train()." + }, + { + "left": "issue:43746", + "right": "issue:44811", + "accept": false, + "reason": "Different code paths: PEFT/local checkpoint loading vs Whisper batch_decode skip_special_tokens handling." + }, + { + "left": "issue:38175", + "right": "issue:44610", + "accept": false, + "reason": "Different models and bugs: SigLIP2 zero probabilities vs OmDet-Turbo processor size mismatch." + }, + { + "left": "issue:44589", + "right": "issue:45042", + "accept": false, + "reason": "Unrelated failures: Float8 storage lookup vs PIL image processor torchvision dependency regression." + }, + { + "left": "issue:43454", + "right": "issue:43577", + "accept": false, + "reason": "Both are model-loading issues, but one is lm_head tying in AyaVision and the other is Blip2 dtype propagation; not the same bug." + }, + { + "left": "issue:43854", + "right": "issue:44977", + "accept": false, + "reason": "Different issues: GLM-4.7-Flash test loading failure vs Qwen3.5 flash-attention generation behavior." + }, + { + "left": "issue:43526", + "right": "issue:44554", + "accept": false, + "reason": "Different subsystems: BEiT label reduction bug vs MPS attention correctness issue." + }, + { + "left": "issue:36010", + "right": "issue:43645", + "accept": false, + "reason": "Both are API regressions, but one is GenerationMixin import resolution and the other is custom model initialization in notebooks." + }, + { + "left": "issue:44568", + "right": "issue:45488", + "accept": false, + "reason": "Both concern tokenization, but they hit different tokenizer implementations and failure modes." + }, + { + "left": "issue:34634", + "right": "issue:36331", + "accept": false, + "reason": "Different training/audio bugs: BarkProcessor voice_preset vs CustomTrainer.compute_loss signature change." + }, + { + "left": "issue:39692", + "right": "issue:44912", + "accept": false, + "reason": "Unrelated: documentation/example errors for SigLIP2 vs MXFP4 quantization fallback for git-oss-20b." + }, + { + "left": "issue:43257", + "right": "issue:43329", + "accept": false, + "reason": "Different problems: Qwen3 MoE weight conversion/loading vs undefined helpers in multimodal video token counting." + }, + { + "left": "issue:43819", + "right": "issue:43927", + "accept": false, + "reason": "Different model bugs: DAC latent reconstruction mismatch vs DiaConfig losing custom token IDs on save/load." + }, + { + "left": "issue:43720", + "right": "issue:44610", + "accept": false, + "reason": "Different failures: BitNet packed-weight unpacking during accelerate loading vs OmDet-Turbo processor output shape/size mismatch." + }, + { + "left": "issue:43749", + "right": "issue:44291", + "accept": false, + "reason": "Both involve loading infrastructure, but one is FSDP CPU RAM efficient loading and the other is init_empty_weights/_is_hf_initialized signature breakage." + }, + { + "left": "issue:43550", + "right": "issue:44912", + "accept": false, + "reason": "Different model/runtime issues: torch.compile SDPA bug vs MXFP4 quantization fallback." + }, + { + "left": "issue:43262", + "right": "issue:44811", + "accept": false, + "reason": "Different audio processor bugs: chat-template sampling rate default vs Whisper decode special-token handling." + }, + { + "left": "issue:44568", + "right": "issue:45356", + "accept": false, + "reason": "Different tokenizer regressions on different models and code paths; not the same defect." + }, + { + "left": "issue:43531", + "right": "issue:45042", + "accept": false, + "reason": "Unrelated: Qwen3-MoE sliding_window bug vs PIL backend torchvision requirement regression." + }, + { + "left": "issue:43643", + "right": "issue:45759", + "accept": false, + "reason": "Both are config/loading-related, but one is trust_remote_code returning missing fields and the other is AutoModelForCausalLM.from_config not unwrapping text_config." + }, + { + "left": "issue:36010", + "right": "issue:36331", + "accept": false, + "reason": "Different APIs: import path for GenerationMixin vs Trainer.compute_loss keyword-argument compatibility." + }, + { + "left": "issue:34567", + "right": "issue:40990", + "accept": false, + "reason": "Unrelated: TrainerState token counter not updating vs GPT-OSS perplexity regression." + }, + { + "left": "issue:43329", + "right": "issue:43450", + "accept": false, + "reason": "Both are video-related, but one is missing functions in token counting and the other is batched video processor shape handling." + }, + { + "left": "issue:43329", + "right": "issue:44479", + "accept": false, + "reason": "Shared video theme only; one is undefined helpers in a video branch, the other is a broader v5.3.0 video input regression." + }, + { + "left": "issue:44849", + "right": "issue:45701", + "accept": false, + "reason": "Different regressions: Qwen3.5 hidden-state output bug vs version-dependent tokenization changes." + }, + { + "left": "issue:43257", + "right": "issue:43931", + "accept": false, + "reason": "Different Qwen-related loading bugs: MoE weight conversion vs VL weight shape mismatch." + }, + { + "left": "issue:43697", + "right": "issue:45357", + "accept": false, + "reason": "Different issues: RTDetrV2 output drift in v5 vs save_pretrained storing incorrect visual encoder keys." + }, + { + "left": "issue:43452", + "right": "issue:43582", + "accept": false, + "reason": "Unrelated: gguf_file/from_pretrained loading breakage vs Apple Silicon caching_allocator_warmup TypeError." + }, + { + "left": "issue:41720", + "right": "issue:43540", + "accept": false, + "reason": "Different Qwen3-family failures: auto device mapping cuda assert vs video input ValueError in Qwen3OmniMoe." + }, + { + "left": "issue:44291", + "right": "issue:45042", + "accept": false, + "reason": "Different bugs in different layers: init_empty_weights argument handling vs image processor torchvision dependency." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 16, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:33357", + "issue:34567", + "issue:36010", + "issue:36331", + "issue:38175", + "issue:42175", + "issue:42915", + "issue:43257", + "issue:43317", + "issue:43452", + "issue:43493", + "issue:43526", + "issue:43540", + "issue:43577", + "issue:43643", + "issue:43646", + "issue:43696", + "issue:43742", + "issue:43749", + "issue:43756", + "issue:43824", + "issue:43906", + "issue:43927", + "issue:43994", + "issue:44062", + "issue:44291", + "issue:44479", + "issue:44488", + "issue:44560", + "issue:44737", + "issue:44756", + "issue:44792", + "issue:44821", + "issue:44898", + "issue:44912", + "issue:44964", + "issue:44987", + "issue:44991", + "issue:45072", + "issue:45081", + "issue:45137", + "issue:45375", + "issue:45381", + "issue:45405", + "issue:45479", + "issue:45538", + "issue:45588" + ], + "soft_pairs": [ + "issue:42915|issue:43742", + "issue:33357|issue:34567", + "issue:43317|issue:45538", + "issue:44987|issue:45538", + "issue:44479|issue:45381", + "issue:43540|issue:44560", + "issue:43927|issue:44912", + "issue:42915|issue:44821", + "issue:43257|issue:43756", + "issue:43577|issue:45072", + "issue:43540|issue:44479", + "issue:42915|issue:45538", + "issue:43257|issue:43696", + "issue:45137|issue:45405", + "issue:38175|issue:44898", + "issue:44737|issue:45588", + "issue:44062|issue:44991", + "issue:43742|issue:44291", + "issue:43906|issue:45381", + "issue:43646|issue:43749", + "issue:43749|issue:44756", + "issue:44062|issue:45081", + "issue:36010|issue:42175", + "issue:34567|issue:36331", + "issue:43643|issue:45375", + "issue:43493|issue:43994", + "issue:42175|issue:43824", + "issue:43452|issue:44488", + "issue:43526|issue:45479", + "issue:44291|issue:44821", + "issue:44792|issue:44964" + ], + "prepared_review_unit_hash": "10475a892ee88241aff13a2d3e0b91d5fea5da0d52da5e458f69c5482d49f8e9", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 31, + "serialized_chars": 36567, + "estimated_input_tokens": 9142, + "estimated_eval_tokens": 18540 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly heterogeneous: most items are distinct bug reports across unrelated models/subsystems. The only strong duplicate-like matches are the Qwen video-input regression pair and the SigLIP2 correctness pair.", + "confidence": 0.66, + "canonical_issue_reason": "issue:44479 is the broadest, most central issue in the set around the shared Qwen video-input regression, and it has the strongest umbrella framing for the accepted duplicate-like report.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:44479 is the best single representative of the cluster\u2019s overlapping video-regression theme; it is broader than the narrower follow-up report and is more suitable as the canonical issue.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:42915", + "right": "issue:43742", + "accept": false, + "reason": "Different models and failure modes: Qwen3Moe FP8 config vs MobileLLM key error." + }, + { + "left": "issue:33357", + "right": "issue:34567", + "accept": false, + "reason": "Unrelated areas: MacOS bus error with CLIP vs TrainerState token-count tracking." + }, + { + "left": "issue:43317", + "right": "issue:45538", + "accept": false, + "reason": "Different problems: offload/loading failure vs CLIPTokenizer max-length constant." + }, + { + "left": "issue:44987", + "right": "issue:45538", + "accept": false, + "reason": "Different loading bugs: model load regression vs tokenizer max-length issue." + }, + { + "left": "issue:44479", + "right": "issue:45381", + "accept": true, + "reason": "Both describe the same Qwen video-input regression; 45381 looks like the narrower symptom/diagnosis of the broader v5.3.0 video-input bug." + }, + { + "left": "issue:43540", + "right": "issue:44560", + "accept": false, + "reason": "Both involve video inputs, but they are different Qwen model variants with different errors; not enough to treat as the same bug." + }, + { + "left": "issue:43927", + "right": "issue:44912", + "accept": false, + "reason": "Different root causes: config token IDs being lost vs MXFP4 quantization fallback/loading behavior." + }, + { + "left": "issue:42915", + "right": "issue:44821", + "accept": false, + "reason": "Different targets and failures: Qwen3Moe FP8 config vs AutoImageProcessor URL loading." + }, + { + "left": "issue:43257", + "right": "issue:43756", + "accept": false, + "reason": "Distinct model-specific issues: MOE weight conversion under accelerate/deepspeed vs Smollm3 RoPE layer mismatch." + }, + { + "left": "issue:43577", + "right": "issue:45072", + "accept": false, + "reason": "Both mention dtype, but they affect different models and code paths; too broad to merge." + }, + { + "left": "issue:43540", + "right": "issue:44479", + "accept": false, + "reason": "Same broad topic of video inputs, but the model families and symptoms differ; not clearly the same concrete bug." + }, + { + "left": "issue:42915", + "right": "issue:45538", + "accept": false, + "reason": "Completely different scopes: FP8/Qwen3Moe loading vs CLIP tokenizer max length." + }, + { + "left": "issue:43257", + "right": "issue:43696", + "accept": false, + "reason": "Different underlying issue classes: MOE conversion/loading vs CUDA OOM." + }, + { + "left": "issue:45137", + "right": "issue:45405", + "accept": false, + "reason": "DeepSpeed ZeRO3 deque failure is unrelated to PEFT version bump." + }, + { + "left": "issue:38175", + "right": "issue:44898", + "accept": false, + "reason": "Both are image-model correctness bugs, but they concern different architectures and failure modes." + }, + { + "left": "issue:44737", + "right": "issue:45588", + "accept": false, + "reason": "XLNet CPU-device bug is unrelated to flash-attention sink-less model crash." + }, + { + "left": "issue:44062", + "right": "issue:44991", + "accept": false, + "reason": "Tokenizer-related, but different models and insufficient evidence of the same special-token bug." + }, + { + "left": "issue:43742", + "right": "issue:44291", + "accept": false, + "reason": "Both are model-loading errors, but the specific causes are different (_is_hf_initialized vs MobileLLM key error)." + }, + { + "left": "issue:43906", + "right": "issue:45381", + "accept": false, + "reason": "Isolated reproduction of a different issue; not the same as the Qwen2.5-VL video regression." + }, + { + "left": "issue:43646", + "right": "issue:43749", + "accept": false, + "reason": "Custom model init breakage is unrelated to FSDP CPU RAM-efficient loading." + }, + { + "left": "issue:43749", + "right": "issue:44756", + "accept": false, + "reason": "Different memory/loading failures with different triggers; not the same bug." + }, + { + "left": "issue:44062", + "right": "issue:45081", + "accept": false, + "reason": "Tokenizer special-arg error vs Mistral regex/backend_tokenizer AttributeError are different failure points." + }, + { + "left": "issue:36010", + "right": "issue:42175", + "accept": false, + "reason": "Import path break for GenerationMixin is unrelated to TensorFlow being absent from a torch extra install." + }, + { + "left": "issue:34567", + "right": "issue:36331", + "accept": false, + "reason": "TrainerState token accounting vs custom compute_loss signature mismatch are separate training issues." + }, + { + "left": "issue:43643", + "right": "issue:45375", + "accept": false, + "reason": "Both concern config fields, but one is trust_remote_code config materialization and the other is a specific Qwen vision config field being dropped." + }, + { + "left": "issue:43493", + "right": "issue:43994", + "accept": true, + "reason": "Both report the same SigLIP2 model discrepancy/correctness problem: HF outputs disagree with expected behavior and produce nonsensical results." + }, + { + "left": "issue:42175", + "right": "issue:43824", + "accept": false, + "reason": "Package/dependency availability issue vs missing model class import; unrelated." + }, + { + "left": "issue:43452", + "right": "issue:44488", + "accept": false, + "reason": "Different loading breakages for different models and APIs." + }, + { + "left": "issue:43526", + "right": "issue:45479", + "accept": false, + "reason": "Different image/label handling issues: BeitImageProcessorFast reduction bug vs sequence-classification loss degeneracy." + }, + { + "left": "issue:44291", + "right": "issue:44821", + "accept": false, + "reason": "General init_empty_weights loading bug vs AutoImageProcessor-from-URL failure are different code paths." + }, + { + "left": "issue:44792", + "right": "issue:44964", + "accept": false, + "reason": "Janus generation test failure and Phi-4 multimodal loading failure are unrelated." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet and stays conservative. The two accepted duplicate-like pairs are reasonably supported by the issue titles/descriptions, and the remaining soft-edge rejections are appropriately cautious. No overstatement stands out." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 17, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:24643", + "issue:28282", + "issue:29127", + "issue:34689", + "issue:38175", + "issue:41762", + "issue:42175", + "issue:43295", + "issue:43452", + "issue:43531", + "issue:43577", + "issue:43582", + "issue:43643", + "issue:43653", + "issue:43697", + "issue:43720", + "issue:43742", + "issue:43756", + "issue:43819", + "issue:44361", + "issue:44442", + "issue:44466", + "issue:44484", + "issue:44514", + "issue:44561", + "issue:44568", + "issue:44589", + "issue:44625", + "issue:44704", + "issue:44811", + "issue:44821", + "issue:44843", + "issue:44877", + "issue:44898", + "issue:44936", + "issue:44938", + "issue:44977", + "issue:45005", + "issue:45042", + "issue:45081", + "issue:45216", + "issue:45290", + "issue:45325", + "issue:45356", + "issue:45381", + "issue:45405", + "issue:45600" + ], + "soft_pairs": [ + "issue:28282|issue:34689", + "issue:45042|issue:45081", + "issue:44484|issue:44704", + "issue:45405|issue:45600", + "issue:43756|issue:45325", + "issue:43742|issue:44589", + "issue:43697|issue:45216", + "issue:28282|issue:29127", + "issue:44514|issue:45381", + "issue:43531|issue:44821", + "issue:43643|issue:44561", + "issue:43582|issue:44977", + "issue:44625|issue:44877", + "issue:43452|issue:43577", + "issue:43720|issue:44898", + "issue:44514|issue:45290", + "issue:43452|issue:44843", + "issue:42175|issue:44938", + "issue:44466|issue:45005", + "issue:43819|issue:44811", + "issue:43295|issue:45356", + "issue:41762|issue:44936", + "issue:43653|issue:44568", + "issue:24643|issue:29127", + "issue:38175|issue:43295", + "issue:44361|issue:44442" + ], + "prepared_review_unit_hash": "4bca7397a15a2ff095c2d0091c6669a6d94828502a7a1e11cf3547d8af6d16cf", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 26, + "serialized_chars": 35227, + "estimated_input_tokens": 8807, + "estimated_eval_tokens": 17870 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly heterogeneous and does not look like a single duplicate set. The only convincing duplicate pair is the tied-weights save_pretrained regression; the rest are unrelated tokenizer, config, model-loading, and vision bugs.", + "confidence": 0.88, + "canonical_issue_reason": "Issue 44466 is the most concrete anchor for the tied-weights regression: it describes the symptom, version boundary, and device-dependent serialization behavior more specifically than 45005.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 44466 is the strongest standalone issue in the cluster because it has the clearest repro target and the most actionable description of the underlying save_pretrained/tied-weights problem.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:28282", + "right": "issue:34689", + "accept": false, + "reason": "Different failures: missing PyTorch import vs. model loading regression for a specific Llama vision model." + }, + { + "left": "issue:45042", + "right": "issue:45081", + "accept": false, + "reason": "Unrelated subsystems: image processor torchvision dependency vs. Mistral tokenizer regex crash." + }, + { + "left": "issue:44484", + "right": "issue:44704", + "accept": false, + "reason": "One is a save_pretrained shard-size question; the other is a cached_file kwargs propagation bug." + }, + { + "left": "issue:45405", + "right": "issue:45600", + "accept": false, + "reason": "Release/versioning problem vs. broken auto_mappings references; not the same bug." + }, + { + "left": "issue:43756", + "right": "issue:45325", + "accept": false, + "reason": "Both involve positional/RoPE logic, but they affect different models and different code paths." + }, + { + "left": "issue:43742", + "right": "issue:44589", + "accept": false, + "reason": "Different loading/runtime failures: missing key for MobileLLM vs. Float8 storage lookup error." + }, + { + "left": "issue:43697", + "right": "issue:45216", + "accept": false, + "reason": "Different regression surfaces: output drift in RTDetrV2 vs. incorrect checkpoint serialization for Qwen3.5." + }, + { + "left": "issue:28282", + "right": "issue:29127", + "accept": false, + "reason": "ImportError for PyTorch is unrelated to the LayoutLMv3 box-validation error message issue." + }, + { + "left": "issue:44514", + "right": "issue:45381", + "accept": false, + "reason": "Both are Qwen2.5-VL chat/vision bugs, but one is batched padding handling and the other is wrong video position ids." + }, + { + "left": "issue:43531", + "right": "issue:44821", + "accept": false, + "reason": "Sliding-window behavior for Qwen3-MoE is unrelated to loading an AutoImageProcessor from a URL." + }, + { + "left": "issue:43643", + "right": "issue:44561", + "accept": false, + "reason": "Both involve trust_remote_code, but one is missing fields in AutoConfig and the other is a removed helper breaking imports." + }, + { + "left": "issue:43582", + "right": "issue:44977", + "accept": false, + "reason": "AppleSilicon allocator TypeError is unrelated to Qwen3.5 flash-attention generation issues." + }, + { + "left": "issue:44625", + "right": "issue:44877", + "accept": false, + "reason": "Config propagation of num_labels is a different problem from strict config blocking granite_speech loading." + }, + { + "left": "issue:43452", + "right": "issue:43577", + "accept": false, + "reason": "gguf_file loading breakage and Blip2 dtype propagation are different model-loading bugs." + }, + { + "left": "issue:43720", + "right": "issue:44898", + "accept": false, + "reason": "Packed-weight unpacking during accelerate loading is unrelated to Perceiver interpolation failure." + }, + { + "left": "issue:44514", + "right": "issue:45290", + "accept": false, + "reason": "Both hit apply_chat_template, but the triggers and failure modes differ enough that they are not the same bug." + }, + { + "left": "issue:43452", + "right": "issue:44843", + "accept": false, + "reason": "Different failure points: gguf_file loading vs. unconditional model_info() call breaking offline mode." + }, + { + "left": "issue:42175", + "right": "issue:44938", + "accept": false, + "reason": "Packaging/backend dependency issue is unrelated to a Python 3.14 load failure." + }, + { + "left": "issue:44466", + "right": "issue:45005", + "accept": true, + "reason": "Both describe the same tied-weights/save_pretrained regression in v5, with 44466 giving the clearest device-dependent serialization symptom." + }, + { + "left": "issue:43819", + "right": "issue:44811", + "accept": false, + "reason": "DAC latent conversion mismatch and Whisper batch_decode skip_special_tokens behavior are unrelated." + }, + { + "left": "issue:43295", + "right": "issue:45356", + "accept": false, + "reason": "Different tokenizer regressions with different symptoms and likely different code paths." + }, + { + "left": "issue:41762", + "right": "issue:44936", + "accept": false, + "reason": "DeepSpeed ZeRO-3 loading IndexError is unrelated to trainer.evaluate() failing after trainer.train()." + }, + { + "left": "issue:43653", + "right": "issue:44568", + "accept": false, + "reason": "Different tokenizer/special-token bugs affecting different models and behaviors." + }, + { + "left": "issue:24643", + "right": "issue:29127", + "accept": false, + "reason": "DeepSpeed weight-dimension RuntimeError is unrelated to LayoutLMv3 box-information validation." + }, + { + "left": "issue:38175", + "right": "issue:43295", + "accept": false, + "reason": "Zero-probabilities output issue is unrelated to processor/tokenizer regression." + }, + { + "left": "issue:44361", + "right": "issue:44442", + "accept": false, + "reason": "Different tokenizer load failures for different model families." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The cluster summary is conservative and grounded in the packet: the tied-weights save_pretrained pair (44466/45005) is the only clearly mergeable duplicate, and the other pairwise rejections stay within the evidence suggested by the titles. No soft-edge verdict appears to overstate sameness." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 18, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:28282", + "issue:30064", + "issue:32090", + "issue:33666", + "issue:34567", + "issue:35141", + "issue:36010", + "issue:41720", + "issue:43122", + "issue:43262", + "issue:43329", + "issue:43452", + "issue:43493", + "issue:43504", + "issue:43525", + "issue:43531", + "issue:43550", + "issue:43572", + "issue:43577", + "issue:43653", + "issue:43688", + "issue:43697", + "issue:43720", + "issue:43742", + "issue:43749", + "issue:43756", + "issue:43825", + "issue:43906", + "issue:43927", + "issue:44361", + "issue:44451", + "issue:44464", + "issue:44479", + "issue:44610", + "issue:44743", + "issue:44821", + "issue:44843", + "issue:44849", + "issue:44913", + "issue:44936", + "issue:44977", + "issue:44991", + "issue:45084", + "issue:45325", + "issue:45356", + "issue:45357", + "issue:45440", + "issue:45588" + ], + "soft_pairs": [ + "issue:30064|issue:33666", + "issue:43653|issue:44361", + "issue:44464|issue:44743", + "issue:43577|issue:43697", + "issue:44843|issue:44991", + "issue:41720|issue:43906", + "issue:44479|issue:44936", + "issue:44977|issue:45356", + "issue:43688|issue:45440", + "issue:43577|issue:45588", + "issue:43531|issue:43742", + "issue:43653|issue:43927", + "issue:44464|issue:44977", + "issue:43550|issue:43927", + "issue:43550|issue:44361", + "issue:43122|issue:43906", + "issue:43504|issue:43720", + "issue:32090|issue:34567", + "issue:43504|issue:43550", + "issue:32090|issue:33666", + "issue:28282|issue:35141", + "issue:43504|issue:44610", + "issue:43452|issue:44451", + "issue:43756|issue:44913", + "issue:43329|issue:45325", + "issue:43742|issue:44821", + "issue:43688|issue:43749", + "issue:43525|issue:43572", + "issue:41720|issue:44464", + "issue:33666|issue:36010", + "issue:43493|issue:43825", + "issue:43262|issue:45084", + "issue:44849|issue:45357" + ], + "prepared_review_unit_hash": "40735fe0efd11e59b73ef79c443cee48b614b9b4bb563b0e373d2f06ab3d48d7", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 33, + "serialized_chars": 37710, + "estimated_input_tokens": 9428, + "estimated_eval_tokens": 19112 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is heterogeneous: it mixes unrelated tokenizer, trainer, vision, config, and generation regressions. None of the soft pairs look like the same underlying bug, so no dedup merges are justified.", + "confidence": 0.96, + "canonical_issue_reason": "No single canonical issue stands out; the items span multiple unrelated subsystems and model families, and the candidate pairs are not close enough to be duplicates.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 44843 is the clearest self-contained regression report, but it is still too specific to serve as a true canonical for this mixed cluster.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:30064", + "right": "issue:33666", + "accept": false, + "reason": "Image processor segmentation-map failure vs Qwen2-VL multi-GPU training; different models and failure modes." + }, + { + "left": "issue:43653", + "right": "issue:44361", + "accept": false, + "reason": "BigBirdTokenizer special-token decode bug vs MLukeTokenizer task AttributeError; not the same code path." + }, + { + "left": "issue:44464", + "right": "issue:44743", + "accept": false, + "reason": "Both affect generation, but one is compiled-forward inconsistency and the other is recurrent-state reset with cache; distinct bugs." + }, + { + "left": "issue:43577", + "right": "issue:43697", + "accept": false, + "reason": "Blip2 dtype-loading issue vs RTDetrV2 output regression; unrelated subsystems and symptoms." + }, + { + "left": "issue:44843", + "right": "issue:44991", + "accept": false, + "reason": "Both are tokenizer-load regressions, but one is an offline model_info call in Mistral regex patching and the other is a specific tokenizer load failure for a different model." + }, + { + "left": "issue:41720", + "right": "issue:43906", + "accept": false, + "reason": "Insufficient evidence of the same defect; one is an A800 device-map cuda assert, the other is only a reproduction of another issue." + }, + { + "left": "issue:44479", + "right": "issue:44936", + "accept": false, + "reason": "Video-input regression for several Qwen VL models vs trainer evaluate/train lifecycle failure; unrelated." + }, + { + "left": "issue:44977", + "right": "issue:45356", + "accept": false, + "reason": "Flash-attention generation bug vs tokenizer codec/regex regression; different layers of the stack." + }, + { + "left": "issue:43688", + "right": "issue:45440", + "accept": false, + "reason": "Both involve MoE models, but one is auxiliary-loss normalization and the other is native-vs-remote implementation divergence; not mergeable into one fix." + }, + { + "left": "issue:43577", + "right": "issue:45588", + "accept": false, + "reason": "Dtype propagation bug vs flash_attention crash on s_aux=None; different failure mechanisms." + }, + { + "left": "issue:43531", + "right": "issue:43742", + "accept": false, + "reason": "Qwen3-MoE sliding_window issue vs MobileLLM load KeyError; unrelated." + }, + { + "left": "issue:43653", + "right": "issue:43927", + "accept": false, + "reason": "Both are token-related, but one is mask-token registration and the other is saved custom token IDs causing generation IndexError; distinct bugs." + }, + { + "left": "issue:44464", + "right": "issue:44977", + "accept": false, + "reason": "Compiled-forward inconsistency vs flash-attention generation failure; different concrete problems." + }, + { + "left": "issue:43550", + "right": "issue:43927", + "accept": false, + "reason": "torch.compile/SDPA failure vs config save/load token-ID loss; unrelated." + }, + { + "left": "issue:43550", + "right": "issue:44361", + "accept": false, + "reason": "Bamba compile+SDPA bug vs MLukeTokenizer AttributeError; no shared underlying issue." + }, + { + "left": "issue:43122", + "right": "issue:43906", + "accept": false, + "reason": "Different tokenization regression vs isolated reproduction of another issue; no evidence they are the same bug." + }, + { + "left": "issue:43504", + "right": "issue:43720", + "accept": false, + "reason": "BEiT legacy-field load bug vs BitNet packed-weight unpacking bug; different loading paths." + }, + { + "left": "issue:32090", + "right": "issue:34567", + "accept": false, + "reason": "Trainer broadcast NoneType error vs token-count bookkeeping not updating; both trainer-related but not the same defect." + }, + { + "left": "issue:43504", + "right": "issue:43550", + "accept": false, + "reason": "Legacy-field pretrained-model load bug vs Bamba torch.compile+SDPA failure; unrelated." + }, + { + "left": "issue:32090", + "right": "issue:33666", + "accept": false, + "reason": "Trainer _gpu_broadcast_one TypeError vs Qwen2-VL multi-GPU training; unrelated." + }, + { + "left": "issue:28282", + "right": "issue:35141", + "accept": false, + "reason": "Missing PyTorch import vs embedding resize reinitialization after post_init; different concerns entirely." + }, + { + "left": "issue:43504", + "right": "issue:44610", + "accept": false, + "reason": "BEiT preset legacy-field bug vs OmDet-Turbo processor size mismatch; different models and code paths." + }, + { + "left": "issue:43452", + "right": "issue:44451", + "accept": false, + "reason": "gguf_file loading breakage vs a specific model load failure; both are load issues but not the same one." + }, + { + "left": "issue:43756", + "right": "issue:44913", + "accept": false, + "reason": "RoPE layer-dropping bug vs rotary_pct reload bug; both config-related but clearly different regressions." + }, + { + "left": "issue:43329", + "right": "issue:45325", + "accept": false, + "reason": "Undefined video-branch helpers in multimodal token counting vs still-image temporal position scaling in Qwen2.5-VL; distinct code paths." + }, + { + "left": "issue:43742", + "right": "issue:44821", + "accept": false, + "reason": "Model KeyError on load vs AutoImageProcessor URL-loading failure; unrelated." + }, + { + "left": "issue:43688", + "right": "issue:43749", + "accept": false, + "reason": "Aux-loss normalization bug vs FSDP CPU RAM efficient loading breakage; different distributed-training issues." + }, + { + "left": "issue:43525", + "right": "issue:43572", + "accept": false, + "reason": "Missing pad_token_id in Llama4Config vs missing pad_token_idx in StableLmConfig; similar theme, but different model/config regressions." + }, + { + "left": "issue:41720", + "right": "issue:44464", + "accept": false, + "reason": "A800 auto device-map cuda assert vs chunked generation inconsistency with compiled forward; unrelated." + }, + { + "left": "issue:33666", + "right": "issue:36010", + "accept": false, + "reason": "Qwen2-VL multi-GPU training issue vs GenerationMixin import error; unrelated." + }, + { + "left": "issue:43493", + "right": "issue:43825", + "accept": false, + "reason": "SigLIP2 implementation discrepancy vs pipeline translation-task error message; different bugs." + }, + { + "left": "issue:43262", + "right": "issue:45084", + "accept": false, + "reason": "Audio processor sampling-rate default bug vs template compilation TypeError; unrelated." + }, + { + "left": "issue:44849", + "right": "issue:45357", + "accept": false, + "reason": "Qwen3.5 output_hidden_states regression vs save_pretrained writing wrong visual-encoder keys; different parts of the model lifecycle." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 19, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:28282", + "issue:32090", + "issue:33357", + "issue:33666", + "issue:36331", + "issue:39692", + "issue:43257", + "issue:43278", + "issue:43295", + "issue:43299", + "issue:43334", + "issue:43493", + "issue:43504", + "issue:43540", + "issue:43572", + "issue:43575", + "issue:43577", + "issue:43606", + "issue:43646", + "issue:43653", + "issue:43688", + "issue:43716", + "issue:43720", + "issue:43756", + "issue:43824", + "issue:43844", + "issue:43927", + "issue:44291", + "issue:44361", + "issue:44423", + "issue:44466", + "issue:44479", + "issue:44488", + "issue:44554", + "issue:44589", + "issue:44625", + "issue:44734", + "issue:44843", + "issue:44964", + "issue:44977", + "issue:45083", + "issue:45200", + "issue:45229", + "issue:45405", + "issue:45520", + "issue:45701", + "issue:45704", + "issue:45759" + ], + "soft_pairs": [ + "issue:43540|issue:43716", + "issue:43720|issue:43927", + "issue:43606|issue:43927", + "issue:43334|issue:43572", + "issue:28282|issue:32090", + "issue:28282|issue:33357", + "issue:43577|issue:44554", + "issue:43688|issue:45704", + "issue:39692|issue:43493", + "issue:33666|issue:36331", + "issue:43334|issue:45759", + "issue:43720|issue:44361", + "issue:44964|issue:45405", + "issue:44479|issue:44625", + "issue:43540|issue:44977", + "issue:44488|issue:44843", + "issue:43493|issue:43844", + "issue:43756|issue:45083", + "issue:45200|issue:45229", + "issue:43257|issue:43299", + "issue:43278|issue:43844", + "issue:43824|issue:44589", + "issue:43646|issue:44466", + "issue:39692|issue:43295", + "issue:44589|issue:45520", + "issue:43688|issue:45701", + "issue:33357|issue:33666", + "issue:43504|issue:43606", + "issue:44423|issue:44734", + "issue:43572|issue:44291", + "issue:43927|issue:44361", + "issue:43504|issue:43653", + "issue:44625|issue:45759", + "issue:43575|issue:45405" + ], + "prepared_review_unit_hash": "0cf7a325194358c5398ae78978d12c74257fd0d106a55629ae6611e78624f5bc", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 34, + "serialized_chars": 38205, + "estimated_input_tokens": 9552, + "estimated_eval_tokens": 19360 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is a heterogeneous mix of unrelated bugs (model loading, tokenizer/config regressions, dtype/device issues, training/serving failures). None of the soft pairs looks like the same underlying defect under a strict duplicate standard.", + "confidence": 0.93, + "canonical_issue_reason": "No single canonical issue: the cluster is too heterogeneous for duplicate consolidation, with only a few loose Qwen/config-loading themes.", + "canonical_pr_reason": null, + "best_issue_reason": "No global best issue; at most, issue 45759 is the closest representative of a small Qwen multimodal config-loading subgroup, but it does not cover the rest of the cluster.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43540", + "right": "issue:43716", + "accept": false, + "reason": "Different models and failures: Qwen3OmniMoe video-processing ValueError vs Mistral-3 image-preprocessor dtype mismatch." + }, + { + "left": "issue:43720", + "right": "issue:43927", + "accept": false, + "reason": "Unrelated code paths: packed-weight loading in BitNet vs DiaConfig token-id persistence/generation IndexError." + }, + { + "left": "issue:43606", + "right": "issue:43927", + "accept": false, + "reason": "CPU offload device mismatch is unrelated to DiaConfig save/load token ID loss." + }, + { + "left": "issue:43334", + "right": "issue:43572", + "accept": false, + "reason": "Both mention missing pad token fields, but they affect different models/configs and likely different fixes." + }, + { + "left": "issue:28282", + "right": "issue:32090", + "accept": false, + "reason": "ImportError about missing PyTorch is unrelated to Trainer broadcast TypeError." + }, + { + "left": "issue:28282", + "right": "issue:33357", + "accept": false, + "reason": "Different failures: missing PyTorch ImportError vs MacOS bus error with a CLIP model." + }, + { + "left": "issue:43577", + "right": "issue:44554", + "accept": false, + "reason": "Different bug classes: BLIP2 dtype propagation vs MPS attention correctness when head dims differ." + }, + { + "left": "issue:43688", + "right": "issue:45704", + "accept": false, + "reason": "Auxiliary-loss normalization in MoE models is unrelated to T5/apex RMSNorm memory leak." + }, + { + "left": "issue:39692", + "right": "issue:43493", + "accept": false, + "reason": "Doc example errors are not the same as HF-vs-JAX implementation discrepancy for SigLIP2." + }, + { + "left": "issue:33666", + "right": "issue:36331", + "accept": false, + "reason": "Multi-GPU training request is not the same as CustomTrainer compute_loss signature regression." + }, + { + "left": "issue:43334", + "right": "issue:45759", + "accept": false, + "reason": "Both are Qwen multimodal config issues, but one is missing pad_token_id and the other is from_config text_config unwrapping; not the same bug." + }, + { + "left": "issue:43720", + "right": "issue:44361", + "accept": false, + "reason": "BitNet accelerate-loading issue is unrelated to MLukeTokenizer AttributeError." + }, + { + "left": "issue:44964", + "right": "issue:45405", + "accept": false, + "reason": "Model loading regression for Phi-4 multimodal is unrelated to an unreleased PEFT minimum-version bump." + }, + { + "left": "issue:44479", + "right": "issue:44625", + "accept": false, + "reason": "Video-input regression across Qwen variants is distinct from num_labels propagation in Qwen3.5 config." + }, + { + "left": "issue:43540", + "right": "issue:44977", + "accept": false, + "reason": "Different Qwen3.5/Qwen3OmniMoe video-processing failure vs flash-attention generation bug." + }, + { + "left": "issue:44488", + "right": "issue:44843", + "accept": false, + "reason": "Loading a specific model is unrelated to offline-mode breakage in AutoTokenizer._patch_mistral_regex." + }, + { + "left": "issue:43493", + "right": "issue:43844", + "accept": false, + "reason": "SigLIP2 implementation mismatch is unrelated to ZeRO-3 gradient inflation." + }, + { + "left": "issue:43756", + "right": "issue:45083", + "accept": false, + "reason": "Smollm3 RoPE-layer mismatch is unrelated to a helper-function behavior issue in qwen3_omni_moe." + }, + { + "left": "issue:45200", + "right": "issue:45229", + "accept": false, + "reason": "Text-only fine-tuning token-id defaulting is unrelated to Gemma4 multi-GPU CUDA OOM." + }, + { + "left": "issue:43257", + "right": "issue:43299", + "accept": false, + "reason": "Both are Qwen MOE-related, but one is weight conversion under accelerate/deepspeed and the other is composite model loading; not the same defect." + }, + { + "left": "issue:43278", + "right": "issue:43844", + "accept": false, + "reason": "Embedding dtype drift during eval is unrelated to gradient growth under HfDeepSpeedConfig + ZeRO-3." + }, + { + "left": "issue:43824", + "right": "issue:44589", + "accept": false, + "reason": "Missing Qwen2.5-VL export is unrelated to Float8 storage lookup failure." + }, + { + "left": "issue:43646", + "right": "issue:44466", + "accept": false, + "reason": "Custom model init breakage in v5 is unrelated to lm_head.weight serialization inconsistency." + }, + { + "left": "issue:39692", + "right": "issue:43295", + "accept": false, + "reason": "A broken doc example is not the same as a regression in custom model code using processor.tokenizer and image passing." + }, + { + "left": "issue:44589", + "right": "issue:45520", + "accept": false, + "reason": "Different import/runtime problems: missing Float8 storage class vs flash_attn KeyError on Python 3.13." + }, + { + "left": "issue:43688", + "right": "issue:45701", + "accept": false, + "reason": "MoE auxiliary-loss normalization is unrelated to tokenization differences across transformer versions." + }, + { + "left": "issue:33357", + "right": "issue:33666", + "accept": false, + "reason": "MacOS bus error in CLIP loading is unrelated to Qwen2-VL multi-GPU training." + }, + { + "left": "issue:43504", + "right": "issue:43606", + "accept": false, + "reason": "BEiT legacy-field loading failure is unrelated to suno/bark-small CPU offload device mismatch." + }, + { + "left": "issue:44423", + "right": "issue:44734", + "accept": false, + "reason": "Both hit serve, but one is continuous-batching multimodal crash and the other is KV-cache continuation tensor indexing; different bugs." + }, + { + "left": "issue:43572", + "right": "issue:44291", + "accept": false, + "reason": "Missing pad_token_idx after 5.0 update is unrelated to init_empty_weights rejecting _is_hf_initialized." + }, + { + "left": "issue:43927", + "right": "issue:44361", + "accept": false, + "reason": "DiaConfig token ID persistence bug is unrelated to MLukeTokenizer task AttributeError." + }, + { + "left": "issue:43504", + "right": "issue:43653", + "accept": false, + "reason": "BEiT legacy-field loading failure is unrelated to BigBirdTokenizer special-token registration." + }, + { + "left": "issue:44625", + "right": "issue:45759", + "accept": false, + "reason": "Both concern Qwen3.5 config handling, but one is num_labels propagation and the other is from_config text_config unwrapping; not the same underlying issue." + }, + { + "left": "issue:43575", + "right": "issue:45405", + "accept": false, + "reason": "Tensor-parallel OOM for a large model is unrelated to a PEFT version requirement bump." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 20, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:15354", + "issue:29942", + "issue:30333", + "issue:37428", + "issue:38175", + "issue:39401", + "issue:43232", + "issue:43278", + "issue:43295", + "issue:43334", + "issue:43493", + "issue:43525", + "issue:43540", + "issue:43606", + "issue:43688", + "issue:43701", + "issue:43844", + "issue:43873", + "issue:43957", + "issue:43994", + "issue:44112", + "issue:44315", + "issue:44464", + "issue:44466", + "issue:44493", + "issue:44521", + "issue:44554", + "issue:44568", + "issue:44610", + "issue:44792", + "issue:44849", + "issue:44898", + "issue:44964", + "issue:45072", + "issue:45084", + "issue:45127", + "issue:45198", + "issue:45200", + "issue:45290", + "issue:45325", + "issue:45356", + "issue:45405", + "issue:45412", + "issue:45446", + "issue:45520", + "issue:45584", + "issue:45701" + ], + "soft_pairs": [ + "issue:44792|issue:45405", + "issue:43606|issue:44112", + "issue:44521|issue:45290", + "issue:43994|issue:44493", + "issue:29942|issue:30333", + "issue:44464|issue:45584", + "issue:43844|issue:44964", + "issue:45084|issue:45520", + "issue:44554|issue:45446", + "issue:44568|issue:45701", + "issue:43688|issue:43873", + "issue:43540|issue:45200", + "issue:15354|issue:38175", + "issue:45072|issue:45198", + "issue:43232|issue:44849", + "issue:39401|issue:44568", + "issue:43278|issue:43701", + "issue:44315|issue:44849", + "issue:44554|issue:45356", + "issue:43295|issue:44610", + "issue:45200|issue:45325", + "issue:43334|issue:43525", + "issue:44898|issue:45412", + "issue:44466|issue:45127", + "issue:43493|issue:43701", + "issue:37428|issue:43957" + ], + "prepared_review_unit_hash": "1e34de14281a8fa704959989494164d873a1d9df73535df32823468d9d310a7e", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 26, + "serialized_chars": 35092, + "estimated_input_tokens": 8773, + "estimated_eval_tokens": 17802 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This set is a loose grab bag of unrelated Transformers issues: tokenization regressions, generation bugs, multimodal processor failures, device/offload problems, and CI/test noise. The proposed soft links are mostly superficial theme matches rather than the same underlying bug.", + "confidence": 0.91, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": "If one issue must represent the set, issue:45701 is the broadest/meta report ('transformers version changes the tokenization') and best captures the recurring version-regression theme, though it is not a true duplicate of the others.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44792", + "right": "issue:45405", + "accept": false, + "reason": "Janus image-generation test failure vs an unreleased PEFT-version constraint; unrelated bugs." + }, + { + "left": "issue:43606", + "right": "issue:44112", + "accept": false, + "reason": "Both are CI/device-related, but one is a CPU offload mismatch on Bark and the other is a stale device override test in GraniteSpeech." + }, + { + "left": "issue:44521", + "right": "issue:45290", + "accept": false, + "reason": "Both involve apply_chat_template, but one is multimodal assistant masks and the other is a crash on tool-call assistant messages; different concrete failures." + }, + { + "left": "issue:43994", + "right": "issue:44493", + "accept": false, + "reason": "SigLIP2 nonsensical outputs vs a broad unexpected position-id key issue; different code paths and symptoms." + }, + { + "left": "issue:29942", + "right": "issue:30333", + "accept": false, + "reason": "FlashAttention 2 test failures and MLflow job-reporting issues are unrelated." + }, + { + "left": "issue:44464", + "right": "issue:45584", + "accept": false, + "reason": "Compiled-forward chunked generation inconsistency vs Whisper empty-transcription generation failure; not the same bug." + }, + { + "left": "issue:43844", + "right": "issue:44964", + "accept": false, + "reason": "ZeRO-3 gradient growth and Phi-4 multimodal load failures are separate training/loading problems." + }, + { + "left": "issue:45084", + "right": "issue:45520", + "accept": false, + "reason": "Non-template node compilation error vs flash_attn import KeyError on Python 3.13; unrelated failures." + }, + { + "left": "issue:44554", + "right": "issue:45446", + "accept": false, + "reason": "MPS attention correctness issue and an AuxRequest version-check bug in flex_attention are different problems." + }, + { + "left": "issue:44568", + "right": "issue:45701", + "accept": false, + "reason": "Both are tokenizer regressions, but 44568 is a specific mdeberta BOS/EOS bug while 45701 is a broad version-change report; not one concrete duplicate." + }, + { + "left": "issue:43688", + "right": "issue:43873", + "accept": false, + "reason": "Aux-loss normalization and quantization offloading are different model-training/runtime issues." + }, + { + "left": "issue:43540", + "right": "issue:45200", + "accept": false, + "reason": "Qwen3OmniMoe video-processing error vs Gemma 4 mm_token_type_ids defaulting; different multimodal bugs." + }, + { + "left": "issue:15354", + "right": "issue:38175", + "accept": false, + "reason": "torch.jit.script GeneratorExp limitation vs SigLIP2 zero-probability outputs; unrelated." + }, + { + "left": "issue:45072", + "right": "issue:45198", + "accept": false, + "reason": "BFloat16 dtype mismatches in inference and Wav2Vec2 save/tokenization failures do not share a concrete code-path." + }, + { + "left": "issue:43232", + "right": "issue:44849", + "accept": false, + "reason": "Generation kwargs sync issue and Qwen3.5 hidden-states bug are separate generation-path regressions." + }, + { + "left": "issue:39401", + "right": "issue:44568", + "accept": false, + "reason": "Both are tokenizer bugs, but offset_mapping handling in Qwen3 and BOS/EOS insertion in mdeberta are distinct failures." + }, + { + "left": "issue:43278", + "right": "issue:43701", + "accept": false, + "reason": "BF16-to-FP32 embedding dtype drift vs resume_from_checkpoint key mismatch; unrelated." + }, + { + "left": "issue:44315", + "right": "issue:44849", + "accept": false, + "reason": "Liger Kernel application with model_init and Qwen3.5 hidden-states output are not the same bug." + }, + { + "left": "issue:44554", + "right": "issue:45356", + "accept": false, + "reason": "MPS attention correctness vs Kimi-K2.5 tokenizer codec handling regression; different subsystems." + }, + { + "left": "issue:43295", + "right": "issue:44610", + "accept": false, + "reason": "Processor/tokenizer API regression and OmDet-Turbo image-size mismatch are both processor-facing but not the same underlying issue." + }, + { + "left": "issue:45200", + "right": "issue:45325", + "accept": false, + "reason": "Gemma 4 token-type defaults and Qwen2.5-VL temporal position-id scaling are unrelated multimodal issues." + }, + { + "left": "issue:43334", + "right": "issue:43525", + "accept": false, + "reason": "Both mention missing pad_token_id, but they affect different model families and likely different config paths; too broad to merge." + }, + { + "left": "issue:44898", + "right": "issue:45412", + "accept": false, + "reason": "Perceiver resolution/interpolation failure and RT-DETR memory leak are unrelated." + }, + { + "left": "issue:44466", + "right": "issue:45127", + "accept": false, + "reason": "Serialization/tied-weights inconsistency and LoRA-merge collapse with extended vocab are different weight-handling bugs." + }, + { + "left": "issue:37428", + "right": "issue:43957", + "accept": false, + "reason": "Flash-attention import error and meta-device model-loading breakage are separate loading/runtime issues." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 21, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:37428", + "issue:38175", + "issue:42915", + "issue:43262", + "issue:43278", + "issue:43295", + "issue:43425", + "issue:43450", + "issue:43493", + "issue:43572", + "issue:43575", + "issue:43606", + "issue:43653", + "issue:43819", + "issue:43824", + "issue:43825", + "issue:43827", + "issue:43844", + "issue:43856", + "issue:43873", + "issue:43901", + "issue:43957", + "issue:43975", + "issue:43994", + "issue:44112", + "issue:44206", + "issue:44360", + "issue:44442", + "issue:44464", + "issue:44514", + "issue:44521", + "issue:44561", + "issue:44610", + "issue:44792", + "issue:44857", + "issue:44936", + "issue:44987", + "issue:45083", + "issue:45127", + "issue:45198", + "issue:45245", + "issue:45290", + "issue:45381", + "issue:45412", + "issue:45563", + "issue:45584", + "issue:45593" + ], + "soft_pairs": [ + "issue:43572|issue:44987", + "issue:43572|issue:43957", + "issue:43873|issue:44464", + "issue:43824|issue:43975", + "issue:43295|issue:43450", + "issue:43844|issue:44936", + "issue:43295|issue:43873", + "issue:38175|issue:45412", + "issue:44514|issue:44521", + "issue:44442|issue:45198", + "issue:43278|issue:43856", + "issue:43295|issue:43994", + "issue:37428|issue:43824", + "issue:42915|issue:43824", + "issue:38175|issue:43493", + "issue:43827|issue:43901", + "issue:43653|issue:44112", + "issue:43450|issue:43493", + "issue:44360|issue:44561", + "issue:43450|issue:43575", + "issue:44792|issue:45563", + "issue:45127|issue:45245", + "issue:43819|issue:45083", + "issue:44206|issue:45584", + "issue:43262|issue:43450", + "issue:43819|issue:44112", + "issue:37428|issue:43425", + "issue:44610|issue:44857", + "issue:43278|issue:43825", + "issue:45290|issue:45381", + "issue:44857|issue:45593", + "issue:43606|issue:45083" + ], + "prepared_review_unit_hash": "6c84a3c27a42b637b5bcee019b73325cd45a8bc6aaef65e34dbecf2b264f69b2", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 32, + "serialized_chars": 36641, + "estimated_input_tokens": 9161, + "estimated_eval_tokens": 18578 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly a set of unrelated transformer bug reports that only share broad subsystem vocabulary. I only accept the SigLIP2 pair as likely the same underlying model discrepancy; the rest look like distinct issues.", + "confidence": 0.78, + "canonical_issue_reason": "issue:43493 is the best canonical issue because it describes the broad SigLIP2 implementation discrepancy at the root of the symptom-specific output failures, making it the most general and reusable duplicate target in this set.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43493 is the strongest representative of the only plausible duplicate chain here: it frames the underlying SigLIP2 HF-vs-JAX mismatch, whereas nearby reports are narrower symptoms.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43572", + "right": "issue:44987", + "accept": false, + "reason": "Both are version/load failures, but one is a missing config field and the other is a model load OOM/regression; different code paths." + }, + { + "left": "issue:43572", + "right": "issue:43957", + "accept": false, + "reason": "Different failure modes: missing StableLm config field vs meta-device model loading breakage." + }, + { + "left": "issue:43873", + "right": "issue:44464", + "accept": false, + "reason": "Quantization/offloading behavior and compiled chunked generation are unrelated bugs." + }, + { + "left": "issue:43824", + "right": "issue:43975", + "accept": false, + "reason": "ImportError for a missing Qwen2.5-VL symbol vs DeepSeek detokenization issue; not the same bug." + }, + { + "left": "issue:43295", + "right": "issue:43450", + "accept": false, + "reason": "Processor.tokenizer regression and batched video shape bug are different processor issues." + }, + { + "left": "issue:43844", + "right": "issue:44936", + "accept": false, + "reason": "DeepSpeed/ZeRO gradient inflation and trainer.evaluate() after train() are distinct training-loop problems." + }, + { + "left": "issue:43295", + "right": "issue:43873", + "accept": false, + "reason": "Custom processor regression and quantization offloading bug are unrelated." + }, + { + "left": "issue:38175", + "right": "issue:45412", + "accept": false, + "reason": "SigLIP2 zero probabilities and RT-DETR memory cleanup are unrelated models and symptoms." + }, + { + "left": "issue:44514", + "right": "issue:44521", + "accept": false, + "reason": "Both involve apply_chat_template, but one is a batched-padding crash and the other is incorrect assistant masks; not enough evidence of one underlying bug." + }, + { + "left": "issue:44442", + "right": "issue:45198", + "accept": false, + "reason": "Different tokenizer/model families with separate tokenizer/save_pretrained failures." + }, + { + "left": "issue:43278", + "right": "issue:43856", + "accept": false, + "reason": "Embedding dtype drift in eval vs Qwen3 MoE memory efficiency are unrelated." + }, + { + "left": "issue:43295", + "right": "issue:43994", + "accept": false, + "reason": "Processor.tokenizer regression is not the same as SigLIP2 nonsensical outputs." + }, + { + "left": "issue:37428", + "right": "issue:43824", + "accept": false, + "reason": "Different missing imports from different modules and model families." + }, + { + "left": "issue:42915", + "right": "issue:43824", + "accept": false, + "reason": "FineGrainedFP8Config failure and missing Qwen2.5-VL export are unrelated." + }, + { + "left": "issue:38175", + "right": "issue:43493", + "accept": true, + "reason": "Both point to the same SigLIP2 implementation problem: HF outputs are wrong/non-sensical relative to expected behavior, with 38175 showing a symptom and 43493 describing the broader mismatch." + }, + { + "left": "issue:43827", + "right": "issue:43901", + "accept": false, + "reason": "Both are docs-related, but they concern different pipeline APIs and different stale-deprecation issues." + }, + { + "left": "issue:43653", + "right": "issue:44112", + "accept": false, + "reason": "Tokenizer special-token registration and a GraniteSpeech CI device override test failure are unrelated." + }, + { + "left": "issue:43450", + "right": "issue:43493", + "accept": false, + "reason": "Video batching shape bug and SigLIP2 implementation discrepancy affect different models and code paths." + }, + { + "left": "issue:44360", + "right": "issue:44561", + "accept": false, + "reason": "DSA indexer design discussion and removal of is_torch_fx_available are unrelated." + }, + { + "left": "issue:43450", + "right": "issue:43575", + "accept": false, + "reason": "Batched video processor shape bug and tensor-parallel OOM on Qwen2-57B are different issues." + }, + { + "left": "issue:44792", + "right": "issue:45563", + "accept": false, + "reason": "Janus image generation test failure and paged generate warning are unrelated generation issues." + }, + { + "left": "issue:45127", + "right": "issue:45245", + "accept": false, + "reason": "LoRA/tied-embeddings collapse and category-count RuntimeError are different model-training failures." + }, + { + "left": "issue:43819", + "right": "issue:45083", + "accept": false, + "reason": "DAC latent forward mismatch and qwen3_omni_moe feature-length helper behavior are unrelated." + }, + { + "left": "issue:44206", + "right": "issue:45584", + "accept": false, + "reason": "Feature extractor center-arg crash and Whisper empty-transcription failure are different audio bugs." + }, + { + "left": "issue:43262", + "right": "issue:43450", + "accept": false, + "reason": "Audio chat-template sampling-rate default and video batched-shape bug are different processor issues." + }, + { + "left": "issue:43819", + "right": "issue:44112", + "accept": false, + "reason": "DAC STE mismatch and GraniteSpeech CI device override failure are unrelated." + }, + { + "left": "issue:37428", + "right": "issue:43425", + "accept": false, + "reason": "Flash-attention import error and Torch 2.10 incompatibility are different compatibility issues." + }, + { + "left": "issue:44610", + "right": "issue:44857", + "accept": false, + "reason": "Processor output-size mismatch and CUDA AMP loss crash are unrelated detection bugs." + }, + { + "left": "issue:43278", + "right": "issue:43825", + "accept": false, + "reason": "Training/eval dtype drift and pipeline docs/error-message wording are unrelated." + }, + { + "left": "issue:45290", + "right": "issue:45381", + "accept": false, + "reason": "Chat-template tool-call crash and Qwen2.5-VL vision_position_ids bug are different failure modes." + }, + { + "left": "issue:44857", + "right": "issue:45593", + "accept": false, + "reason": "Float16/CUDA loss crash and missing auxiliary losses in denoising-off mode are different detection training bugs." + }, + { + "left": "issue:43606", + "right": "issue:45083", + "accept": false, + "reason": "CPU offload device mismatch and qwen3_omni_moe feature-length helper behavior are unrelated." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet and stays conservative: it treats most pairs as unrelated and only accepts the SigLIP2 pair, which is plausible from the titles. The canonical choice of issue:43493 is also reasonable as the broader SigLIP2 discrepancy report. No obvious overstatement beyond what the packet supports." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 22, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:29127", + "issue:29942", + "issue:32090", + "issue:34689", + "issue:35141", + "issue:37428", + "issue:38175", + "issue:39692", + "issue:40444", + "issue:43335", + "issue:43388", + "issue:43404", + "issue:43504", + "issue:43526", + "issue:43550", + "issue:43606", + "issue:43611", + "issue:43653", + "issue:43720", + "issue:43819", + "issue:43931", + "issue:44112", + "issue:44292", + "issue:44351", + "issue:44448", + "issue:44462", + "issue:44493", + "issue:44561", + "issue:44568", + "issue:44743", + "issue:44779", + "issue:44855", + "issue:44857", + "issue:44977", + "issue:45072", + "issue:45083", + "issue:45200", + "issue:45229", + "issue:45310", + "issue:45372", + "issue:45459", + "issue:45478", + "issue:45491", + "issue:45520", + "issue:45563", + "issue:45588", + "issue:45593", + "issue:45712" + ], + "soft_pairs": [ + "issue:44448|issue:44743", + "issue:45200|issue:45588", + "issue:43335|issue:43550", + "issue:44462|issue:44561", + "issue:44561|issue:44779", + "issue:37428|issue:44351", + "issue:43611|issue:43931", + "issue:43388|issue:43526", + "issue:39692|issue:43526", + "issue:43819|issue:44493", + "issue:45310|issue:45478", + "issue:43335|issue:43653", + "issue:29942|issue:32090", + "issue:43404|issue:45200", + "issue:44855|issue:45083", + "issue:44292|issue:45520", + "issue:39692|issue:44493", + "issue:43335|issue:45083", + "issue:43550|issue:44112", + "issue:43335|issue:43504", + "issue:44448|issue:44779", + "issue:38175|issue:44493", + "issue:43335|issue:45072", + "issue:44977|issue:45229", + "issue:43720|issue:44112", + "issue:43335|issue:45593", + "issue:45459|issue:45712", + "issue:44568|issue:45491", + "issue:44448|issue:44977", + "issue:29127|issue:32090", + "issue:43606|issue:44857", + "issue:40444|issue:43526", + "issue:45372|issue:45588", + "issue:44857|issue:45072", + "issue:34689|issue:35141", + "issue:44977|issue:45563" + ], + "prepared_review_unit_hash": "d9666ce86a147524e89affa9712db5eaa65089a2006b28968307198ed3d2cf94", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 36, + "serialized_chars": 38870, + "estimated_input_tokens": 9718, + "estimated_eval_tokens": 19692 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly a grab-bag of unrelated Transformers bug reports. The only clear duplicate pair is the two identical Qwen3.5 MoE from_pretrained errors; the rest share broad themes like model loading, tokenizers, or flash-attention but point to different code paths and should stay separate.", + "confidence": 0.94, + "canonical_issue_reason": "issue:45310 is the earlier report of the same Qwen3.5 MoE from_pretrained failure and is the cleaner canonical target.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45310 is the best issue candidate because it is the earliest and directly matches the duplicate report in issue:45478.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44448", + "right": "issue:44743", + "accept": false, + "reason": "Different problems: Pegasus output regression vs Qwen3.5 recurrent-state reset in modular code." + }, + { + "left": "issue:45200", + "right": "issue:45588", + "accept": false, + "reason": "Unrelated: Gemma 4 token-type IDs defaulting vs flash-attention crash on sink-less models." + }, + { + "left": "issue:43335", + "right": "issue:43550", + "accept": false, + "reason": "Different models and failures: SwitchTransformers sparse-layer config vs Bamba torch.compile/SDPA." + }, + { + "left": "issue:44462", + "right": "issue:44561", + "accept": false, + "reason": "Tokenizer.json loading bug is unrelated to removal of is_torch_fx_available breaking trust_remote_code models." + }, + { + "left": "issue:44561", + "right": "issue:44779", + "accept": false, + "reason": "Different tokenizer/runtime regressions with different symptoms and likely different fixes." + }, + { + "left": "issue:37428", + "right": "issue:44351", + "accept": false, + "reason": "Both are import errors, but for different missing symbols and different affected features." + }, + { + "left": "issue:43611", + "right": "issue:43931", + "accept": false, + "reason": "Model loading breaks for different reasons: base_model_prefix handling vs weight-shape mismatch." + }, + { + "left": "issue:43388", + "right": "issue:43526", + "accept": false, + "reason": "Both concern labels, but one is gather_for_metrics truncation and the other is BeitImageProcessorFast reduce_labels behavior." + }, + { + "left": "issue:39692", + "right": "issue:43526", + "accept": false, + "reason": "Documentation/example errors are unrelated to the Beit label reduction bug." + }, + { + "left": "issue:43819", + "right": "issue:44493", + "accept": false, + "reason": "DAC latent conversion mismatch is unrelated to unexpected position-id keys." + }, + { + "left": "issue:45310", + "right": "issue:45478", + "accept": true, + "reason": "Same title and same concrete failure: transformers>=5.4.0 Qwen3.5 MoE from_pretrained error." + }, + { + "left": "issue:43335", + "right": "issue:43653", + "accept": false, + "reason": "Different tokenizer/model areas: SwitchTransformers config vs BigBirdTokenizer special-token registration." + }, + { + "left": "issue:29942", + "right": "issue:32090", + "accept": false, + "reason": "Flash Attention 2 test failures are unrelated to Trainer _gpu_broadcast_one NoneType errors." + }, + { + "left": "issue:43404", + "right": "issue:45200", + "accept": false, + "reason": "Different bugs: untied lm_head weights vs missing mm_token_type_ids default." + }, + { + "left": "issue:44855", + "right": "issue:45083", + "accept": false, + "reason": "Python 3.13 JIT parsing issue is unrelated to qwen3_omni_moe feature-length computation." + }, + { + "left": "issue:44292", + "right": "issue:45520", + "accept": false, + "reason": "Qwen-3-8B-NVFP4 runtime issue is not the same as a flash_attn import-utils KeyError on Python 3.13." + }, + { + "left": "issue:39692", + "right": "issue:44493", + "accept": false, + "reason": "SigLIP2 doc/example mistakes are unrelated to the position-id key regression." + }, + { + "left": "issue:43335", + "right": "issue:45083", + "accept": false, + "reason": "Different subsystems and failure modes; no shared underlying bug." + }, + { + "left": "issue:43550", + "right": "issue:44112", + "accept": false, + "reason": "Bamba SDPA compile failure is unrelated to a GraniteSpeech CI device-override test." + }, + { + "left": "issue:43335", + "right": "issue:43504", + "accept": false, + "reason": "SwitchTransformers sparse-layer config bug is unrelated to BEiT legacy-field loading failure." + }, + { + "left": "issue:44448", + "right": "issue:44779", + "accept": false, + "reason": "Different model regressions: Pegasus output mismatch vs DeepSeek tokenizer behavior change." + }, + { + "left": "issue:38175", + "right": "issue:44493", + "accept": false, + "reason": "SigLIP2 zero probabilities and unexpected position-id keys are different bugs." + }, + { + "left": "issue:43335", + "right": "issue:45072", + "accept": false, + "reason": "SwitchTransformers config bug and bfloat16 dtype mismatch in inference are unrelated." + }, + { + "left": "issue:44977", + "right": "issue:45229", + "accept": false, + "reason": "Qwen3.5 flash-attention generation bug is unrelated to Gemma4 multi-GPU CUDA OOM." + }, + { + "left": "issue:43720", + "right": "issue:44112", + "accept": false, + "reason": "Packed-weight loading failure in BitNet is unrelated to a stale GraniteSpeech CI test." + }, + { + "left": "issue:43335", + "right": "issue:45593", + "accept": false, + "reason": "SwitchTransformers sparse-layer creation and D-FINE auxiliary-loss behavior are different issues." + }, + { + "left": "issue:45459", + "right": "issue:45712", + "accept": false, + "reason": "Tokenizer error masking due to missing protobuf is unrelated to dummy-class cleanup/check_repo failures." + }, + { + "left": "issue:44568", + "right": "issue:45491", + "accept": false, + "reason": "Tokenizer special-token behavior is unrelated to Gemma3 NaN embeddings from mixed-length batching." + }, + { + "left": "issue:44448", + "right": "issue:44977", + "accept": false, + "reason": "Different model families and symptoms: Pegasus output drift vs Qwen3.5 flash-attention generation issues." + }, + { + "left": "issue:29127", + "right": "issue:32090", + "accept": false, + "reason": "LayoutLMV3 error-message clarity is unrelated to Trainer broadcast NoneType errors." + }, + { + "left": "issue:43606", + "right": "issue:44857", + "accept": false, + "reason": "CPU-offload device mismatch and float16 AMP CUDA crash are distinct bugs." + }, + { + "left": "issue:40444", + "right": "issue:43526", + "accept": false, + "reason": "Qwen2.5-VL iterable-dataset finetuning failure is unrelated to BeitImageProcessorFast label reduction." + }, + { + "left": "issue:45372", + "right": "issue:45588", + "accept": false, + "reason": "Gemma 4 processor loading ImportError is unrelated to flash-attention sink-less-model crash." + }, + { + "left": "issue:44857", + "right": "issue:45072", + "accept": false, + "reason": "Different precision-related failures in different models and code paths." + }, + { + "left": "issue:34689", + "right": "issue:35141", + "accept": false, + "reason": "Model-loading regression for Llama 3.2 Vision is unrelated to token-embedding reinit after resize." + }, + { + "left": "issue:44977", + "right": "issue:45563", + "accept": false, + "reason": "Qwen3.5 flash-attention generation bug is unrelated to a stale paging warning for num_return_sequences." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet and stays conservative. The duplicate call on issue:45310 / issue:45478 is supported by the identical title and matching failure description, and the other soft-edge rejections are framed narrowly enough to avoid overstating similarity." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 23, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:25251", + "issue:29942", + "issue:30064", + "issue:32090", + "issue:35141", + "issue:36010", + "issue:37428", + "issue:42915", + "issue:42994", + "issue:43065", + "issue:43066", + "issue:43278", + "issue:43381", + "issue:43412", + "issue:43421", + "issue:43550", + "issue:43653", + "issue:43701", + "issue:43873", + "issue:43927", + "issue:43976", + "issue:43994", + "issue:44186", + "issue:44206", + "issue:44265", + "issue:44361", + "issue:44448", + "issue:44451", + "issue:44462", + "issue:44466", + "issue:44479", + "issue:44493", + "issue:44743", + "issue:44857", + "issue:44871", + "issue:44977", + "issue:45083", + "issue:45198", + "issue:45245", + "issue:45372", + "issue:45412", + "issue:45440", + "issue:45459", + "issue:45491", + "issue:45520", + "issue:45563", + "issue:45588", + "issue:45698" + ], + "soft_pairs": [ + "issue:43701|issue:45083", + "issue:43927|issue:45698", + "issue:43550|issue:45083", + "issue:43066|issue:44462", + "issue:43412|issue:45588", + "issue:45245|issue:45563", + "issue:43873|issue:44493", + "issue:29942|issue:35141", + "issue:45372|issue:45459", + "issue:36010|issue:37428", + "issue:44448|issue:44466", + "issue:43065|issue:45412", + "issue:25251|issue:32090", + "issue:43278|issue:43381", + "issue:44857|issue:45588", + "issue:44451|issue:45520", + "issue:43653|issue:44743", + "issue:44186|issue:45198", + "issue:43976|issue:45440", + "issue:43994|issue:45412", + "issue:44265|issue:45198", + "issue:44361|issue:45198", + "issue:42915|issue:42994", + "issue:43421|issue:45491", + "issue:44871|issue:44977", + "issue:29942|issue:30064", + "issue:43065|issue:43994", + "issue:44206|issue:44479" + ], + "prepared_review_unit_hash": "9aa90e8e356efdb5534fa79bd40b129e10b955e842351496bc07b85b6bb90992", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 28, + "serialized_chars": 36052, + "estimated_input_tokens": 9013, + "estimated_eval_tokens": 18282 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The cluster is very heterogeneous: most items are unrelated bugs/regressions across tokenization, export, quantization, flash-attention, and model-specific behavior. I found no true duplicate pair among the soft edges.", + "confidence": 0.94, + "canonical_issue_reason": "No single canonical issue stands out because the cluster does not cohere around one underlying bug; the items span many distinct code paths and symptoms.", + "canonical_pr_reason": "No PRs are present in this cluster.", + "best_issue_reason": "If one issue must be used as a representative, issue 44462 is the cleanest self-contained report with a concrete, actionable loading symptom around AutoTokenizer/tokenizer.json.", + "best_pr_reason": "No PRs are present, so there is no best PR candidate.", + "soft_edge_verdicts": [ + { + "left": "issue:43701", + "right": "issue:45083", + "accept": false, + "reason": "Checkpoint-resume key mismatch and qwen3_omni_moe feature-length behavior are different subsystems and different bugs." + }, + { + "left": "issue:43927", + "right": "issue:45698", + "accept": false, + "reason": "Both involve save/load, but one is about custom token IDs in config and the other about loading the wrong custom module." + }, + { + "left": "issue:43550", + "right": "issue:45083", + "accept": false, + "reason": "torch.compile/SDPA failure in Bamba-9B-v2 is unrelated to the qwen3_omni_moe helper-function issue." + }, + { + "left": "issue:43066", + "right": "issue:44462", + "accept": false, + "reason": "Both touch tokenizers, but one is a decoder-type mismatch in v5 while the other is ignoring tokenizer.json from the repo." + }, + { + "left": "issue:43412", + "right": "issue:45588", + "accept": false, + "reason": "Executorch export/erfinv/training-guard failure and flash_attention s_aux=None crash are different code paths." + }, + { + "left": "issue:45245", + "right": "issue:45563", + "accept": false, + "reason": "Category-count overflow and a stale num_return_sequences warning are unrelated." + }, + { + "left": "issue:43873", + "right": "issue:44493", + "accept": false, + "reason": "Quantization offloading behavior and unexpected position-id key warnings are not the same bug." + }, + { + "left": "issue:29942", + "right": "issue:35141", + "accept": false, + "reason": "Flash Attention 2 test failures and embedding reinit in post_init are distinct regressions." + }, + { + "left": "issue:45372", + "right": "issue:45459", + "accept": false, + "reason": "ReasoningEffort import breakage and protobuf error masking affect different tokenizer/processor paths." + }, + { + "left": "issue:36010", + "right": "issue:37428", + "accept": false, + "reason": "These are both import errors, but for different symbols in different modules." + }, + { + "left": "issue:44448", + "right": "issue:44466", + "accept": false, + "reason": "Pegasus output differences and lm_head.weight serialization inconsistency are separate model-generation vs save/load issues." + }, + { + "left": "issue:43065", + "right": "issue:45412", + "accept": false, + "reason": "Sam3PixelDecoder dummy Conv2d and RT-DETR memory release are unrelated." + }, + { + "left": "issue:25251", + "right": "issue:32090", + "accept": false, + "reason": "Pipeline top_k nesting behavior and Trainer _gpu_broadcast_one NoneType error are unrelated failures." + }, + { + "left": "issue:43278", + "right": "issue:43381", + "accept": false, + "reason": "Embedding dtype drift in eval and gradient checkpointing being disallowed in eval are different eval-mode problems." + }, + { + "left": "issue:44857", + "right": "issue:45588", + "accept": false, + "reason": "LwDetrImageLoss AMP/CUDA crash and flash_attention s_aux=None crash do not share the same underlying defect." + }, + { + "left": "issue:44451", + "right": "issue:45520", + "accept": false, + "reason": "Model load failure and Python 3.13 flash_attn KeyError are different import/runtime problems." + }, + { + "left": "issue:43653", + "right": "issue:44743", + "accept": false, + "reason": "BigBirdTokenizer special-token decoding and Qwen recurrent-state reset are unrelated." + }, + { + "left": "issue:44186", + "right": "issue:45198", + "accept": false, + "reason": "LayoutLMv2Tokenizer padding/truncation crash and Wav2Vec2 save_pretrained/tokenization failure are different tokenizer bugs." + }, + { + "left": "issue:43976", + "right": "issue:45440", + "accept": false, + "reason": "Python-version incompatibility and DeepseekV3MoE divergence are not the same issue." + }, + { + "left": "issue:43994", + "right": "issue:45412", + "accept": false, + "reason": "Siglip2 nonsensical outputs and RT-DETR memory leakage are unrelated." + }, + { + "left": "issue:44265", + "right": "issue:45198", + "accept": false, + "reason": "torch.export.export with torch_compilable_check and Wav2Vec2 save/tokenization failure are different defects." + }, + { + "left": "issue:44361", + "right": "issue:45198", + "accept": false, + "reason": "MLukeTokenizer AttributeError and Wav2Vec2 save/tokenization failure are unrelated tokenizer issues." + }, + { + "left": "issue:42915", + "right": "issue:42994", + "accept": false, + "reason": "Both mention quantization, but one is a FineGrainedFP8Config runtime failure and the other is quantized model saving failure." + }, + { + "left": "issue:43421", + "right": "issue:45491", + "accept": false, + "reason": "Runtime special-token post-processor updates and Gemma3 NaN embeddings are different bugs." + }, + { + "left": "issue:44871", + "right": "issue:44977", + "accept": false, + "reason": "Gemma eos_token_id mismatch and Qwen3.5 flash-attention generation regression are unrelated." + }, + { + "left": "issue:29942", + "right": "issue:30064", + "accept": false, + "reason": "Flash Attention 2 test failures and void segmentation map processing failure are unrelated." + }, + { + "left": "issue:44206", + "right": "issue:44479", + "accept": false, + "reason": "Unsupported center argument in LasrFeatureExtractor and Qwen VL video regression are different issues." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 24, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:29127", + "issue:29942", + "issue:32090", + "issue:33357", + "issue:34567", + "issue:36010", + "issue:39692", + "issue:41720", + "issue:42915", + "issue:43262", + "issue:43278", + "issue:43329", + "issue:43334", + "issue:43381", + "issue:43441", + "issue:43526", + "issue:43531", + "issue:43575", + "issue:43653", + "issue:43701", + "issue:43723", + "issue:43756", + "issue:43824", + "issue:43873", + "issue:43931", + "issue:44186", + "issue:44206", + "issue:44279", + "issue:44442", + "issue:44464", + "issue:44554", + "issue:44610", + "issue:44734", + "issue:44743", + "issue:44811", + "issue:44857", + "issue:44977", + "issue:45072", + "issue:45103", + "issue:45229", + "issue:45245", + "issue:45372", + "issue:45406", + "issue:45412", + "issue:45440", + "issue:45491", + "issue:45588", + "issue:45593" + ], + "soft_pairs": [ + "issue:44734|issue:45406", + "issue:33357|issue:39692", + "issue:44610|issue:45412", + "issue:29127|issue:34567", + "issue:45103|issue:45372", + "issue:43873|issue:45229", + "issue:43756|issue:45440", + "issue:45372|issue:45406", + "issue:43723|issue:44554", + "issue:43278|issue:43575", + "issue:43278|issue:43329", + "issue:43381|issue:43701", + "issue:43653|issue:44442", + "issue:45440|issue:45593", + "issue:41720|issue:45229", + "issue:29942|issue:36010", + "issue:43824|issue:44279", + "issue:32090|issue:33357", + "issue:43334|issue:43531", + "issue:43526|issue:44186", + "issue:44977|issue:45245", + "issue:43526|issue:44857", + "issue:42915|issue:43931", + "issue:43441|issue:45588", + "issue:43262|issue:44206", + "issue:44464|issue:45491", + "issue:43526|issue:44610", + "issue:44743|issue:44811", + "issue:44442|issue:45072" + ], + "prepared_review_unit_hash": "9100251c14a0e352eff00225799b65ae5670905cd88fdd9b5532dabc3cdaf994", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 29, + "serialized_chars": 36476, + "estimated_input_tokens": 9119, + "estimated_eval_tokens": 18494 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is a broad false-positive mix of unrelated issues across tokenizers, model loading, training, attention, quantization, and docs. None of the soft pairs describe the same concrete bug or change, so no duplicates should be merged here.", + "confidence": 0.98, + "canonical_issue_reason": "No canonical issue: the items are heterogeneous and do not share one underlying bug, code path, or fix surface.", + "canonical_pr_reason": null, + "best_issue_reason": "No single best issue for deduplication; none stands out as the representative of a real duplicate group.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44734", + "right": "issue:45406", + "accept": false, + "reason": "Different failures: KV-cache continuation tensor indexing vs Gemma4Processor missing _tokenizer." + }, + { + "left": "issue:33357", + "right": "issue:39692", + "accept": false, + "reason": "Unrelated: MacOS CLIP bus error vs SigLIP2 doc example/model-processor mismatch and quantization issue." + }, + { + "left": "issue:44610", + "right": "issue:45412", + "accept": false, + "reason": "Different subsystems: processor image-size mismatch vs RT-DETR memory not released on deletion." + }, + { + "left": "issue:29127", + "right": "issue:34567", + "accept": false, + "reason": "Different problems: LayoutLMv3 error clarity vs TrainerState token-count tracking not updating." + }, + { + "left": "issue:45103", + "right": "issue:45372", + "accept": false, + "reason": "Unrelated: auto_docstring crash with future annotations vs mistral_common import breakage for Gemma 4 loading." + }, + { + "left": "issue:43873", + "right": "issue:45229", + "accept": false, + "reason": "Different issues: quantization offloading behavior vs Gemma4 multi-GPU CUDA OOM." + }, + { + "left": "issue:43756", + "right": "issue:45440", + "accept": false, + "reason": "Different model-specific bugs: Smollm3 RoPE layer handling vs DeepseekV3MoE divergence from remote implementation." + }, + { + "left": "issue:45372", + "right": "issue:45406", + "accept": false, + "reason": "ImportError in Gemma4 processor loading is unrelated to serve-time _tokenizer attribute crash." + }, + { + "left": "issue:43723", + "right": "issue:44554", + "accept": false, + "reason": "Different areas: AutoTokenizer loading regression vs MPS attention correctness bug." + }, + { + "left": "issue:43278", + "right": "issue:43575", + "accept": false, + "reason": "Embedding dtype mismatch in eval vs tensor-parallel OOM on model load are unrelated." + }, + { + "left": "issue:43278", + "right": "issue:43329", + "accept": false, + "reason": "No shared code path: dtype change in evaluate vs multimodal token counting bug." + }, + { + "left": "issue:43381", + "right": "issue:43701", + "accept": false, + "reason": "Gradient checkpointing eval-mode restriction is unrelated to checkpoint resume key mismatch." + }, + { + "left": "issue:43653", + "right": "issue:44442", + "accept": false, + "reason": "Different tokenizer bugs: special-token registration vs AutoTokenizer failing to load a specific tokenizer class." + }, + { + "left": "issue:45440", + "right": "issue:45593", + "accept": false, + "reason": "Unrelated model correctness issue vs D-FINE auxiliary-loss behavior." + }, + { + "left": "issue:41720", + "right": "issue:45229", + "accept": false, + "reason": "Different GPU inference failures: Qwen3 auto device mapping assert vs Gemma4 multi-GPU OOM." + }, + { + "left": "issue:29942", + "right": "issue:36010", + "accept": false, + "reason": "Flash Attention test failures are unrelated to a GenerationMixin import error." + }, + { + "left": "issue:43824", + "right": "issue:44279", + "accept": false, + "reason": "Specific missing symbol import vs generic dependency issue; not the same bug." + }, + { + "left": "issue:32090", + "right": "issue:33357", + "accept": false, + "reason": "Trainer broadcast NoneType error and MacOS CLIP bus error are unrelated." + }, + { + "left": "issue:43334", + "right": "issue:43531", + "accept": false, + "reason": "Different Qwen3 variants and different failures: missing pad_token_id vs sliding_window behavior." + }, + { + "left": "issue:43526", + "right": "issue:44186", + "accept": false, + "reason": "BEiT label reduction bug vs LayoutLMv2 tokenizer crash on NER/padding are unrelated." + }, + { + "left": "issue:44977", + "right": "issue:45245", + "accept": false, + "reason": "Flash-attention generation issue is unrelated to the 2^24 category limit runtime error." + }, + { + "left": "issue:43526", + "right": "issue:44857", + "accept": false, + "reason": "Different modules and symptoms: image-label reduction vs LwDetrImageLoss AMP crash." + }, + { + "left": "issue:42915", + "right": "issue:43931", + "accept": false, + "reason": "Qwen3Moe FP8 config failure is not the same as Qwen3-VL weight-shape mismatch." + }, + { + "left": "issue:43441", + "right": "issue:45588", + "accept": false, + "reason": "Both mention flash attention, but they affect different code paths and concrete crashes." + }, + { + "left": "issue:43262", + "right": "issue:44206", + "accept": false, + "reason": "Audio chat-template sample-rate defaulting is unrelated to an unsupported feature-extractor argument." + }, + { + "left": "issue:44464", + "right": "issue:45491", + "accept": false, + "reason": "Compiled-forward chunked generation inconsistency is unrelated to Gemma3 NaN embeddings from sliding-window attention." + }, + { + "left": "issue:43526", + "right": "issue:44610", + "accept": false, + "reason": "Image label reduction bug and OmDet-Turbo input-size mismatch are distinct processor issues." + }, + { + "left": "issue:44743", + "right": "issue:44811", + "accept": false, + "reason": "Recurrent-state cache reset bug is unrelated to Whisper batch_decode skip_special_tokens handling." + }, + { + "left": "issue:44442", + "right": "issue:45072", + "accept": false, + "reason": "Tokenizer-loading failure and dtype mismatches in bfloat16 inference are unrelated." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 25, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:29127", + "issue:33290", + "issue:33357", + "issue:34567", + "issue:34689", + "issue:35141", + "issue:38175", + "issue:39692", + "issue:42907", + "issue:43295", + "issue:43316", + "issue:43381", + "issue:43452", + "issue:43493", + "issue:43525", + "issue:43550", + "issue:43575", + "issue:43611", + "issue:43646", + "issue:43701", + "issue:43761", + "issue:43827", + "issue:43873", + "issue:43931", + "issue:43994", + "issue:44077", + "issue:44079", + "issue:44188", + "issue:44206", + "issue:44387", + "issue:44442", + "issue:44479", + "issue:44568", + "issue:44743", + "issue:44898", + "issue:44964", + "issue:44977", + "issue:45072", + "issue:45216", + "issue:45229", + "issue:45290", + "issue:45440", + "issue:45479", + "issue:45491", + "issue:45584", + "issue:45588", + "issue:45600", + "issue:45698" + ], + "soft_pairs": [ + "issue:43452|issue:45588", + "issue:43575|issue:43701", + "issue:29127|issue:35141", + "issue:43646|issue:44077", + "issue:45072|issue:45229", + "issue:43575|issue:44387", + "issue:44977|issue:45584", + "issue:45588|issue:45600", + "issue:33357|issue:38175", + "issue:43316|issue:43525", + "issue:44479|issue:45290", + "issue:44964|issue:45440", + "issue:43761|issue:44079", + "issue:43931|issue:45216", + "issue:42907|issue:45698", + "issue:33290|issue:34567", + "issue:34689|issue:43611", + "issue:39692|issue:45491", + "issue:44442|issue:44898", + "issue:43761|issue:44188", + "issue:43493|issue:43827", + "issue:43381|issue:43873", + "issue:44568|issue:45479", + "issue:43550|issue:44743", + "issue:43295|issue:45584", + "issue:43994|issue:44206" + ], + "prepared_review_unit_hash": "2aa24714c0edb08e816c0448714ca132825e9b25012224bbdd5cf84104deffcb", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 26, + "serialized_chars": 35627, + "estimated_input_tokens": 8907, + "estimated_eval_tokens": 18070 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The cluster is very heterogeneous: mostly unrelated bugs/regressions spanning loading, quantization, attention, tokenization, docs, and model-specific failures. None of the soft pairs look like the same underlying bug, so all are rejected.", + "confidence": 0.27, + "canonical_issue_reason": "No single canonical duplicate stands out; the issues do not converge on one concrete bug and instead cover many distinct failure modes.", + "canonical_pr_reason": null, + "best_issue_reason": "If forced to anchor the cluster, issue 45698 is the broadest representative of the recurring save/load/custom-module theme, but it is still not a true duplicate of the rest.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43452", + "right": "issue:45588", + "accept": false, + "reason": "Different code paths: tokenizer/gguf loading vs flash-attention sink handling." + }, + { + "left": "issue:43575", + "right": "issue:43701", + "accept": false, + "reason": "Both are regressions, but one is tensor-parallel OOM and the other is a checkpoint key mismatch." + }, + { + "left": "issue:29127", + "right": "issue:35141", + "accept": false, + "reason": "Unrelated bugs: LayoutLMv3 error messaging vs embedding reinitialization after resize." + }, + { + "left": "issue:43646", + "right": "issue:44077", + "accept": false, + "reason": "Both mention initialization, but they describe different issues and model-specific constraints." + }, + { + "left": "issue:45072", + "right": "issue:45229", + "accept": false, + "reason": "Bfloat16 dtype mismatch in CI vs multi-GPU CUDA OOM are different failure modes." + }, + { + "left": "issue:43575", + "right": "issue:44387", + "accept": false, + "reason": "Both involve memory pressure, but the causes are different: tp load OOM vs int4 reserved-memory growth." + }, + { + "left": "issue:44977", + "right": "issue:45584", + "accept": false, + "reason": "Different subsystems and symptoms: flash-attention generation vs empty-transcription chat-template failure." + }, + { + "left": "issue:45588", + "right": "issue:45600", + "accept": false, + "reason": "Flash-attention sink handling and auto_mappings stale config references are unrelated." + }, + { + "left": "issue:33357", + "right": "issue:38175", + "accept": false, + "reason": "Different models and behaviors: MacOS bus error vs zero-probability output." + }, + { + "left": "issue:43316", + "right": "issue:43525", + "accept": false, + "reason": "Both are API discrepancies, but for different configs and different missing fields." + }, + { + "left": "issue:44479", + "right": "issue:45290", + "accept": false, + "reason": "Video-input regression in Qwen VL models is unrelated to assistant tool-call handling in chat templates." + }, + { + "left": "issue:44964", + "right": "issue:45440", + "accept": false, + "reason": "Model loading failure for Phi-4 multimodal vs DeepSeekV3 implementation divergence are distinct bugs." + }, + { + "left": "issue:43761", + "right": "issue:44079", + "accept": false, + "reason": "Hidden-states omission in CLIPVisionModel is not the same as ModelOutput key assignment logic." + }, + { + "left": "issue:43931", + "right": "issue:45216", + "accept": false, + "reason": "Both are Qwen3-family load/save issues, but one is a shape-mismatch load error and the other is a bad saved checkpoint." + }, + { + "left": "issue:42907", + "right": "issue:45698", + "accept": false, + "reason": "Save failure for dequantized models vs wrong custom-module loading after save_pretrained are different save/load bugs." + }, + { + "left": "issue:33290", + "right": "issue:34567", + "accept": false, + "reason": "OOM in deepspeed/adafactor is unrelated to TrainerState token counting." + }, + { + "left": "issue:34689", + "right": "issue:43611", + "accept": false, + "reason": "Both are loading regressions, but they involve different root causes and model-loading paths." + }, + { + "left": "issue:39692", + "right": "issue:45491", + "accept": false, + "reason": "Documentation example errors are unrelated to Gemma3 NaN embeddings in mixed-length batching." + }, + { + "left": "issue:44442", + "right": "issue:44898", + "accept": false, + "reason": "Tokenizer loading failure and Perceiver resolution failure are separate model-specific bugs." + }, + { + "left": "issue:43761", + "right": "issue:44188", + "accept": false, + "reason": "CLIP hidden_states regression is unrelated to attention-kernel branching under torch.compile." + }, + { + "left": "issue:43493", + "right": "issue:43827", + "accept": false, + "reason": "SigLIP2 implementation mismatch and stale docs referencing pipeline() are unrelated." + }, + { + "left": "issue:43381", + "right": "issue:43873", + "accept": false, + "reason": "Gradient checkpointing in eval mode and quantization offloading are different behaviors." + }, + { + "left": "issue:44568", + "right": "issue:45479", + "accept": false, + "reason": "Tokenizer special-token behavior is unrelated to the zero-loss classification bug." + }, + { + "left": "issue:43550", + "right": "issue:44743", + "accept": false, + "reason": "torch.compile/SDPA failure in Bamba is not the same as Qwen recurrent-state reset." + }, + { + "left": "issue:43295", + "right": "issue:45584", + "accept": false, + "reason": "Processor/tokenizer regression in save/load is unrelated to Whisper empty-transcription generation failure." + }, + { + "left": "issue:43994", + "right": "issue:44206", + "accept": false, + "reason": "SigLIP2 inference quality issue and LasrFeatureExtractor crash are different defects." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 26, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:25251", + "issue:28282", + "issue:29127", + "issue:32090", + "issue:33290", + "issue:33357", + "issue:34689", + "issue:36010", + "issue:41628", + "issue:41950", + "issue:42175", + "issue:43232", + "issue:43295", + "issue:43316", + "issue:43381", + "issue:43388", + "issue:43404", + "issue:43441", + "issue:43475", + "issue:43540", + "issue:43550", + "issue:43606", + "issue:43643", + "issue:43701", + "issue:43720", + "issue:43742", + "issue:43825", + "issue:43856", + "issue:43873", + "issue:44077", + "issue:44162", + "issue:44190", + "issue:44442", + "issue:44479", + "issue:44496", + "issue:44514", + "issue:44792", + "issue:44855", + "issue:44871", + "issue:44877", + "issue:44913", + "issue:45127", + "issue:45137", + "issue:45200", + "issue:45372", + "issue:45584", + "issue:45656", + "issue:45759" + ], + "soft_pairs": [ + "issue:43316|issue:45759", + "issue:41628|issue:45372", + "issue:43643|issue:44913", + "issue:43720|issue:44442", + "issue:32090|issue:33290", + "issue:44190|issue:44496", + "issue:43873|issue:44871", + "issue:34689|issue:36010", + "issue:43381|issue:43856", + "issue:44077|issue:44479", + "issue:44855|issue:45137", + "issue:33290|issue:45656", + "issue:43540|issue:44514", + "issue:43316|issue:44877", + "issue:29127|issue:33357", + "issue:43381|issue:43742", + "issue:42175|issue:43316", + "issue:43404|issue:43441", + "issue:33290|issue:34689", + "issue:25251|issue:28282", + "issue:44162|issue:44792", + "issue:41950|issue:43475", + "issue:43232|issue:43388", + "issue:43295|issue:43825", + "issue:44871|issue:45584", + "issue:45127|issue:45200", + "issue:43550|issue:43701", + "issue:43606|issue:43701", + "issue:43295|issue:44871" + ], + "prepared_review_unit_hash": "7c89439a7dd20679f763df775b7ab3684c00bc9edd50dbfeea8717b8992341d9", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 29, + "serialized_chars": 36559, + "estimated_input_tokens": 9140, + "estimated_eval_tokens": 18536 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This set is heterogeneous: it spans unrelated bugs in config loading, generation, multimodal processors, deepspeed, offloading, and training scripts. I don\u2019t see evidence that any soft pair is the same underlying issue, so all candidate duplicate links should be rejected.", + "confidence": 0.96, + "canonical_issue_reason": "No single canonical issue stands out because the items are not one bug cluster; if a representative issue must be chosen, issue:41950 is the clearest open, concrete, and actionable report.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:41950 is the best representative issue only in the sense that it is open, specific, and has a clear failure mode; however, it is not a duplicate of the rest of the set.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43316", + "right": "issue:45759", + "accept": false, + "reason": "Both involve config-related multimodal/model behavior, but they describe different bugs: one is a Gemma3TextConfig API mismatch, the other is from_config not unwrapping text_config for Qwen composite configs." + }, + { + "left": "issue:41628", + "right": "issue:45372", + "accept": false, + "reason": "Both are import errors, but they stem from different missing symbols and different code paths (AutoImageProcessor vs ReasoningEffort/Gemma 4 processor loading)." + }, + { + "left": "issue:43643", + "right": "issue:44913", + "accept": false, + "reason": "Both concern config fields surviving load/reload, but the reported failures are different: missing fields with trust_remote_code vs rotary_pct resetting on GPTNeoXConfig reload." + }, + { + "left": "issue:43720", + "right": "issue:44442", + "accept": false, + "reason": "Unrelated failures: packed BitNet weights during accelerate loading vs AutoTokenizer failing to load a specific tokenizer." + }, + { + "left": "issue:32090", + "right": "issue:33290", + "accept": false, + "reason": "Both mention Trainer/DeepSpeed, but one is a NoneType broadcast error and the other is an optimizer OOM; these are not the same bug." + }, + { + "left": "issue:44190", + "right": "issue:44496", + "accept": false, + "reason": "Different loading failures: local dataset handling in a no-trainer script vs an unrecognized model config missing model_type." + }, + { + "left": "issue:43873", + "right": "issue:44871", + "accept": false, + "reason": "One is a quantization/offloading memory behavior issue; the other is an eos_token_id configuration mismatch in Gemma-3." + }, + { + "left": "issue:34689", + "right": "issue:36010", + "accept": false, + "reason": "Both are model-loading/import regressions, but they are distinct problems affecting different APIs and code paths." + }, + { + "left": "issue:43381", + "right": "issue:43856", + "accept": false, + "reason": "Gradient checkpointing in eval mode and Qwen3 MoE memory usage are separate training behaviors, not the same defect." + }, + { + "left": "issue:44077", + "right": "issue:44479", + "accept": false, + "reason": "Different model/runtime failures: optional post_init handling for patchtsmixer vs video-input regression for Qwen VL models." + }, + { + "left": "issue:44855", + "right": "issue:45137", + "accept": false, + "reason": "Python 3.13 decorator/comment parsing failure in DebertaV2Model is unrelated to the DeepSpeed ZeRO3 deque IndexError." + }, + { + "left": "issue:33290", + "right": "issue:45656", + "accept": false, + "reason": "Both involve DeepSpeed optimizer behavior, but one reports OOM with Adafactor while the other reports a duplicated optimizer step call." + }, + { + "left": "issue:43540", + "right": "issue:44514", + "accept": false, + "reason": "Both are multimodal/video-related, but one is a Qwen3OmniMoe video processing ValueError and the other is a batched apply_chat_template crash in Qwen2_5_VLProcessor." + }, + { + "left": "issue:43316", + "right": "issue:44877", + "accept": false, + "reason": "Both involve config objects, but they are different bugs: API inconsistency in Gemma3TextConfig vs strict config blocking granite_speech loading." + }, + { + "left": "issue:29127", + "right": "issue:33357", + "accept": false, + "reason": "Completely unrelated issues: LayoutLMv3 error messaging vs a MacOS bus error on a CLIP model." + }, + { + "left": "issue:43381", + "right": "issue:43742", + "accept": false, + "reason": "No overlap beyond generic training/loading context; gradient checkpointing eval behavior is unrelated to MobileLLM key errors." + }, + { + "left": "issue:42175", + "right": "issue:43316", + "accept": false, + "reason": "Install/backend packaging for TensorFlow is unrelated to a Gemma3TextConfig API discrepancy." + }, + { + "left": "issue:43404", + "right": "issue:43441", + "accept": false, + "reason": "Both touch Mistral-family models, but one is tied weights in model conversion/loading and the other is a FlashAttention runtime failure; not the same code-path bug." + }, + { + "left": "issue:33290", + "right": "issue:34689", + "accept": false, + "reason": "Different failures in different subsystems: DeepSpeed Adafactor OOM vs Llama 3.2 Vision model loading regression." + }, + { + "left": "issue:25251", + "right": "issue:28282", + "accept": false, + "reason": "Pipeline top_k nesting behavior is unrelated to an AutoModel PyTorch import error." + }, + { + "left": "issue:44162", + "right": "issue:44792", + "accept": false, + "reason": "ESM2 breakage and a janus test failure are different model-specific issues." + }, + { + "left": "issue:41950", + "right": "issue:43475", + "accept": false, + "reason": "Both are vision/video model reports, but one is a pipeline processor-selection bug and the other is a missing attribute in Sam3VisionEncoderOutput." + }, + { + "left": "issue:43232", + "right": "issue:43388", + "accept": false, + "reason": "Generation kwargs after sync_gpus and metric gathering label truncation are unrelated problems." + }, + { + "left": "issue:43295", + "right": "issue:43825", + "accept": false, + "reason": "Processor/tokenizer regression and pipeline error-message wording are different bugs with different symptoms and fixes." + }, + { + "left": "issue:44871", + "right": "issue:45584", + "accept": false, + "reason": "Both involve special tokens, but one is a config inconsistency in Gemma-3 and the other is Whisper failing on empty transcription after align_special_tokens." + }, + { + "left": "issue:45127", + "right": "issue:45200", + "accept": false, + "reason": "LoRA merge with extended vocab causing collapse is unrelated to Gemma 4 requiring mm_token_type_ids defaults." + }, + { + "left": "issue:43550", + "right": "issue:43701", + "accept": false, + "reason": "A torch.compile/SDPA model failure is unrelated to a resume_from_checkpoint key mismatch." + }, + { + "left": "issue:43606", + "right": "issue:43701", + "accept": false, + "reason": "CPU offload device mismatch for bark-small is unrelated to the checkpoint key mismatch issue." + }, + { + "left": "issue:43295", + "right": "issue:44871", + "accept": false, + "reason": "These are separate regressions: processor.tokenizer access in custom code vs Gemma-3 eos_token_id inconsistency." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 27, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:33290", + "issue:35141", + "issue:36331", + "issue:41720", + "issue:42617", + "issue:43065", + "issue:43335", + "issue:43452", + "issue:43575", + "issue:43646", + "issue:43673", + "issue:43742", + "issue:43761", + "issue:43827", + "issue:43873", + "issue:43906", + "issue:43931", + "issue:43994", + "issue:44060", + "issue:44077", + "issue:44190", + "issue:44242", + "issue:44464", + "issue:44479", + "issue:44496", + "issue:44521", + "issue:44534", + "issue:44610", + "issue:44617", + "issue:44734", + "issue:44743", + "issue:44792", + "issue:44855", + "issue:44871", + "issue:44913", + "issue:44936", + "issue:44964", + "issue:44977", + "issue:45127", + "issue:45245", + "issue:45290", + "issue:45357", + "issue:45381", + "issue:45584", + "issue:45588", + "issue:45600", + "issue:45698" + ], + "soft_pairs": [ + "issue:43673|issue:44743", + "issue:43761|issue:44479", + "issue:33290|issue:35141", + "issue:33290|issue:36331", + "issue:43065|issue:44617", + "issue:43452|issue:44496", + "issue:43575|issue:43742", + "issue:44936|issue:45698", + "issue:43646|issue:44913", + "issue:42617|issue:43575", + "issue:43761|issue:44242", + "issue:43575|issue:43873", + "issue:44077|issue:44534", + "issue:43931|issue:44060", + "issue:44479|issue:44734", + "issue:44617|issue:45584", + "issue:44464|issue:44871", + "issue:43065|issue:44610", + "issue:43827|issue:43994", + "issue:43906|issue:45290", + "issue:44792|issue:44977", + "issue:44534|issue:44913", + "issue:41720|issue:45245", + "issue:44190|issue:44964", + "issue:44792|issue:45600", + "issue:44855|issue:45588", + "issue:44521|issue:45381", + "issue:45127|issue:45357", + "issue:43335|issue:44855" + ], + "prepared_review_unit_hash": "8424c1f4cb75b643a62837b83d6e81d61b5e4e1177be761e74dafa34ad37bdb4", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 29, + "serialized_chars": 35843, + "estimated_input_tokens": 8961, + "estimated_eval_tokens": 18178 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is a heterogeneous mix of unrelated bugs/regressions (OOMs, config/load failures, generation/cache issues, multimodal model regressions, and docs/CI problems). I do not see a meaningful duplicate set or a single canonical issue for the group.", + "confidence": 0.97, + "canonical_issue_reason": "No single canonical issue: the items span distinct code paths and failure modes, so the cluster looks like a false-positive mix rather than one underlying bug.", + "canonical_pr_reason": null, + "best_issue_reason": "No global best issue is suitable because there is no coherent duplicate family to represent; the issues are too diverse in subsystem and root cause.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43673", + "right": "issue:44743", + "accept": false, + "reason": "Both involve cache/state behavior, but one is a v5 chunked-prefill cache missing bug and the other is Qwen3.5 recurrent-state reset with seq_len>1; different model paths and causes." + }, + { + "left": "issue:43761", + "right": "issue:44479", + "accept": false, + "reason": "Both are multimodal regressions, but CLIPVision hidden_states handling and Qwen video input position handling are different bugs in different models." + }, + { + "left": "issue:33290", + "right": "issue:35141", + "accept": false, + "reason": "OOM with Adafactor/DeepSpeed is unrelated to token embedding reinitialization after resize and post_init." + }, + { + "left": "issue:33290", + "right": "issue:36331", + "accept": false, + "reason": "Memory blowup in training infrastructure is not the same as a CustomTrainer compute_loss signature mismatch." + }, + { + "left": "issue:43065", + "right": "issue:44617", + "accept": false, + "reason": "Same model family, but one is a dummy Conv2d implementation issue and the other is a CUDA OOM during video inference; not the same fix." + }, + { + "left": "issue:43452", + "right": "issue:44496", + "accept": false, + "reason": "Both are model-loading failures, but one is gguf/AutoTokenizer-AutoModel handling and the other is missing model_type in config.json." + }, + { + "left": "issue:43575", + "right": "issue:43742", + "accept": false, + "reason": "Both fail on loading, but one is TP-related OOM for Qwen2-57B and the other is a config/key error for MobileLLM." + }, + { + "left": "issue:44936", + "right": "issue:45698", + "accept": false, + "reason": "trainer.evaluate after train and wrong custom module reload after save_pretrained are separate lifecycle bugs." + }, + { + "left": "issue:43646", + "right": "issue:44913", + "accept": false, + "reason": "Transformers 5 custom model init breakage is unrelated to GPTNeoX rotary_pct not persisting on reload." + }, + { + "left": "issue:42617", + "right": "issue:43575", + "accept": false, + "reason": "3d_parallel.py execution failure is unrelated to Qwen2 TP OOM during model loading." + }, + { + "left": "issue:43761", + "right": "issue:44242", + "accept": false, + "reason": "CLIPVision hidden_states regression and missing load-balancing loss are different model components and failure modes." + }, + { + "left": "issue:43575", + "right": "issue:43873", + "accept": false, + "reason": "Both mention loading/memory, but TP OOM and quantization offloading not working are distinct problems." + }, + { + "left": "issue:44077", + "right": "issue:44534", + "accept": false, + "reason": "Optional post_init restriction in patchtsmixer is unrelated to junk-filled non-persistent buffers in v5." + }, + { + "left": "issue:43931", + "right": "issue:44060", + "accept": false, + "reason": "Weight shape mismatch in Qwen3-VL and incorrect tied-weights warning in Qwen3-Next are different bugs." + }, + { + "left": "issue:44479", + "right": "issue:44734", + "accept": false, + "reason": "Video-input regression in Qwen multimodal models is unrelated to KV-cache continuation tensor indexing in transformers serve." + }, + { + "left": "issue:44617", + "right": "issue:45584", + "accept": false, + "reason": "Sam3Video CUDA OOM and Whisper empty-transcription failure after align_special_tokens are unrelated." + }, + { + "left": "issue:44464", + "right": "issue:44871", + "accept": false, + "reason": "Chunked generation inconsistency with compiled forward is unrelated to Gemma eos_token_id config mismatch." + }, + { + "left": "issue:43065", + "right": "issue:44610", + "accept": false, + "reason": "Sam3PixelDecoder dummy layer issue is unrelated to OmDet-Turbo processor/model input-size mismatch." + }, + { + "left": "issue:43827", + "right": "issue:43994", + "accept": false, + "reason": "Docs still referencing pipeline() and Siglip2 nonsensical outputs are different issues." + }, + { + "left": "issue:43906", + "right": "issue:45290", + "accept": false, + "reason": "A reproduction of issue 38071 is not the same as apply_chat_template crashing on assistant tool-call messages." + }, + { + "left": "issue:44792", + "right": "issue:44977", + "accept": false, + "reason": "Janus image-generation test failure and Qwen3.5 flash-attention generation bug are unrelated." + }, + { + "left": "issue:44534", + "right": "issue:44913", + "accept": false, + "reason": "Non-persistent buffer junk in v5 and GPTNeoX rotary_pct reload default are distinct configuration/runtime bugs." + }, + { + "left": "issue:41720", + "right": "issue:45245", + "accept": false, + "reason": "Qwen3 auto device mapping cudaErrorAssert is unrelated to the 2^24 category limit RuntimeError." + }, + { + "left": "issue:44190", + "right": "issue:44964", + "accept": false, + "reason": "Local dataset loading in no-trainer image classification and Phi-4 multimodal model loading are different paths." + }, + { + "left": "issue:44792", + "right": "issue:45600", + "accept": false, + "reason": "Janus test failure is unrelated to stale auto_mappings references for removed Sam3LiteText configs." + }, + { + "left": "issue:44855", + "right": "issue:45588", + "accept": false, + "reason": "Python 3.13 IndentationError in DebertaV2 import is unrelated to flash_attention's s_aux=None crash." + }, + { + "left": "issue:44521", + "right": "issue:45381", + "accept": false, + "reason": "Both are multimodal/chat-template issues, but all-zero assistant masks and wrong vision_position_ids are different concrete bugs." + }, + { + "left": "issue:45127", + "right": "issue:45357", + "accept": false, + "reason": "LoRA merge collapse with extended vocab is unrelated to incorrect visual encoder keys saved by Qwen3.5." + }, + { + "left": "issue:43335", + "right": "issue:44855", + "accept": false, + "reason": "SwitchTransformers sparse-layer creation and a Python 3.13 DebertaV2 import parse failure are unrelated." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 28, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:36246", + "issue:38175", + "issue:38617", + "issue:41628", + "issue:43065", + "issue:43116", + "issue:43295", + "issue:43335", + "issue:43404", + "issue:43479", + "issue:43540", + "issue:43550", + "issue:43643", + "issue:43644", + "issue:43650", + "issue:43761", + "issue:43976", + "issue:44077", + "issue:44112", + "issue:44186", + "issue:44190", + "issue:44220", + "issue:44246", + "issue:44464", + "issue:44466", + "issue:44496", + "issue:44561", + "issue:44683", + "issue:44734", + "issue:44743", + "issue:44855", + "issue:44877", + "issue:44913", + "issue:44933", + "issue:44938", + "issue:45003", + "issue:45216", + "issue:45245", + "issue:45276", + "issue:45362", + "issue:45375", + "issue:45406", + "issue:45446", + "issue:45542", + "issue:45563", + "issue:45584", + "issue:45600", + "issue:45632" + ], + "soft_pairs": [ + "issue:44938|issue:45542", + "issue:43976|issue:44855", + "issue:44077|issue:44877", + "issue:43644|issue:44913", + "issue:43550|issue:44190", + "issue:43761|issue:44743", + "issue:43540|issue:44734", + "issue:45276|issue:45584", + "issue:43650|issue:44220", + "issue:44112|issue:45600", + "issue:43116|issue:44190", + "issue:43295|issue:45276", + "issue:43335|issue:44190", + "issue:45245|issue:45584", + "issue:36246|issue:41628", + "issue:43404|issue:45216", + "issue:43479|issue:45375", + "issue:44933|issue:45362", + "issue:44734|issue:45632", + "issue:44464|issue:45563", + "issue:44186|issue:44734", + "issue:38175|issue:43065", + "issue:44561|issue:45632", + "issue:44683|issue:45446", + "issue:44466|issue:44913", + "issue:44246|issue:45003", + "issue:36246|issue:38617", + "issue:45406|issue:45632", + "issue:43643|issue:44496", + "issue:45563|issue:45584" + ], + "prepared_review_unit_hash": "5b0a488241905aeea3224192b954b039470a9e9718dea880e65c69d4e9af6f38", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 30, + "serialized_chars": 36765, + "estimated_input_tokens": 9192, + "estimated_eval_tokens": 18640 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This set is highly heterogeneous: the soft pairs mostly share only broad themes like imports, configs, or regressions, but they do not describe the same concrete bug/change. I would not merge any of the proposed pairs.", + "confidence": 0.89, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": "No issue is a strong duplicate hub for this set. If a single representative is needed, issue 44246 is the broadest user-facing runtime/reporting bug with active discussion, but it is not a true stand-in for the others.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44938", + "right": "issue:45542", + "accept": false, + "reason": "Different failures: Python 3.14 import/load problem vs missing TensorFlow backend when only tensorboard is installed." + }, + { + "left": "issue:43976", + "right": "issue:44855", + "accept": false, + "reason": "Both are Python-version-related import errors, but they involve different models, symptoms, and root causes." + }, + { + "left": "issue:44077", + "right": "issue:44877", + "accept": false, + "reason": "Both touch config validation, but one is about optional post_init on patchtsmixer and the other about loading granite_speech config." + }, + { + "left": "issue:43644", + "right": "issue:44913", + "accept": false, + "reason": "One is a non-persistent buffer initialization regression; the other is a config field not persisting on reload." + }, + { + "left": "issue:43550", + "right": "issue:44190", + "accept": false, + "reason": "Completely different areas: SDPA/torch.compile on a model vs example script dataset loading." + }, + { + "left": "issue:43761", + "right": "issue:44743", + "accept": false, + "reason": "Different models and different bugs: CLIP hidden_states omission vs recurrent state reset with cache." + }, + { + "left": "issue:43540", + "right": "issue:44734", + "accept": false, + "reason": "Video input validation in Qwen3OmniMoe is unrelated to serve KV-cache continuation indexing." + }, + { + "left": "issue:45276", + "right": "issue:45584", + "accept": false, + "reason": "Token embedding resize wiring vs Whisper empty-transcription generation failure are distinct issues." + }, + { + "left": "issue:43650", + "right": "issue:44220", + "accept": false, + "reason": "The titles are unrelated; one is not a real technical bug description and the other is a feature-specific audio feature extraction issue." + }, + { + "left": "issue:44112", + "right": "issue:45600", + "accept": false, + "reason": "Both mention CI, but they are different failures in different code paths and different models." + }, + { + "left": "issue:43116", + "right": "issue:44190", + "accept": false, + "reason": "Example-script multi-label classification output bug vs local dataset loading failure are not the same problem." + }, + { + "left": "issue:43295", + "right": "issue:45276", + "accept": false, + "reason": "Processor/tokenizer regression is unrelated to token embedding resizing behavior." + }, + { + "left": "issue:43335", + "right": "issue:44190", + "accept": false, + "reason": "SwitchTransformers config layer construction bug is unrelated to the image-classification dataset loader issue." + }, + { + "left": "issue:45245", + "right": "issue:45584", + "accept": false, + "reason": "Generic category-count runtime error is different from Whisper post-processing/empty transcription handling." + }, + { + "left": "issue:36246", + "right": "issue:41628", + "accept": false, + "reason": "Both are import errors, but they expose different missing symbols from different modules and likely different regressions." + }, + { + "left": "issue:43404", + "right": "issue:45216", + "accept": false, + "reason": "Weight tying in Mistral3 is a different bug from incorrect Qwen3.5 checkpoint saving." + }, + { + "left": "issue:43479", + "right": "issue:45375", + "accept": false, + "reason": "Both involve multimodal config fields, but they affect different models and different missing/defaulted fields." + }, + { + "left": "issue:44933", + "right": "issue:45362", + "accept": false, + "reason": "A missing image_utils import is unrelated to Qwen3.5 chat crashes." + }, + { + "left": "issue:44734", + "right": "issue:45632", + "accept": false, + "reason": "Serve KV-cache indexing crash and trust_remote_code cache-path collision are unrelated." + }, + { + "left": "issue:44464", + "right": "issue:45563", + "accept": false, + "reason": "Compiled forward generation inconsistency is not the same as a stale warning about num_return_sequences." + }, + { + "left": "issue:44186", + "right": "issue:44734", + "accept": false, + "reason": "Tokenizer padding/NER crash is unrelated to serve-side KV-cache tensor indexing." + }, + { + "left": "issue:38175", + "right": "issue:43065", + "accept": false, + "reason": "Zero probabilities in SigLIP2 and dummy Conv2d in Sam3PixelDecoder are entirely different model bugs." + }, + { + "left": "issue:44561", + "right": "issue:45632", + "accept": false, + "reason": "Removal of is_torch_fx_available affects remote-code imports broadly; cache-path collisions are a separate trust_remote_code bug." + }, + { + "left": "issue:44683", + "right": "issue:45446", + "accept": false, + "reason": "Torch >= 2.9 flex_attention failure and incorrect version check for AuxRequest import are different flex_attention issues." + }, + { + "left": "issue:44466", + "right": "issue:44913", + "accept": false, + "reason": "Serialization of tied lm_head weights is unrelated to GPTNeoX rotary_pct reload behavior." + }, + { + "left": "issue:44246", + "right": "issue:45003", + "accept": false, + "reason": "Import slowdowns and unsafe sys.modules access may be related in spirit, but they are not the same concrete bug." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 29, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:41950", + "issue:42175", + "issue:42907", + "issue:43295", + "issue:43299", + "issue:43381", + "issue:43388", + "issue:43441", + "issue:43479", + "issue:43525", + "issue:43526", + "issue:43531", + "issue:43577", + "issue:43673", + "issue:43688", + "issue:43824", + "issue:43827", + "issue:43854", + "issue:43901", + "issue:43976", + "issue:44016", + "issue:44062", + "issue:44075", + "issue:44112", + "issue:44220", + "issue:44242", + "issue:44279", + "issue:44336", + "issue:44351", + "issue:44368", + "issue:44496", + "issue:44683", + "issue:44933", + "issue:44964", + "issue:45020", + "issue:45042", + "issue:45200", + "issue:45216", + "issue:45278", + "issue:45335", + "issue:45357", + "issue:45405", + "issue:45542", + "issue:45584", + "issue:45593", + "issue:45600", + "issue:45759" + ], + "soft_pairs": [ + "issue:41950|issue:44016", + "issue:44964|issue:45600", + "issue:42907|issue:43441", + "issue:43388|issue:43688", + "issue:43295|issue:45335", + "issue:44112|issue:44964", + "issue:44496|issue:45042", + "issue:43854|issue:44964", + "issue:44016|issue:44220", + "issue:43673|issue:44336", + "issue:43299|issue:45216", + "issue:42175|issue:44279", + "issue:43531|issue:45600", + "issue:43526|issue:43901", + "issue:45335|issue:45584", + "issue:44075|issue:44368", + "issue:43381|issue:45593", + "issue:43299|issue:45357", + "issue:44496|issue:45759", + "issue:41950|issue:45020", + "issue:43901|issue:44496", + "issue:44242|issue:45200", + "issue:44016|issue:44062", + "issue:43577|issue:45216", + "issue:43976|issue:45405", + "issue:44351|issue:45278", + "issue:43824|issue:44933", + "issue:43388|issue:43827", + "issue:43479|issue:43525", + "issue:44683|issue:45542" + ], + "prepared_review_unit_hash": "cd2e966ad2fb706b9e7214bbaa23986451317590debb6c1e0c4cf1ed98d2c8b5", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 30, + "serialized_chars": 36062, + "estimated_input_tokens": 9016, + "estimated_eval_tokens": 18288 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The cluster is mostly a mix of unrelated Transformers regressions, import errors, and docs/test issues. None of the soft-pair candidates look like true duplicates.", + "confidence": 0.84, + "canonical_issue_reason": "issue:45020 is the broadest representative report here: a general regression affecting remote_code model loading across recent Transformers versions.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45020 is the best umbrella issue in this set because it captures the widest upstream regression surface.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:41950", + "right": "issue:44016", + "accept": false, + "reason": "Different problems: video pipeline processor lookup vs notebook syntax error." + }, + { + "left": "issue:44964", + "right": "issue:45600", + "accept": false, + "reason": "Different model-loading bug vs stale auto_mappings CI references." + }, + { + "left": "issue:42907", + "right": "issue:43441", + "accept": false, + "reason": "Different code paths: save/dequantization vs FlashAttention model loading." + }, + { + "left": "issue:43388", + "right": "issue:43688", + "accept": false, + "reason": "Metrics batching bug vs MoE auxiliary-loss normalization; unrelated." + }, + { + "left": "issue:43295", + "right": "issue:45335", + "accept": false, + "reason": "Processor/tokenizer regression vs t5gemma embedding resize bug." + }, + { + "left": "issue:44112", + "right": "issue:44964", + "accept": false, + "reason": "CI device override test issue vs Phi-4 multimodal loading failure." + }, + { + "left": "issue:44496", + "right": "issue:45042", + "accept": false, + "reason": "Unrecognized model/config issue vs PIL image processor torchvision dependency bug." + }, + { + "left": "issue:43854", + "right": "issue:44964", + "accept": false, + "reason": "Both are load failures, but for different models and different underlying causes." + }, + { + "left": "issue:44016", + "right": "issue:44220", + "accept": false, + "reason": "Notebook syntax error is unrelated to audio feature extraction failure." + }, + { + "left": "issue:43673", + "right": "issue:44336", + "accept": false, + "reason": "Cache/memory issue vs ANSI terminal reporting bug; unrelated." + }, + { + "left": "issue:43299", + "right": "issue:45216", + "accept": false, + "reason": "Qwen loading regression vs Qwen save_pretrained regression; different concrete bugs." + }, + { + "left": "issue:42175", + "right": "issue:44279", + "accept": false, + "reason": "Backend dependency packaging issue vs generic dependency problem; too broad to merge." + }, + { + "left": "issue:43531", + "right": "issue:45600", + "accept": false, + "reason": "Qwen3-MoE sliding window bug vs removed Sam3LiteText config references." + }, + { + "left": "issue:43526", + "right": "issue:43901", + "accept": false, + "reason": "ImageProcessorFast label reduction bug vs documentation mismatch." + }, + { + "left": "issue:45335", + "right": "issue:45584", + "accept": false, + "reason": "Embedding resize bug vs Whisper empty-transcription generation failure." + }, + { + "left": "issue:44075", + "right": "issue:44368", + "accept": false, + "reason": "Optimizer argument handling vs LoRA warning about tie_word_embeddings." + }, + { + "left": "issue:43381", + "right": "issue:45593", + "accept": false, + "reason": "Gradient checkpointing eval-mode bug vs D-FINE auxiliary-loss behavior." + }, + { + "left": "issue:43299", + "right": "issue:45357", + "accept": false, + "reason": "Loading failure vs save_pretrained checkpoint-key regression; not same bug." + }, + { + "left": "issue:44496", + "right": "issue:45759", + "accept": false, + "reason": "Unrecognized model registration vs from_config text_config unwrapping; related area but different defect." + }, + { + "left": "issue:41950", + "right": "issue:45020", + "accept": false, + "reason": "General version-regression theme only; different subsystems and concrete failures." + }, + { + "left": "issue:43901", + "right": "issue:44496", + "accept": false, + "reason": "Docs behavior note vs model-loading config error; unrelated." + }, + { + "left": "issue:44242", + "right": "issue:45200", + "accept": false, + "reason": "MoE load-balancing loss bug vs Gemma 4 mm_token_type_ids defaulting issue." + }, + { + "left": "issue:44016", + "right": "issue:44062", + "accept": false, + "reason": "Notebook syntax error vs AddedToken keyword-argument conflict." + }, + { + "left": "issue:43577", + "right": "issue:45216", + "accept": false, + "reason": "BLIP2 dtype propagation vs Qwen save_pretrained checkpoint regression." + }, + { + "left": "issue:43976", + "right": "issue:45405", + "accept": false, + "reason": "Python support breakage vs unreleased PEFT version bump." + }, + { + "left": "issue:44351", + "right": "issue:45278", + "accept": false, + "reason": "Specific HybridCache import failure vs broad many-import-errors report; not the same bug." + }, + { + "left": "issue:43824", + "right": "issue:44933", + "accept": false, + "reason": "Different missing imports from different modules and causes." + }, + { + "left": "issue:43388", + "right": "issue:43827", + "accept": false, + "reason": "Training metric batching bug vs docs still referencing removed pipeline API." + }, + { + "left": "issue:43479", + "right": "issue:43525", + "accept": false, + "reason": "Different config initialization issue vs missing pad_token_id attribute." + }, + { + "left": "issue:44683", + "right": "issue:45542", + "accept": false, + "reason": "Compiled flex_attention torch-version bug vs TensorBoard/TensorFlow packaging issue." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 30, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:32090", + "issue:35141", + "issue:38175", + "issue:39692", + "issue:41950", + "issue:42994", + "issue:43065", + "issue:43122", + "issue:43295", + "issue:43299", + "issue:43317", + "issue:43441", + "issue:43450", + "issue:43650", + "issue:43653", + "issue:43723", + "issue:43784", + "issue:43854", + "issue:43901", + "issue:43976", + "issue:43994", + "issue:44162", + "issue:44261", + "issue:44263", + "issue:44279", + "issue:44297", + "issue:44361", + "issue:44462", + "issue:44466", + "issue:44521", + "issue:44617", + "issue:44623", + "issue:44671", + "issue:44683", + "issue:44749", + "issue:44779", + "issue:44898", + "issue:44987", + "issue:45081", + "issue:45276", + "issue:45290", + "issue:45325", + "issue:45397", + "issue:45405", + "issue:45542", + "issue:45593", + "issue:45600", + "issue:45712" + ], + "soft_pairs": [ + "issue:44462|issue:45081", + "issue:44279|issue:45405", + "issue:43299|issue:43854", + "issue:39692|issue:43901", + "issue:32090|issue:35141", + "issue:41950|issue:43650", + "issue:43295|issue:44779", + "issue:44297|issue:44462", + "issue:38175|issue:44617", + "issue:43976|issue:44279", + "issue:44462|issue:44779", + "issue:44521|issue:45325", + "issue:43976|issue:44987", + "issue:45600|issue:45712", + "issue:43450|issue:44279", + "issue:43065|issue:44898", + "issue:42994|issue:43317", + "issue:44466|issue:45276", + "issue:44261|issue:44263", + "issue:43441|issue:44683", + "issue:43650|issue:45290", + "issue:44623|issue:45397", + "issue:43723|issue:44162", + "issue:43784|issue:45542", + "issue:44623|issue:44749", + "issue:44671|issue:44779", + "issue:43653|issue:45593", + "issue:44987|issue:45542", + "issue:43901|issue:43994", + "issue:43122|issue:43650", + "issue:44361|issue:45593" + ], + "prepared_review_unit_hash": "50c5b93e896db6ce1eb9065d8b60b2ebe2ca3b4b652291773692f750d61fa29d", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 31, + "serialized_chars": 36962, + "estimated_input_tokens": 9241, + "estimated_eval_tokens": 18738 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This looks like a loose collection of Transformers regressions, with several tokenizer/model-loading and multimodal-processing complaints, but the proposed soft edges mostly share only broad subsystem overlap rather than the same concrete bug. No pair here is a safe duplicate merge.", + "confidence": 0.63, + "canonical_issue_reason": "issue:43723 is the broadest and most central tokenizer-loading regression in the set; it can anchor the nearby tokenizer/processor complaints better than the model-specific or unrelated issues.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43723 is the clearest representative issue: generic, regression-oriented, and closer in scope to several tokenizer-related reports than the others.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44462", + "right": "issue:45081", + "accept": false, + "reason": "Both concern tokenizer loading, but one is AutoTokenizer ignoring repository tokenizer.json while the other is a Mistral regex patch crash; different failure points and fixes." + }, + { + "left": "issue:44279", + "right": "issue:45405", + "accept": false, + "reason": "Unrelated topics: a dependency problem vs a bumped PEFT minimum version. No shared bug." + }, + { + "left": "issue:43299", + "right": "issue:43854", + "accept": false, + "reason": "Both are model-loading failures, but they target different models and likely different load-time issues." + }, + { + "left": "issue:39692", + "right": "issue:43901", + "accept": false, + "reason": "One is a SigLIP2 example/model mismatch and quantization issue; the other is a docs mismatch about return_all_scores. Not the same bug." + }, + { + "left": "issue:32090", + "right": "issue:35141", + "accept": false, + "reason": "Trainer gpu broadcast TypeError vs output embedding reinit after resizing embeddings. Different code paths and symptoms." + }, + { + "left": "issue:41950", + "right": "issue:43650", + "accept": false, + "reason": "Video-classification pipeline processor lookup vs a placeholder issue title; no meaningful overlap." + }, + { + "left": "issue:43295", + "right": "issue:44779", + "accept": false, + "reason": "Both mention tokenizer behavior, but one is a processor/tokenizer API regression and the other is Deepseek tokenization correctness; too model-specific to merge." + }, + { + "left": "issue:44297", + "right": "issue:44462", + "accept": false, + "reason": "Related to tokenizer save/load metadata, but one is tokenizer_class mismatch on save and the other is AutoTokenizer ignoring tokenizer.json on load." + }, + { + "left": "issue:38175", + "right": "issue:44617", + "accept": false, + "reason": "SigLIP2 zero probabilities and Sam3Video CUDA OOM are completely different issues." + }, + { + "left": "issue:43976", + "right": "issue:44279", + "accept": false, + "reason": "Python-version support breakage vs a generic dependency issue; not the same underlying bug." + }, + { + "left": "issue:44462", + "right": "issue:44779", + "accept": false, + "reason": "Both are tokenizer-related, but one is repository tokenizer.json being ignored and the other is incorrect Deepseek tokenization results; different regressions." + }, + { + "left": "issue:44521", + "right": "issue:45325", + "accept": false, + "reason": "Both touch multimodal chat/VL behavior, but assistant mask generation and temporal rope position IDs are separate problems." + }, + { + "left": "issue:43976", + "right": "issue:44987", + "accept": false, + "reason": "Environment/version incompatibility vs a specific model load failure; not duplicates." + }, + { + "left": "issue:45600", + "right": "issue:45712", + "accept": false, + "reason": "Both are CI/repo hygiene problems, but one is stale auto_mappings references and the other is leftover dummy classes. Different fixes." + }, + { + "left": "issue:43450", + "right": "issue:44279", + "accept": false, + "reason": "Video processor shape bug vs dependency issue; unrelated." + }, + { + "left": "issue:43065", + "right": "issue:44898", + "accept": false, + "reason": "Sam3PixelDecoder dummy layer vs Perceiver non-default-resolution failure. Different models and failures." + }, + { + "left": "issue:42994", + "right": "issue:43317", + "accept": false, + "reason": "Both are quantization/load related, but saving a quantized model and loading a dequantized model with device_map=auto are different paths." + }, + { + "left": "issue:44466", + "right": "issue:45276", + "accept": false, + "reason": "Both involve embeddings/tied weights, but one is device-dependent serialization and the other is resize_token_embeddings not updating Gemma4 embeddings/output heads." + }, + { + "left": "issue:44261", + "right": "issue:44263", + "accept": false, + "reason": "MLA rms_norm_eps precision concern vs torch.split return values in GlmMoeDsaIndexer. Different components." + }, + { + "left": "issue:43441", + "right": "issue:44683", + "accept": false, + "reason": "Attention backend failures, but one is Ministral-3 with FlashAttention in v5 RC and the other is compiled flex_attention on torch >= 2.9. Not the same bug." + }, + { + "left": "issue:43650", + "right": "issue:45290", + "accept": false, + "reason": "Placeholder title vs a concrete chat-template crash with tool calls. No duplicate signal." + }, + { + "left": "issue:44623", + "right": "issue:45397", + "accept": false, + "reason": "Processor save_pretrained missing files vs Gemma-4 zero3 from_pretrained loading. Related area, different issue." + }, + { + "left": "issue:43723", + "right": "issue:44162", + "accept": false, + "reason": "Tokenizer-loading regression vs ESM2 model breakage; different model families and failure modes." + }, + { + "left": "issue:43784", + "right": "issue:45542", + "accept": false, + "reason": "Import-time NameError in sentence-transformers integration vs backend selection error when only tensorboard is installed. Different cause." + }, + { + "left": "issue:44623", + "right": "issue:44749", + "accept": false, + "reason": "Processor save_pretrained files missing vs a tokenizer performance regression after upgrade. Not duplicates." + }, + { + "left": "issue:44671", + "right": "issue:44779", + "accept": false, + "reason": "Both are v5 correctness regressions in NLP models, but CamemBERT masked LM output and Deepseek tokenization are separate concrete bugs." + }, + { + "left": "issue:43653", + "right": "issue:45593", + "accept": false, + "reason": "BigBirdTokenizer special-token registration bug vs D-FINE auxiliary-loss behavior. Unrelated." + }, + { + "left": "issue:44987", + "right": "issue:45542", + "accept": false, + "reason": "Model loading failure vs TensorFlow backend/installation issue. No overlap." + }, + { + "left": "issue:43901", + "right": "issue:43994", + "accept": false, + "reason": "Documentation/API mismatch vs SigLIP2 nonsensical outputs. Different scope." + }, + { + "left": "issue:43122", + "right": "issue:43650", + "accept": false, + "reason": "Tokenizer behavior changed across versions vs placeholder issue title. No duplicate." + }, + { + "left": "issue:44361", + "right": "issue:45593", + "accept": false, + "reason": "MLukeTokenizer AttributeError vs D-FINE denoising/aux-loss behavior. Completely different bugs." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 31, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:16998", + "issue:22355", + "issue:36246", + "issue:41628", + "issue:41950", + "issue:42491", + "issue:42757", + "issue:43262", + "issue:43295", + "issue:43299", + "issue:43352", + "issue:43475", + "issue:43525", + "issue:43673", + "issue:43761", + "issue:43824", + "issue:43881", + "issue:43901", + "issue:43976", + "issue:44016", + "issue:44188", + "issue:44220", + "issue:44242", + "issue:44246", + "issue:44291", + "issue:44295", + "issue:44297", + "issue:44336", + "issue:44351", + "issue:44373", + "issue:44462", + "issue:44485", + "issue:44496", + "issue:44521", + "issue:44561", + "issue:44589", + "issue:44617", + "issue:44661", + "issue:44857", + "issue:44871", + "issue:44964", + "issue:45250", + "issue:45356", + "issue:45362", + "issue:45397", + "issue:45542", + "issue:45593", + "issue:45759" + ], + "soft_pairs": [ + "issue:44016|issue:44291", + "issue:43262|issue:44521", + "issue:43475|issue:44016", + "issue:44297|issue:45397", + "issue:43673|issue:43761", + "issue:44188|issue:44336", + "issue:43525|issue:45759", + "issue:43525|issue:43881", + "issue:44589|issue:45362", + "issue:43295|issue:44617", + "issue:42491|issue:43299", + "issue:41628|issue:43976", + "issue:43901|issue:44246", + "issue:44336|issue:44373", + "issue:43352|issue:45250", + "issue:16998|issue:22355", + "issue:44242|issue:45593", + "issue:44295|issue:45397", + "issue:44661|issue:45542", + "issue:41950|issue:44220", + "issue:44246|issue:44496", + "issue:44462|issue:45356", + "issue:36246|issue:42757", + "issue:44336|issue:44485", + "issue:36246|issue:44351", + "issue:43673|issue:44561", + "issue:43824|issue:43976", + "issue:43976|issue:44964", + "issue:44246|issue:44871", + "issue:44242|issue:44857" + ], + "prepared_review_unit_hash": "ce782dec63463b2cbd8429aa5adcf7090407f73c3f25f89d868382405a98a69e", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 30, + "serialized_chars": 36525, + "estimated_input_tokens": 9132, + "estimated_eval_tokens": 18520 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "These soft-similarity pairs are mostly superficial; they share broad themes like imports, loading, tokenizers, or multimodal processing, but they do not appear to be the same underlying bug. I would reject all soft edges.", + "confidence": 0.96, + "canonical_issue_reason": "No single canonical issue: the set spans unrelated bugs (imports, model loading, tokenizer/pipeline regressions, docs), so no item cleanly represents the whole cluster.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 44521 is the strongest standalone representative because it is a current, well-documented multimodal regression with inbound references, but it still does not unify the rest of the set.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44016", + "right": "issue:44291", + "accept": false, + "reason": "Notebook syntax error vs. init_empty_weights TypeError; unrelated failures." + }, + { + "left": "issue:43262", + "right": "issue:44521", + "accept": false, + "reason": "Both touch apply_chat_template, but one is audio sampling-rate defaulting and the other is multimodal assistant masks; different bugs." + }, + { + "left": "issue:43475", + "right": "issue:44016", + "accept": false, + "reason": "SAM 3 vision encoder attribute error is unrelated to a notebook syntax error." + }, + { + "left": "issue:44297", + "right": "issue:45397", + "accept": false, + "reason": "Tokenizer save_pretrained metadata mismatch is unrelated to gemma-4 zero3 loading." + }, + { + "left": "issue:43673", + "right": "issue:43761", + "accept": false, + "reason": "Chunked prefill cache issue and CLIP hidden_states regression are different code paths." + }, + { + "left": "issue:44188", + "right": "issue:44336", + "accept": false, + "reason": "torch.compile attention-kernel divergence vs ANSI codes in loading_report; no overlap beyond general utilities." + }, + { + "left": "issue:43525", + "right": "issue:45759", + "accept": false, + "reason": "Both are model-loading/config issues, but Llama4 pad_token_id and Qwen multimodal text_config unwrapping are distinct failures." + }, + { + "left": "issue:43525", + "right": "issue:43881", + "accept": false, + "reason": "Different loading errors for different models; no clear shared concrete bug." + }, + { + "left": "issue:44589", + "right": "issue:45362", + "accept": false, + "reason": "Float8 storage deserialization and Qwen3.5 chat crash are unrelated." + }, + { + "left": "issue:43295", + "right": "issue:44617", + "accept": false, + "reason": "processor.tokenizer regression vs CUDA OOM; not the same underlying problem." + }, + { + "left": "issue:42491", + "right": "issue:43299", + "accept": false, + "reason": "Both involve Qwen/MoE loading regressions, but they affect different model families and failure modes." + }, + { + "left": "issue:41628", + "right": "issue:43976", + "accept": false, + "reason": "ImportError for AutoImageProcessor is unrelated to Python version support incompatibility." + }, + { + "left": "issue:43901", + "right": "issue:44246", + "accept": false, + "reason": "Docs mismatch for return_all_scores vs intermittent import slowness; different concerns." + }, + { + "left": "issue:44336", + "right": "issue:44373", + "accept": false, + "reason": "ANSI color handling in loading_report is unrelated to a docstring typo." + }, + { + "left": "issue:43352", + "right": "issue:45250", + "accept": false, + "reason": "Specific model lacks Flash Attention 2 support; the other is just a generic Flash Attention topic." + }, + { + "left": "issue:16998", + "right": "issue:22355", + "accept": false, + "reason": "model_max_length question and missing transformers.onnx module are unrelated." + }, + { + "left": "issue:44242", + "right": "issue:45593", + "accept": false, + "reason": "Both are loss-related when an option is disabled, but they affect different models and mechanisms." + }, + { + "left": "issue:44295", + "right": "issue:45397", + "accept": false, + "reason": "position_ids buffer access error vs zero3 from_pretrained issue; no shared code path." + }, + { + "left": "issue:44661", + "right": "issue:45542", + "accept": false, + "reason": "add-new-model-like tokenizer mapping bug is unrelated to the TensorFlow backend import error." + }, + { + "left": "issue:41950", + "right": "issue:44220", + "accept": false, + "reason": "Video-classification processor lookup and _torch_extract_fbank_features() are different pipeline components." + }, + { + "left": "issue:44246", + "right": "issue:44496", + "accept": false, + "reason": "Slow import is unrelated to unrecognized model config loading." + }, + { + "left": "issue:44462", + "right": "issue:45356", + "accept": false, + "reason": "AutoTokenizer ignoring tokenizer.json and Kimi-K2.5 codec regression are different tokenizer bugs." + }, + { + "left": "issue:36246", + "right": "issue:42757", + "accept": false, + "reason": "Different missing imports from different packages; same symptom class, not the same bug." + }, + { + "left": "issue:44336", + "right": "issue:44485", + "accept": false, + "reason": "ANSI-output bug and GLM-5 RoPE discussion are unrelated." + }, + { + "left": "issue:36246", + "right": "issue:44351", + "accept": false, + "reason": "Both are import errors, but for different missing symbols and modules." + }, + { + "left": "issue:43673", + "right": "issue:44561", + "accept": false, + "reason": "cache-related chunked prefill regression and removed helper breaking trust_remote_code are unrelated." + }, + { + "left": "issue:43824", + "right": "issue:43976", + "accept": false, + "reason": "Missing Qwen2_5_VL export and Python version support issue are unrelated." + }, + { + "left": "issue:43976", + "right": "issue:44964", + "accept": false, + "reason": "Python compatibility regression and Phi-4 multimodal loading failure are different issues." + }, + { + "left": "issue:44246", + "right": "issue:44871", + "accept": false, + "reason": "Import timing problem vs Gemma-3 eos_token_id mismatch; no shared bug." + }, + { + "left": "issue:44242", + "right": "issue:44857", + "accept": false, + "reason": "Different loss/AMP failures in different models; too broad to merge." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 32, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:36246", + "issue:41628", + "issue:41950", + "issue:42886", + "issue:43065", + "issue:43232", + "issue:43257", + "issue:43262", + "issue:43408", + "issue:43441", + "issue:43475", + "issue:43526", + "issue:43540", + "issue:43596", + "issue:43653", + "issue:43673", + "issue:43901", + "issue:44162", + "issue:44246", + "issue:44295", + "issue:44336", + "issue:44368", + "issue:44466", + "issue:44496", + "issue:44734", + "issue:44749", + "issue:44779", + "issue:44877", + "issue:44977", + "issue:44987", + "issue:44995", + "issue:45003", + "issue:45137", + "issue:45200", + "issue:45229", + "issue:45292", + "issue:45335", + "issue:45362", + "issue:45375", + "issue:45397", + "issue:45459", + "issue:45561", + "issue:45563", + "issue:45584", + "issue:45588", + "issue:45656", + "issue:45701", + "issue:45759" + ], + "soft_pairs": [ + "issue:45362|issue:45375", + "issue:44162|issue:45200", + "issue:44295|issue:44749", + "issue:44779|issue:45292", + "issue:43441|issue:43540", + "issue:45292|issue:45584", + "issue:44877|issue:45375", + "issue:36246|issue:43596", + "issue:44987|issue:45459", + "issue:45292|issue:45701", + "issue:41628|issue:41950", + "issue:44779|issue:45701", + "issue:45561|issue:45563", + "issue:43901|issue:45003", + "issue:45375|issue:45759", + "issue:43065|issue:43408", + "issue:43232|issue:44336", + "issue:43257|issue:45137", + "issue:43526|issue:45003", + "issue:44995|issue:45200", + "issue:44496|issue:45003", + "issue:43262|issue:44246", + "issue:42886|issue:45229", + "issue:43475|issue:43540", + "issue:44466|issue:45335", + "issue:44995|issue:45588", + "issue:43653|issue:44162", + "issue:44368|issue:45563", + "issue:45397|issue:45656", + "issue:44368|issue:44977", + "issue:43673|issue:44734" + ], + "prepared_review_unit_hash": "d035c946b33a38e7aefba61d37ac04b970a6d8e17c2a11c151e9e80669c52969", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 31, + "serialized_chars": 37083, + "estimated_input_tokens": 9271, + "estimated_eval_tokens": 18798 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is heterogeneous: most items are distinct Transformers regressions across tokenization, config loading, generation, DeepSpeed, and SAM3/Qwen/Gemma code paths. None of the soft pairs look like the same concrete bug/change, so I would not merge any of them as duplicates.", + "confidence": 0.68, + "canonical_issue_reason": "issue:45292 is the best representative of the small embedding/weight-resizing regression subset: it is concrete, broadly applicable, and closer to other nearby resize/tied-weight reports than the more model-specific tickets.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45292 is the clearest standalone bug in the set and the most reusable triage anchor for this cluster; it has a direct API symptom and broad model impact.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:45362", + "right": "issue:45375", + "accept": false, + "reason": "Both mention Qwen 3.5, but one is a chat/runtime crash and the other is a strict-config field drop; different code paths." + }, + { + "left": "issue:44162", + "right": "issue:45200", + "accept": false, + "reason": "ESM2 loading is unrelated to Gemma 4 mm_token_type_ids defaulting; different models and failures." + }, + { + "left": "issue:44295", + "right": "issue:44749", + "accept": false, + "reason": "One is a position_ids buffer bug, the other is a post-upgrade tokenization slowdown; not the same issue." + }, + { + "left": "issue:44779", + "right": "issue:45292", + "accept": false, + "reason": "DeepSeek tokenizer output regression and token embedding resize propagation are different bugs." + }, + { + "left": "issue:43441", + "right": "issue:43540", + "accept": false, + "reason": "FlashAttention crash in Ministral-3 and Qwen3OmniMoe video-input validation are unrelated." + }, + { + "left": "issue:45292", + "right": "issue:45584", + "accept": false, + "reason": "Resize-token-embeddings affects embedding weights, while Whisper empty-transcription failure is about special-token/chat-template handling." + }, + { + "left": "issue:44877", + "right": "issue:45375", + "accept": false, + "reason": "Both involve config handling, but one is strict loading of granite_speech and the other is a missing Qwen3_5MoeVisionConfig field." + }, + { + "left": "issue:36246", + "right": "issue:43596", + "accept": false, + "reason": "Importing a missing Qwen2.5 VL processor and a ZeRO3/BertModel index error are clearly different bugs." + }, + { + "left": "issue:44987", + "right": "issue:45459", + "accept": false, + "reason": "Model loading failure on transformers>=5.1.0 and protobuf-related tokenizer error masking are different failure points." + }, + { + "left": "issue:45292", + "right": "issue:45701", + "accept": false, + "reason": "Resize-token-embeddings behavior and version-dependent tokenization changes are not the same concrete regression." + }, + { + "left": "issue:41628", + "right": "issue:41950", + "accept": false, + "reason": "AutoImageProcessor import failure and video-classification pipeline choosing image processors are related only by naming, not the same bug." + }, + { + "left": "issue:44779", + "right": "issue:45701", + "accept": false, + "reason": "Both concern tokenization, but one is DeepSeek-specific incorrect output and the other is a broader version-dependent tokenization change." + }, + { + "left": "issue:45561", + "right": "issue:45563", + "accept": false, + "reason": "A pytest-xdist file race and a stale generation warning are unrelated test vs generation issues." + }, + { + "left": "issue:43901", + "right": "issue:45003", + "accept": false, + "reason": "Docs drift for return_all_scores is unrelated to the sys.modules access bug in modeling_utils." + }, + { + "left": "issue:45375", + "right": "issue:45759", + "accept": false, + "reason": "One is a missing config field being dropped, the other is AutoModelForCausalLM.from_config not unwrapping text_config." + }, + { + "left": "issue:43065", + "right": "issue:43408", + "accept": false, + "reason": "SAM3 pixel-decoder dummy conv and sam3_video/sam3_tracker warning are different SAM3 issues with different symptoms." + }, + { + "left": "issue:43232", + "right": "issue:44336", + "accept": false, + "reason": "Generation kwargs after sync_gpus and ANSI codes in loading_report are unrelated." + }, + { + "left": "issue:43257", + "right": "issue:45137", + "accept": false, + "reason": "Both mention DeepSpeed, but Qwen3 MOE weight conversion and ZeRO3 deque underflow are different code paths." + }, + { + "left": "issue:43526", + "right": "issue:45003", + "accept": false, + "reason": "BeitImageProcessorFast label reduction and unsafe sys.modules access are unrelated." + }, + { + "left": "issue:44995", + "right": "issue:45200", + "accept": false, + "reason": "GlmMoeDsa stale indexer cache and Gemma 4 token-type defaults are unrelated." + }, + { + "left": "issue:44496", + "right": "issue:45003", + "accept": false, + "reason": "Unrecognized model config and sys.modules access are separate core-loading failures." + }, + { + "left": "issue:43262", + "right": "issue:44246", + "accept": false, + "reason": "Audio processor sampling-rate defaults and occasional slow imports are unrelated." + }, + { + "left": "issue:42886", + "right": "issue:45229", + "accept": false, + "reason": "Offline tokenizer cache loading and Gemma 4 multi-GPU OOM are different problems." + }, + { + "left": "issue:43475", + "right": "issue:43540", + "accept": false, + "reason": "SAM3 vision encoder output attribute error and Qwen3OmniMoe video input ValueError are distinct model bugs." + }, + { + "left": "issue:44466", + "right": "issue:45335", + "accept": false, + "reason": "Tied-weight serialization inconsistency is not the same as resize_token_embeddings failing to update decoder.embed_tokens." + }, + { + "left": "issue:44995", + "right": "issue:45588", + "accept": false, + "reason": "A stale cache crash in GlmMoeDsa and flash_attention.py s_aux=None handling are different issues." + }, + { + "left": "issue:43653", + "right": "issue:44162", + "accept": false, + "reason": "BigBirdTokenizer special-token registration and ESM2 broken loading are unrelated." + }, + { + "left": "issue:44368", + "right": "issue:45563", + "accept": false, + "reason": "Both are warnings, but one is tie_word_embeddings config noise and the other is num_return_sequences warning drift." + }, + { + "left": "issue:45397", + "right": "issue:45656", + "accept": false, + "reason": "Gemma-4 ZeRO3 loading and duplicate optimizer.step() under DeepSpeed are different failures." + }, + { + "left": "issue:44368", + "right": "issue:44977", + "accept": false, + "reason": "Tie-word-embeddings warning noise is unrelated to Qwen3.5 flash-attention generation failure." + }, + { + "left": "issue:43673", + "right": "issue:44734", + "accept": false, + "reason": "GenerationMixin cache missing in chunked_prefill and KV-cache continuation indexing are both cache-related but not the same concrete bug." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 33, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:28282", + "issue:30990", + "issue:41628", + "issue:42222", + "issue:42371", + "issue:42673", + "issue:42994", + "issue:43232", + "issue:43452", + "issue:43475", + "issue:43482", + "issue:43525", + "issue:43643", + "issue:43646", + "issue:43673", + "issue:43704", + "issue:43827", + "issue:43931", + "issue:44038", + "issue:44060", + "issue:44077", + "issue:44162", + "issue:44297", + "issue:44361", + "issue:44368", + "issue:44623", + "issue:44661", + "issue:44749", + "issue:44869", + "issue:44933", + "issue:45003", + "issue:45081", + "issue:45092", + "issue:45103", + "issue:45137", + "issue:45200", + "issue:45230", + "issue:45310", + "issue:45356", + "issue:45375", + "issue:45381", + "issue:45397", + "issue:45459", + "issue:45478", + "issue:45561", + "issue:45600", + "issue:45656" + ], + "soft_pairs": [ + "issue:44060|issue:44077", + "issue:45003|issue:45561", + "issue:45137|issue:45600", + "issue:43931|issue:44368", + "issue:44933|issue:45459", + "issue:44869|issue:45356", + "issue:44297|issue:44623", + "issue:43646|issue:44162", + "issue:44361|issue:44749", + "issue:43232|issue:43673", + "issue:44623|issue:45656", + "issue:28282|issue:30990", + "issue:43673|issue:43827", + "issue:43452|issue:45003", + "issue:42994|issue:43482", + "issue:44749|issue:45310", + "issue:43475|issue:43643", + "issue:44749|issue:45656", + "issue:45230|issue:45397", + "issue:44038|issue:44623", + "issue:42371|issue:45092", + "issue:43475|issue:45381", + "issue:45081|issue:45103", + "issue:41628|issue:44933", + "issue:45375|issue:45478", + "issue:43475|issue:44661", + "issue:42673|issue:43704", + "issue:42994|issue:43525", + "issue:45310|issue:45375", + "issue:42222|issue:45200" + ], + "prepared_review_unit_hash": "8a45655a08e6d8a93d8fc4e2698ffe6d9edd405b96aa877f87921227b76fef7e", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 30, + "serialized_chars": 36049, + "estimated_input_tokens": 9013, + "estimated_eval_tokens": 18282 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The set is mostly a heterogeneous grab bag of Transformers/model-loading regressions, not one tight duplicate cluster. The only clear mini-cluster is the Qwen3.5-MoE loading/config issue, where one report names the likely root cause and the others look like symptom duplicates.", + "confidence": 0.37, + "canonical_issue_reason": "issue:45375 is the best anchor because it identifies the likely root cause (missing `deepstack_visual_indexes` being dropped by `@strict`) behind the Qwen3.5-MoE `from_pretrained` failures; the other Qwen3.5-MoE reports are downstream symptom duplicates.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45375 is the most actionable and root-cause-oriented report in the only clear duplicate subcluster, and it best explains the related loading errors in 45310/45478.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44060", + "right": "issue:44077", + "accept": false, + "reason": "Different model families and different warnings; no shared concrete bug path." + }, + { + "left": "issue:45003", + "right": "issue:45561", + "accept": false, + "reason": "`sys.modules` access bug vs xdist race on `captured_info.txt`; unrelated failures." + }, + { + "left": "issue:45137", + "right": "issue:45600", + "accept": false, + "reason": "DeepSpeed ZeRO3 deque error vs removed Sam3LiteText config references; different code paths." + }, + { + "left": "issue:43931", + "right": "issue:44368", + "accept": false, + "reason": "Both mention Qwen, but one is a weight-shape load error and the other is a tied-weights warning; not the same bug." + }, + { + "left": "issue:44933", + "right": "issue:45459", + "accept": false, + "reason": "Import missing from `image_utils` vs protobuf exception masking tokenizer errors; different root causes." + }, + { + "left": "issue:44869", + "right": "issue:45356", + "accept": false, + "reason": "Whisper timestamp decode crash vs Kimi tokenizer codec/regression warning; unrelated." + }, + { + "left": "issue:44297", + "right": "issue:44623", + "accept": false, + "reason": "Tokenizer save metadata mismatch vs processor save missing files; same area, but not the same concrete bug." + }, + { + "left": "issue:43646", + "right": "issue:44162", + "accept": false, + "reason": "Broad v5 custom-model init breakage vs an ESM2-specific breakage; too different to merge." + }, + { + "left": "issue:44361", + "right": "issue:44749", + "accept": false, + "reason": "Tokenizer AttributeError vs a data-filtering slowdown regression; not the same failure." + }, + { + "left": "issue:43232", + "right": "issue:43673", + "accept": false, + "reason": "Different generation internals (`sync_gpus` kwargs update vs chunked prefill cache missing)." + }, + { + "left": "issue:44623", + "right": "issue:45656", + "accept": false, + "reason": "Processor serialization bug vs DeepSpeed optimizer stepping twice; unrelated." + }, + { + "left": "issue:28282", + "right": "issue:30990", + "accept": false, + "reason": "ImportError for PyTorch vs a loading hang in Sentence Transformers; different symptoms and likely causes." + }, + { + "left": "issue:43673", + "right": "issue:43827", + "accept": false, + "reason": "Generation cache regression vs docs still referencing removed `pipeline()` APIs; not a duplicate bug." + }, + { + "left": "issue:43452", + "right": "issue:45003", + "accept": false, + "reason": "GGUF loading breakage vs unsafe `sys.modules` access; different code paths." + }, + { + "left": "issue:42994", + "right": "issue:43482", + "accept": false, + "reason": "Quantized saving failure vs Qwen2.5-GGUF loading failure; opposite sides of the workflow and different bugs." + }, + { + "left": "issue:44749", + "right": "issue:45310", + "accept": false, + "reason": "Performance regression on upgraded Transformers vs Qwen3.5-MoE loading error; unrelated." + }, + { + "left": "issue:43475", + "right": "issue:43643", + "accept": false, + "reason": "SAM3 vision encoder attribute error vs missing fields returned from `AutoConfig`; different models and failures." + }, + { + "left": "issue:45230", + "right": "issue:45397", + "accept": false, + "reason": "Generic bug report vs Gemma-4 DeepSpeed ZeRO3 loading issue; no evidence they match." + }, + { + "left": "issue:44038", + "right": "issue:44623", + "accept": false, + "reason": "Qwen3-VL-Moe bug vs processor save missing files; not the same underlying problem." + }, + { + "left": "issue:42371", + "right": "issue:45092", + "accept": false, + "reason": "TF32 settings docs issue vs InternVL2 meta-init incompatibility; documentation vs runtime bug." + }, + { + "left": "issue:43475", + "right": "issue:45381", + "accept": false, + "reason": "Different video-model failures (`Sam3VisionEncoderOutput` attribute vs Qwen2.5-VL position IDs)." + }, + { + "left": "issue:45081", + "right": "issue:45103", + "accept": false, + "reason": "Tokenizer regex crash vs docstring/annotations crash; unrelated subsystems." + }, + { + "left": "issue:41628", + "right": "issue:44933", + "accept": false, + "reason": "Different import failures in different modules; not the same missing symbol bug." + }, + { + "left": "issue:45375", + "right": "issue:45478", + "accept": true, + "reason": "Same Qwen3.5-MoE loading regression; 45375 explains the likely root cause that would surface as 45478's `from_pretrained` error." + }, + { + "left": "issue:43475", + "right": "issue:44661", + "accept": false, + "reason": "SAM3 output attribute issue vs tokenizer mapping failure when adding a model; unrelated." + }, + { + "left": "issue:42673", + "right": "issue:43704", + "accept": false, + "reason": "Same title, but they are separate reports with no evidence of a shared code-path bug from the provided text." + }, + { + "left": "issue:42994", + "right": "issue:43525", + "accept": false, + "reason": "Quantized save failure vs missing `pad_token_id` on `Llama4Config`; different bugs." + }, + { + "left": "issue:45310", + "right": "issue:45375", + "accept": true, + "reason": "Same Qwen3.5-MoE failure family; 45375 pinpoints the missing config field that plausibly causes 45310's `from_pretrained` error." + }, + { + "left": "issue:42222", + "right": "issue:45200", + "accept": false, + "reason": "VitPose model breakage vs Gemma-4 token-type-id defaulting; different model-specific issues." + }, + { + "left": "issue:44623", + "right": "issue:45656", + "accept": false, + "reason": "Processor save regression vs DeepSpeed optimizer-step duplication; not the same bug." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "Grounded overall and conservative in the soft-edge calls. The summary appropriately treats the packet as mostly heterogeneous and limits the duplicate claim to the Qwen3.5-MoE loading/config trio, which is the only clearly supported mini-cluster from the titles provided." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 34, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:30333", + "issue:30990", + "issue:33290", + "issue:35141", + "issue:36010", + "issue:41628", + "issue:42175", + "issue:42831", + "issue:42898", + "issue:42994", + "issue:43232", + "issue:43278", + "issue:43377", + "issue:43479", + "issue:43723", + "issue:43784", + "issue:43824", + "issue:43874", + "issue:43901", + "issue:44038", + "issue:44060", + "issue:44079", + "issue:44164", + "issue:44188", + "issue:44222", + "issue:44242", + "issue:44295", + "issue:44351", + "issue:44373", + "issue:44448", + "issue:44496", + "issue:44568", + "issue:44623", + "issue:44734", + "issue:44779", + "issue:44792", + "issue:44964", + "issue:45092", + "issue:45103", + "issue:45125", + "issue:45137", + "issue:45310", + "issue:45356", + "issue:45405", + "issue:45561", + "issue:45600", + "issue:45656", + "issue:45712" + ], + "soft_pairs": [ + "issue:42831|issue:44222", + "issue:41628|issue:44351", + "issue:44222|issue:45310", + "issue:42994|issue:44164", + "issue:42175|issue:45712", + "issue:43377|issue:44242", + "issue:36010|issue:41628", + "issue:44038|issue:45310", + "issue:44496|issue:45561", + "issue:45561|issue:45600", + "issue:44964|issue:45137", + "issue:42898|issue:44568", + "issue:30990|issue:35141", + "issue:43901|issue:45561", + "issue:43232|issue:44792", + "issue:43232|issue:44188", + "issue:43784|issue:43824", + "issue:44188|issue:44373", + "issue:43278|issue:45137", + "issue:44623|issue:45310", + "issue:44734|issue:45103", + "issue:30333|issue:33290", + "issue:42898|issue:45356", + "issue:43723|issue:44568", + "issue:44060|issue:45125", + "issue:42898|issue:44779", + "issue:45405|issue:45712", + "issue:43874|issue:45092", + "issue:42898|issue:44448", + "issue:43479|issue:44079", + "issue:44188|issue:44964", + "issue:44295|issue:45656" + ], + "prepared_review_unit_hash": "45df2ddf32ced97ef49075c2cb6c421075e9ed25609c76dd9efc1b94bbe9cb4e", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 32, + "serialized_chars": 37267, + "estimated_input_tokens": 9317, + "estimated_eval_tokens": 18890 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is highly heterogeneous: mostly unrelated bug reports spanning tokenizer regressions, import errors, save/load issues, multimodal model failures, and DeepSpeed/compile edge cases. None of the soft-edge pairs look like the same underlying defect, so all are rejected.", + "confidence": 0.92, + "canonical_issue_reason": "issue:42898 is the broadest, most central regression in the set: a v5 behavior change affecting tokenization semantics rather than a model-specific one-off. It best represents the general \u2018transformers v5 breaking behavior\u2019 theme visible across several nearby issues.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:42898 is the strongest representative issue because it is concrete, broadly relevant, and not tied to a single model family or integration path.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:42831", + "right": "issue:44222", + "accept": false, + "reason": "Both mention FP8, but one is an accuracy issue and the other is a save_pretrained failure for MoE; different code paths." + }, + { + "left": "issue:41628", + "right": "issue:44351", + "accept": false, + "reason": "Both are import errors, but they concern different missing symbols and different modules." + }, + { + "left": "issue:44222", + "right": "issue:45310", + "accept": false, + "reason": "One is a FP8 save_pretrained bug and the other is a Qwen3.5 from_pretrained error; not the same defect." + }, + { + "left": "issue:42994", + "right": "issue:44164", + "accept": false, + "reason": "Both involve serialization, but one is quantized model saving and the other is extra_state handling in save/from_pretrained." + }, + { + "left": "issue:42175", + "right": "issue:45712", + "accept": false, + "reason": "Packaging/backend availability versus dummy class leakage into torch-less environments are different problems." + }, + { + "left": "issue:43377", + "right": "issue:44242", + "accept": false, + "reason": "Missing padding-mask support in an encoder is unrelated to load-balancing loss gating on router logits." + }, + { + "left": "issue:36010", + "right": "issue:41628", + "accept": false, + "reason": "Different missing imports (`GenerationMixin` vs `AutoImageProcessor`) and different failure surfaces." + }, + { + "left": "issue:44038", + "right": "issue:45310", + "accept": false, + "reason": "Both touch Qwen/MoE and v5, but the reported failures are distinct and no shared code path is shown." + }, + { + "left": "issue:44496", + "right": "issue:45561", + "accept": false, + "reason": "A model config loading error and a pytest-xdist file race are unrelated." + }, + { + "left": "issue:45561", + "right": "issue:45600", + "accept": false, + "reason": "Concurrency on captured_info.txt is unrelated to stale auto_mappings references." + }, + { + "left": "issue:44964", + "right": "issue:45137", + "accept": false, + "reason": "Multimodal model loading and a DeepSpeed ZeRO3 deque error are separate bugs." + }, + { + "left": "issue:42898", + "right": "issue:44568", + "accept": false, + "reason": "Both are tokenization regressions in v5, but they affect different tokenizer behaviors and models." + }, + { + "left": "issue:30990", + "right": "issue:35141", + "accept": false, + "reason": "A loading hang and an embedding reinitialization bug are not the same underlying issue." + }, + { + "left": "issue:43901", + "right": "issue:45561", + "accept": false, + "reason": "A documentation mismatch is unrelated to a test utility race condition." + }, + { + "left": "issue:43232", + "right": "issue:44792", + "accept": false, + "reason": "Generation kwargs handling and a Janus image-generation test failure do not share the same concrete bug." + }, + { + "left": "issue:43232", + "right": "issue:44188", + "accept": false, + "reason": "Generation bookkeeping after sync_gpus is unrelated to torch.compile-dependent attention kernel divergence." + }, + { + "left": "issue:43784", + "right": "issue:43824", + "accept": false, + "reason": "Both are import failures, but for different names in different packages/models." + }, + { + "left": "issue:44188", + "right": "issue:44373", + "accept": false, + "reason": "A runtime kernel divergence bug is unrelated to a wrong docstring." + }, + { + "left": "issue:43278", + "right": "issue:45137", + "accept": false, + "reason": "BF16/FP32 dtype drift during eval is unrelated to a DeepSpeed deque underflow." + }, + { + "left": "issue:44623", + "right": "issue:45310", + "accept": false, + "reason": "Missing files in processor.save_pretrained and a Qwen3.5 from_pretrained error are different operations and bugs." + }, + { + "left": "issue:44734", + "right": "issue:45103", + "accept": false, + "reason": "A serving KV-cache indexing crash and an auto-docstring annotation crash are unrelated." + }, + { + "left": "issue:44779", + "right": "issue:44568", + "accept": false, + "reason": "Both are tokenizer regressions, but they involve different models and different tokenization behaviors." + }, + { + "left": "issue:44792", + "right": "issue:44964", + "accept": false, + "reason": "Different multimodal models and different failure modes; not mergeable into one fix." + }, + { + "left": "issue:45092", + "right": "issue:45137", + "accept": false, + "reason": "Remote-code/meta-init incompatibility is unrelated to a ZeRO3 deque error." + }, + { + "left": "issue:45103", + "right": "issue:44734", + "accept": false, + "reason": "Reversed pair of the same two unrelated issues; still not a duplicate." + }, + { + "left": "issue:45125", + "right": "issue:45310", + "accept": false, + "reason": "Both mention Qwen3.5 MoE, but one is tensor-parallel metadata and the other is a from_pretrained failure; too different to treat as one bug." + }, + { + "left": "issue:45356", + "right": "issue:42898", + "accept": false, + "reason": "Two separate tokenizer regressions affecting different models and behaviors." + }, + { + "left": "issue:45405", + "right": "issue:45712", + "accept": false, + "reason": "A dependency version bump issue is unrelated to dummy class cleanup." + }, + { + "left": "issue:43874", + "right": "issue:45092", + "accept": false, + "reason": "Different multimodal model families and different missing-method/meta-init problems." + }, + { + "left": "issue:42898", + "right": "issue:44448", + "accept": false, + "reason": "A tokenizer behavior change and a model output mismatch across v4/v5 are not the same defect." + }, + { + "left": "issue:43479", + "right": "issue:44079", + "accept": false, + "reason": "Default config initialization and ModelOutput key assignment are unrelated internals." + }, + { + "left": "issue:44188", + "right": "issue:44964", + "accept": false, + "reason": "Compile-dependent attention divergence does not match a model loading failure." + }, + { + "left": "issue:44295", + "right": "issue:45656", + "accept": false, + "reason": "A position_ids buffer access error and double optimizer stepping under DeepSpeed are different bugs." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 35, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:28282", + "issue:30333", + "issue:30990", + "issue:34689", + "issue:35141", + "issue:36331", + "issue:38175", + "issue:41628", + "issue:42175", + "issue:42757", + "issue:42831", + "issue:42886", + "issue:42898", + "issue:42994", + "issue:43295", + "issue:43479", + "issue:43644", + "issue:43723", + "issue:43824", + "issue:43873", + "issue:43976", + "issue:44206", + "issue:44222", + "issue:44261", + "issue:44263", + "issue:44295", + "issue:44297", + "issue:44360", + "issue:44373", + "issue:44483", + "issue:44484", + "issue:44485", + "issue:44514", + "issue:44623", + "issue:44869", + "issue:44871", + "issue:44913", + "issue:44933", + "issue:44945", + "issue:44977", + "issue:45092", + "issue:45103", + "issue:45310", + "issue:45357", + "issue:45406", + "issue:45459", + "issue:45656", + "issue:45712" + ], + "soft_pairs": [ + "issue:42898|issue:43295", + "issue:44933|issue:45712", + "issue:42175|issue:44933", + "issue:41628|issue:45459", + "issue:43976|issue:44933", + "issue:44222|issue:44297", + "issue:44297|issue:45656", + "issue:43479|issue:44869", + "issue:44206|issue:45103", + "issue:42831|issue:44623", + "issue:44222|issue:44623", + "issue:30333|issue:36331", + "issue:41628|issue:42757", + "issue:30333|issue:34689", + "issue:43723|issue:44977", + "issue:44297|issue:45310", + "issue:28282|issue:30333", + "issue:30333|issue:35141", + "issue:44222|issue:44484", + "issue:44483|issue:45406", + "issue:44263|issue:44360", + "issue:43873|issue:44945", + "issue:44263|issue:45092", + "issue:44514|issue:45103", + "issue:43479|issue:43644", + "issue:44373|issue:44485", + "issue:44295|issue:44623", + "issue:43723|issue:44871", + "issue:45092|issue:45357", + "issue:44261|issue:45092", + "issue:43479|issue:44913", + "issue:30990|issue:38175", + "issue:41628|issue:43824", + "issue:42886|issue:42994" + ], + "prepared_review_unit_hash": "14f28dffa56d704361441560d585f26a54f614daa74bb28b78392701f258f765", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 34, + "serialized_chars": 37709, + "estimated_input_tokens": 9428, + "estimated_eval_tokens": 19112 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is heterogeneous: most items are unrelated Transformers regressions spanning imports, tokenizers, multimodal processors, training, quantization, and save/load behavior. The soft-similarity pairs share broad themes only, not the same concrete bug or change, so none should be merged.", + "confidence": 0.82, + "canonical_issue_reason": "issue:41628 is the most representative user-facing import regression in the set and sits near several other import/load-related reports, making it the best single issue to anchor the cluster.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:41628 is the broadest and clearest representative of the import/load subset; it is actionable and more general than the model-specific or training-specific items around it.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:42898", + "right": "issue:43295", + "accept": false, + "reason": "Both are v5 regressions, but one is tokenization cleanup behavior and the other is processor/tokenizer attribute access for custom model code; different code paths." + }, + { + "left": "issue:44933", + "right": "issue:45712", + "accept": false, + "reason": "Both involve image_utils-related symbols, but one is a bad import target while the other is dummy PT classes leaking into dir() without torch; not the same bug." + }, + { + "left": "issue:42175", + "right": "issue:44933", + "accept": false, + "reason": "TensorFlow backend packaging and a missing image_utils import are different dependency/import problems." + }, + { + "left": "issue:41628", + "right": "issue:45459", + "accept": false, + "reason": "Different missing symbols in different modules: transformers import exposure vs tokenizer protobuf error handling." + }, + { + "left": "issue:43976", + "right": "issue:44933", + "accept": false, + "reason": "Python version compatibility regression is unrelated to the image_utils import issue." + }, + { + "left": "issue:44222", + "right": "issue:44297", + "accept": false, + "reason": "FP8 save_pretrained failure and tokenizer save metadata mismatch affect different save paths." + }, + { + "left": "issue:44297", + "right": "issue:45656", + "accept": false, + "reason": "Tokenizer config serialization and deepspeed optimizer-step behavior are unrelated." + }, + { + "left": "issue:43479", + "right": "issue:44869", + "accept": false, + "reason": "Multimodal config default initialization and Whisper timestamp decoding are different model-specific bugs." + }, + { + "left": "issue:44206", + "right": "issue:45103", + "accept": false, + "reason": "Feature extractor argument regression and auto_docstring annotation crash are separate code paths." + }, + { + "left": "issue:42831", + "right": "issue:44623", + "accept": false, + "reason": "FP8 accuracy issues and processor.save_pretrained file omissions are not the same underlying defect." + }, + { + "left": "issue:44222", + "right": "issue:44623", + "accept": false, + "reason": "Both affect saving, but FP8/moe serialization and processor artifact saving are distinct failures." + }, + { + "left": "issue:30333", + "right": "issue:36331", + "accept": false, + "reason": "MLflow reporting on failed jobs and a Trainer compute_loss signature break are unrelated." + }, + { + "left": "issue:41628", + "right": "issue:42757", + "accept": false, + "reason": "Missing AutoImageProcessor import and missing is_offline_mode from huggingface_hub are different import regressions." + }, + { + "left": "issue:30333", + "right": "issue:34689", + "accept": false, + "reason": "MLflow status reporting and Llama 3.2 model loading are unrelated bugs." + }, + { + "left": "issue:43723", + "right": "issue:44977", + "accept": false, + "reason": "Tokenizer loading from cache and flash-attention generation errors are different failure modes." + }, + { + "left": "issue:44297", + "right": "issue:45310", + "accept": false, + "reason": "Tokenizer save metadata mismatch and Qwen3.5 MoE from_pretrained regression are separate issues." + }, + { + "left": "issue:28282", + "right": "issue:30333", + "accept": false, + "reason": "PyTorch missing ImportError and MLflow job reporting are not the same problem." + }, + { + "left": "issue:30333", + "right": "issue:35141", + "accept": false, + "reason": "Failed-job reporting and embedding resize/post_init reinitialization are unrelated." + }, + { + "left": "issue:44222", + "right": "issue:44484", + "accept": false, + "reason": "FP8 save failure and a max_shard_size behavior question are not duplicates." + }, + { + "left": "issue:44483", + "right": "issue:45406", + "accept": false, + "reason": "Chat completions request handling and Gemma4Processor _tokenizer access crash are different API paths." + }, + { + "left": "issue:44263", + "right": "issue:44360", + "accept": false, + "reason": "Torch.split indexing behavior and a missing ReLU in the DSA indexer are different bugs in different logic." + }, + { + "left": "issue:43873", + "right": "issue:44945", + "accept": false, + "reason": "Quantization offloading and pipeline parallelism output corruption are separate distributed-inference issues." + }, + { + "left": "issue:44263", + "right": "issue:45092", + "accept": false, + "reason": "torch.split return handling and remote-code meta initialization incompatibility do not share the same concrete defect." + }, + { + "left": "issue:44514", + "right": "issue:45103", + "accept": false, + "reason": "Batched chat template padding crash and auto_docstring annotation crash are unrelated processor/docstring issues." + }, + { + "left": "issue:43479", + "right": "issue:43644", + "accept": false, + "reason": "Default multimodal config initialization and non-persistent buffer junk are different initialization bugs." + }, + { + "left": "issue:44373", + "right": "issue:44485", + "accept": false, + "reason": "Wrong docstring for position_ids and GLM-5 RoPE implementation are not the same bug." + }, + { + "left": "issue:44295", + "right": "issue:44623", + "accept": false, + "reason": "Buffer readback after registration and processor.save_pretrained missing files are different serialization/storage issues." + }, + { + "left": "issue:43723", + "right": "issue:44871", + "accept": false, + "reason": "Tokenizer loading failure and eos_token_id config inconsistency are distinct tokenizer-related bugs." + }, + { + "left": "issue:45092", + "right": "issue:45357", + "accept": false, + "reason": "Transformers v5 meta-init incompatibility and incorrect visual encoder keys on save_pretrained are separate model lifecycle problems." + }, + { + "left": "issue:44261", + "right": "issue:45092", + "accept": false, + "reason": "MLA layernorm epsilon precision error and remote-code meta initialization incompatibility are unrelated." + }, + { + "left": "issue:30990", + "right": "issue:38175", + "accept": false, + "reason": "Sentence Transformers loading hang and zero probabilities in siglip2-base-patch16-224 are different model behaviors." + }, + { + "left": "issue:41628", + "right": "issue:43824", + "accept": false, + "reason": "Both are import errors, but they concern different symbols and model families; not the same missing code path." + }, + { + "left": "issue:42886", + "right": "issue:42994", + "accept": false, + "reason": "Offline tokenizer cache loading and quantized model saving failures are unrelated." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 36, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:22355", + "issue:28282", + "issue:30990", + "issue:33357", + "issue:41628", + "issue:42371", + "issue:42757", + "issue:42831", + "issue:42886", + "issue:42907", + "issue:43066", + "issue:43116", + "issue:43122", + "issue:43232", + "issue:43408", + "issue:43475", + "issue:43576", + "issue:43723", + "issue:43756", + "issue:43784", + "issue:43937", + "issue:44038", + "issue:44060", + "issue:44117", + "issue:44261", + "issue:44263", + "issue:44295", + "issue:44297", + "issue:44360", + "issue:44373", + "issue:44442", + "issue:44485", + "issue:44514", + "issue:44521", + "issue:44625", + "issue:44704", + "issue:44743", + "issue:44779", + "issue:44829", + "issue:44945", + "issue:45200", + "issue:45245", + "issue:45278", + "issue:45290", + "issue:45310", + "issue:45593", + "issue:45600", + "issue:45753" + ], + "soft_pairs": [ + "issue:22355|issue:28282", + "issue:42831|issue:45310", + "issue:44263|issue:44373", + "issue:44945|issue:45753", + "issue:42831|issue:44038", + "issue:42757|issue:45278", + "issue:43576|issue:43784", + "issue:42886|issue:43066", + "issue:43232|issue:45593", + "issue:44117|issue:44704", + "issue:30990|issue:33357", + "issue:42371|issue:44263", + "issue:44704|issue:45290", + "issue:43756|issue:44060", + "issue:42907|issue:44945", + "issue:43122|issue:44514", + "issue:44295|issue:44297", + "issue:43232|issue:45245", + "issue:43723|issue:44779", + "issue:43408|issue:43475", + "issue:43723|issue:44442", + "issue:43122|issue:44521", + "issue:43784|issue:45600", + "issue:42907|issue:43937", + "issue:41628|issue:45278", + "issue:44261|issue:44373", + "issue:44360|issue:44485", + "issue:43232|issue:44743", + "issue:43116|issue:45200", + "issue:44625|issue:44829", + "issue:43122|issue:44779" + ], + "prepared_review_unit_hash": "5c489127a4c42cad715d8a6e163ae4dda5e9d789bbe6f427a4da87f76567a1d2", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 31, + "serialized_chars": 36587, + "estimated_input_tokens": 9147, + "estimated_eval_tokens": 18550 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This looks like a loose bundle of mostly unrelated Transformers issues, not a tight duplicate cluster. No soft edge is convincing enough to merge; the closest theme is the broad import-regression report, which is the best anchor if one must be chosen.", + "confidence": 0.63, + "canonical_issue_reason": "issue:45278 is the broadest umbrella-style report in the set (many import errors after upgrading), and it is the closest match to the other import-related regressions.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45278 is the best global representative because it is the most general and likely to subsume other import-related failures in this bundle.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:22355", + "right": "issue:28282", + "accept": false, + "reason": "Both are import/load failures, but one is about transformers.onnx and the other about missing PyTorch; different root causes." + }, + { + "left": "issue:42831", + "right": "issue:45310", + "accept": false, + "reason": "FineGrainedFP8 accuracy drift and a Qwen3.5 MOE from_pretrained error are unrelated problems." + }, + { + "left": "issue:44263", + "right": "issue:44373", + "accept": false, + "reason": "A torch.split return-value issue and a docstring bug are not the same underlying defect." + }, + { + "left": "issue:44945", + "right": "issue:45753", + "accept": false, + "reason": "Pipeline-parallelism output corruption and a specific Qwen3_5 image infinite loop are distinct model/runtime bugs." + }, + { + "left": "issue:42831", + "right": "issue:44038", + "accept": false, + "reason": "Different symptoms and code paths: FP8 accuracy vs Qwen3-VL-Moe loading/usage." + }, + { + "left": "issue:42757", + "right": "issue:45278", + "accept": false, + "reason": "45278 is a broad import-regression report, but 42757 is a specific huggingface_hub symbol import error; not enough to call them the same bug." + }, + { + "left": "issue:43576", + "right": "issue:43784", + "accept": false, + "reason": "An env command regression and a sentence-transformers import NameError are different failures." + }, + { + "left": "issue:42886", + "right": "issue:43066", + "accept": false, + "reason": "Offline cache loading and wrong tokenizer decoder type are different tokenizer issues." + }, + { + "left": "issue:43232", + "right": "issue:45593", + "accept": false, + "reason": "Generation state handling after sync_gpus is unrelated to D-FINE auxiliary-loss behavior." + }, + { + "left": "issue:44117", + "right": "issue:44704", + "accept": false, + "reason": "TOKENIZER_MAPPING_NAMES/ from_pretrained logic differs from AutoProcessor kwargs forwarding." + }, + { + "left": "issue:30990", + "right": "issue:33357", + "accept": false, + "reason": "Sentence Transformers loading hang and a MacOS bus error on CLIP are separate issues." + }, + { + "left": "issue:42371", + "right": "issue:44263", + "accept": false, + "reason": "TF32 settings guidance and torch.split behavior have no direct overlap." + }, + { + "left": "issue:44704", + "right": "issue:45290", + "accept": false, + "reason": "Cached-file kwargs forwarding is unrelated to apply_chat_template crashing on tool-call assistant messages." + }, + { + "left": "issue:43756", + "right": "issue:44060", + "accept": false, + "reason": "Smollm3 RoPE layer handling and Qwen3-Next tied-weights warnings are different model-specific bugs." + }, + { + "left": "issue:42907", + "right": "issue:44945", + "accept": false, + "reason": "Saving dequantized Ministrals and incorrect pipeline-parallel output are unrelated." + }, + { + "left": "issue:43122", + "right": "issue:44514", + "accept": false, + "reason": "Both touch tokenization/chat templating, but one is a version-to-version tokenization difference and the other is a batched multimodal crash." + }, + { + "left": "issue:44295", + "right": "issue:44297", + "accept": false, + "reason": "Reading a registered position_ids buffer and saving tokenizer_class metadata are unrelated." + }, + { + "left": "issue:43232", + "right": "issue:45245", + "accept": false, + "reason": "Generation cache behavior and the 2^24 category limit error are not the same bug." + }, + { + "left": "issue:43723", + "right": "issue:44779", + "accept": false, + "reason": "A generic tokenizer load issue and incorrect DeepSeek tokenization are different failure modes." + }, + { + "left": "issue:43408", + "right": "issue:43475", + "accept": false, + "reason": "Model-type mismatch warning and missing fpn_position_embeddings are related to the same family, but not the same concrete defect." + }, + { + "left": "issue:43723", + "right": "issue:44442", + "accept": false, + "reason": "Generic AutoTokenizer load regression and FastSpeech2ConformerTokenizer load failure are distinct." + }, + { + "left": "issue:43122", + "right": "issue:44521", + "accept": false, + "reason": "Different tokenization across versions and zero assistant masks for multimodal inputs are not the same issue." + }, + { + "left": "issue:43784", + "right": "issue:45600", + "accept": false, + "reason": "Sentence-transformers import NameError and broken auto_mappings CI are unrelated." + }, + { + "left": "issue:42907", + "right": "issue:43937", + "accept": false, + "reason": "Dequantized save behavior and invalid GenerationConfig for GLM-5 are separate problems." + }, + { + "left": "issue:41628", + "right": "issue:45278", + "accept": false, + "reason": "Both are import-related, but 41628 is a single missing AutoImageProcessor import while 45278 is a broad upgrade regression; not enough evidence of one root cause." + }, + { + "left": "issue:44261", + "right": "issue:44373", + "accept": false, + "reason": "A missing rms_norm_eps config value and a wrong docstring do not share the same bug." + }, + { + "left": "issue:44360", + "right": "issue:44485", + "accept": false, + "reason": "DSA indexer activation choice and GLM-5 RoPE implementation are unrelated." + }, + { + "left": "issue:43232", + "right": "issue:44743", + "accept": false, + "reason": "Both involve generation/cache semantics, but the sync_gpus bug and recurrent-state reset bug are different code paths." + }, + { + "left": "issue:43116", + "right": "issue:45200", + "accept": false, + "reason": "Example-script multi-label output is unrelated to mm_token_type_ids defaults for Gemma 4 fine-tuning." + }, + { + "left": "issue:44625", + "right": "issue:44829", + "accept": false, + "reason": "num_labels propagation and flash_attention_3 degenerate training are different sequence-classification issues." + }, + { + "left": "issue:43122", + "right": "issue:44779", + "accept": false, + "reason": "Generic version-to-version tokenization changes and DeepSeek-specific incorrect tokenization are not clearly the same defect." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 37, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:29127", + "issue:30990", + "issue:33453", + "issue:36296", + "issue:38617", + "issue:39401", + "issue:42222", + "issue:42371", + "issue:42757", + "issue:42831", + "issue:42907", + "issue:43064", + "issue:43066", + "issue:43116", + "issue:43122", + "issue:43441", + "issue:43452", + "issue:43493", + "issue:43673", + "issue:43784", + "issue:43824", + "issue:43825", + "issue:43874", + "issue:43937", + "issue:43976", + "issue:43992", + "issue:44038", + "issue:44075", + "issue:44222", + "issue:44295", + "issue:44297", + "issue:44351", + "issue:44485", + "issue:44488", + "issue:44521", + "issue:44704", + "issue:44779", + "issue:44936", + "issue:44964", + "issue:44977", + "issue:45083", + "issue:45216", + "issue:45278", + "issue:45446", + "issue:45478", + "issue:45563", + "issue:45753" + ], + "soft_pairs": [ + "issue:29127|issue:30990", + "issue:42907|issue:44964", + "issue:44038|issue:44297", + "issue:42907|issue:43452", + "issue:44038|issue:44295", + "issue:39401|issue:44488", + "issue:36296|issue:43116", + "issue:43824|issue:44351", + "issue:44038|issue:44222", + "issue:43066|issue:44779", + "issue:39401|issue:44521", + "issue:45216|issue:45478", + "issue:42757|issue:44351", + "issue:43122|issue:44977", + "issue:43992|issue:44704", + "issue:42907|issue:44222", + "issue:43784|issue:45446", + "issue:43937|issue:44485", + "issue:42222|issue:43441", + "issue:42371|issue:43874", + "issue:43976|issue:45446", + "issue:43493|issue:45753", + "issue:42831|issue:45753", + "issue:44075|issue:44936", + "issue:43116|issue:43825", + "issue:44222|issue:45478", + "issue:44038|issue:45478", + "issue:43116|issue:43493", + "issue:43064|issue:43673", + "issue:45083|issue:45446", + "issue:43937|issue:45563", + "issue:33453|issue:39401", + "issue:39401|issue:44779", + "issue:38617|issue:45278" + ], + "prepared_review_unit_hash": "47783cbf09764d6fe4b3847189916d7374639fb0d3e5c6d7deffa2ff6333faac", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 47, + "item_count": 47, + "soft_pair_count": 34, + "serialized_chars": 36776, + "estimated_input_tokens": 9194, + "estimated_eval_tokens": 18644 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is heterogeneous: import errors, tokenizer regressions, model-specific loading/saving issues, training bugs, and unrelated warnings. None of the soft edges look like the same concrete bug or change, so all candidate duplicate links should be rejected.", + "confidence": 0.87, + "canonical_issue_reason": "Issue 45278 is the broadest umbrella for the import-error subset, but the overall cluster is too mixed to justify a true canonical duplicate.", + "canonical_pr_reason": null, + "best_issue_reason": "45278 is the best representative only for the import-regression subset; it is broader than the others and most likely to absorb similar reports.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:29127", + "right": "issue:30990", + "accept": false, + "reason": "Different subsystems and symptoms: layoutlmv3 box validation vs sentence-transformers loading hang." + }, + { + "left": "issue:42907", + "right": "issue:44964", + "accept": false, + "reason": "Both are model-loading/saving related, but they are different model families and failure modes." + }, + { + "left": "issue:44038", + "right": "issue:44297", + "accept": false, + "reason": "Qwen3-VL-Moe runtime bug vs tokenizer save metadata mismatch; not the same code path." + }, + { + "left": "issue:42907", + "right": "issue:43452", + "accept": false, + "reason": "Dequantized save failure vs gguf/AutoTokenizer loading breakage are different bugs." + }, + { + "left": "issue:44038", + "right": "issue:44295", + "accept": false, + "reason": "Model-specific generation/load issue vs position_ids buffer read error; unrelated." + }, + { + "left": "issue:39401", + "right": "issue:44488", + "accept": false, + "reason": "Tokenizer offset mapping bug vs inability to load a specific model checkpoint." + }, + { + "left": "issue:36296", + "right": "issue:43116", + "accept": false, + "reason": "Tensor-parallel training state bug vs example-script multi-label inference issue." + }, + { + "left": "issue:43824", + "right": "issue:44351", + "accept": false, + "reason": "Both are import errors, but for different missing symbols from different modules." + }, + { + "left": "issue:44038", + "right": "issue:44222", + "accept": false, + "reason": "Qwen3-VL-Moe bug vs FP8 save_pretrained bug; same broad area, different concrete failure." + }, + { + "left": "issue:43066", + "right": "issue:44779", + "accept": false, + "reason": "Different tokenizer regressions affecting different models and code paths." + }, + { + "left": "issue:39401", + "right": "issue:44521", + "accept": false, + "reason": "Tokenizer offset mapping bug vs multimodal chat-template assistant mask bug." + }, + { + "left": "issue:45216", + "right": "issue:45478", + "accept": false, + "reason": "Both mention Qwen3.5 and v5.4+, but one is save_pretrained corruption and the other is from_pretrained failure." + }, + { + "left": "issue:42757", + "right": "issue:44351", + "accept": false, + "reason": "Import errors from different packages: huggingface_hub vs transformers." + }, + { + "left": "issue:43122", + "right": "issue:44977", + "accept": false, + "reason": "Tokenization regression vs flash-attention generation failure; not the same bug." + }, + { + "left": "issue:43992", + "right": "issue:44704", + "accept": false, + "reason": "Encoder weight loading issue vs AutoProcessor kwargs propagation issue." + }, + { + "left": "issue:42907", + "right": "issue:44222", + "accept": false, + "reason": "Different save/pretrained bugs in different quantization formats and model families." + }, + { + "left": "issue:43784", + "right": "issue:45446", + "accept": false, + "reason": "Sentence-transformers import NameError vs flex_attention version check bug are unrelated." + }, + { + "left": "issue:43937", + "right": "issue:44485", + "accept": false, + "reason": "GenerationConfig validation problem vs RoPE implementation discussion; different concrete issues." + }, + { + "left": "issue:42222", + "right": "issue:43441", + "accept": false, + "reason": "vitpose breakage vs Ministral FlashAttention failure; unrelated model paths." + }, + { + "left": "issue:42371", + "right": "issue:43874", + "accept": false, + "reason": "TF32 API deprecation notice vs missing image-patch helper; not the same bug." + }, + { + "left": "issue:43976", + "right": "issue:45446", + "accept": false, + "reason": "Python version compatibility problem vs a specific PyTorch import/version check." + }, + { + "left": "issue:43493", + "right": "issue:45753", + "accept": false, + "reason": "SigLIP2 implementation discrepancy vs Qwen3_5 infinite-loop-on-image issue." + }, + { + "left": "issue:42831", + "right": "issue:45753", + "accept": false, + "reason": "FP8 accuracy issue vs multimodal generation loop are different failures." + }, + { + "left": "issue:44075", + "right": "issue:44936", + "accept": false, + "reason": "Optimizer argument handling vs trainer.evaluate after train; both trainer-related but not the same bug." + }, + { + "left": "issue:43116", + "right": "issue:43825", + "accept": false, + "reason": "Empty multi-label results vs stale translation-task error message; unrelated example/pipeline issues." + }, + { + "left": "issue:44222", + "right": "issue:45478", + "accept": false, + "reason": "Both involve Qwen3.5/MoE, but one is FP8 save_pretrained and the other is from_pretrained loading." + }, + { + "left": "issue:44038", + "right": "issue:45478", + "accept": false, + "reason": "Different Qwen model variants and different failure modes; not mergeable as one fix." + }, + { + "left": "issue:43116", + "right": "issue:43493", + "accept": false, + "reason": "Classification example bug vs SigLIP2 model discrepancy are unrelated." + }, + { + "left": "issue:43064", + "right": "issue:43673", + "accept": false, + "reason": "Trainer/FSDP optimizer-state bug vs GenerationMixin cache issue; different code paths." + }, + { + "left": "issue:45083", + "right": "issue:45446", + "accept": false, + "reason": "qwen3_omni_moe feature-length helper bug vs flex_attention import/version check bug." + }, + { + "left": "issue:43937", + "right": "issue:45563", + "accept": false, + "reason": "GLM-5 generation-config validation vs paged-generate warning text are unrelated." + }, + { + "left": "issue:33453", + "right": "issue:39401", + "accept": false, + "reason": "Generic tokenizer loading regression vs Qwen3 offset_mapping bug; not the same defect." + }, + { + "left": "issue:39401", + "right": "issue:44779", + "accept": false, + "reason": "Both are tokenizer regressions, but they affect different models and outputs, so not one concrete bug." + }, + { + "left": "issue:38617", + "right": "issue:45278", + "accept": false, + "reason": "45278 is a broad import-error umbrella, but 38617 is a specific missing-symbol report and not clearly the same underlying bug." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 38, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:22355", + "issue:29127", + "issue:38617", + "issue:39401", + "issue:41628", + "issue:42175", + "issue:42222", + "issue:42371", + "issue:42673", + "issue:43012", + "issue:43066", + "issue:43122", + "issue:43381", + "issue:43408", + "issue:43421", + "issue:43450", + "issue:43577", + "issue:43742", + "issue:43827", + "issue:43937", + "issue:43986", + "issue:44075", + "issue:44117", + "issue:44220", + "issue:44261", + "issue:44291", + "issue:44410", + "issue:44483", + "issue:44484", + "issue:44496", + "issue:44829", + "issue:44945", + "issue:44964", + "issue:44977", + "issue:45071", + "issue:45072", + "issue:45081", + "issue:45216", + "issue:45278", + "issue:45310", + "issue:45405", + "issue:45446", + "issue:45459", + "issue:45464", + "issue:45478", + "issue:45542", + "issue:45588", + "issue:45656" + ], + "soft_pairs": [ + "issue:44117|issue:45588", + "issue:22355|issue:29127", + "issue:43827|issue:45446", + "issue:43381|issue:44945", + "issue:43937|issue:44977", + "issue:43450|issue:43986", + "issue:45405|issue:45446", + "issue:39401|issue:43421", + "issue:44117|issue:45081", + "issue:42673|issue:45072", + "issue:38617|issue:41628", + "issue:44829|issue:45216", + "issue:43408|issue:43742", + "issue:42222|issue:45542", + "issue:44410|issue:45310", + "issue:44291|issue:45071", + "issue:42175|issue:42222", + "issue:44484|issue:45310", + "issue:43937|issue:44496", + "issue:44483|issue:45464", + "issue:44075|issue:45656", + "issue:43066|issue:43122", + "issue:44964|issue:45278", + "issue:43012|issue:43577", + "issue:38617|issue:44220", + "issue:42371|issue:44261", + "issue:42222|issue:45459", + "issue:44410|issue:45478", + "issue:44484|issue:45478" + ], + "prepared_review_unit_hash": "1394739772ea01ba1b9c4f5cdbe55d98fc530467a74e2c5eea30d8e351a0e513", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 29, + "serialized_chars": 36336, + "estimated_input_tokens": 9084, + "estimated_eval_tokens": 18424 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is highly heterogeneous: most items are unrelated model, tokenizer, optimizer, or import-regression bugs. None of the soft pairs look like true duplicates, so all soft edges are rejected. The closest thing to a canonical issue is the broad upgrade/import-regression report, but overall the cluster is not a clean duplicate set.", + "confidence": 0.33, + "canonical_issue_reason": "Issue 45278 is the broadest umbrella report in the set, covering multiple import failures after upgrading to 5.5.0, but the cluster is too mixed for a strong canonical duplicate.", + "canonical_pr_reason": null, + "best_issue_reason": "45278 is the most general, user-facing issue and the best fit if one must anchor the cluster, though it still only matches a subset of the items.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44117", + "right": "issue:45588", + "accept": false, + "reason": "Different failure modes: tokenizer mapping fallback vs flash-attention crash on sink-less models." + }, + { + "left": "issue:22355", + "right": "issue:29127", + "accept": false, + "reason": "Unrelated bugs: missing transformers.onnx module vs LayoutLMv3 error-message clarity." + }, + { + "left": "issue:43827", + "right": "issue:45446", + "accept": false, + "reason": "Docs/pipeline removal mismatch is unrelated to the flex_attention PyTorch version check bug." + }, + { + "left": "issue:43381", + "right": "issue:44945", + "accept": false, + "reason": "Gradient checkpointing in eval mode and incorrect pipeline-parallel output are different code paths." + }, + { + "left": "issue:43937", + "right": "issue:44977", + "accept": false, + "reason": "Both affect generation, but one is invalid GenerationConfig and the other is a Qwen3.5 flash-attention issue." + }, + { + "left": "issue:43450", + "right": "issue:43986", + "accept": false, + "reason": "Batched video processor shape bug is unrelated to AutoProcessor crashing without torchvision." + }, + { + "left": "issue:45405", + "right": "issue:45446", + "accept": false, + "reason": "PEFT version pinning and a flex_attention import/version check are unrelated." + }, + { + "left": "issue:39401", + "right": "issue:43421", + "accept": false, + "reason": "Tokenizer offset_mapping regression is different from post-processor not updating after runtime special-token changes." + }, + { + "left": "issue:44117", + "right": "issue:45081", + "accept": false, + "reason": "Tokenizer mapping None and Mistral regex patch backend_tokenizer errors are different bugs." + }, + { + "left": "issue:42673", + "right": "issue:45072", + "accept": false, + "reason": "VRAM leak across dataloader threads is unrelated to dtype mismatches in inference." + }, + { + "left": "issue:38617", + "right": "issue:41628", + "accept": false, + "reason": "Both are import errors, but they concern different missing symbols and different root causes." + }, + { + "left": "issue:44829", + "right": "issue:45216", + "accept": false, + "reason": "Flash-attention training degeneration is unrelated to a save_pretrained checkpoint regression." + }, + { + "left": "issue:43408", + "right": "issue:43742", + "accept": false, + "reason": "Model-type warning for sam3_tracker is unrelated to a MobileLLM loading key error." + }, + { + "left": "issue:42222", + "right": "issue:45542", + "accept": false, + "reason": "Broken vitpose models and missing TensorFlow backend from tensorboard-only install are unrelated." + }, + { + "left": "issue:44410", + "right": "issue:45310", + "accept": false, + "reason": "Missing Qwen3Next projections and Qwen3.5 MoE from_pretrained error are different model-specific loading bugs." + }, + { + "left": "issue:44291", + "right": "issue:45071", + "accept": false, + "reason": "init_empty_weights argument handling and PreTrainedConfig type checking are separate regressions." + }, + { + "left": "issue:42175", + "right": "issue:42222", + "accept": false, + "reason": "TensorFlow backend packaging issue and vitpose model breakage are unrelated." + }, + { + "left": "issue:44484", + "right": "issue:45310", + "accept": false, + "reason": "Chat/completions API failure is a different surface from Qwen3.5 MoE from_pretrained loading error." + }, + { + "left": "issue:43937", + "right": "issue:44496", + "accept": false, + "reason": "Invalid GenerationConfig and unrecognized model/config problems are different validation failures." + }, + { + "left": "issue:44483", + "right": "issue:45464", + "accept": false, + "reason": "These both touch chat/completions, but one is a request-acceptance bug and the other is streaming inference failure." + }, + { + "left": "issue:44075", + "right": "issue:45656", + "accept": false, + "reason": "SGD args ignored and deepspeed double-step are distinct optimizer bugs." + }, + { + "left": "issue:43066", + "right": "issue:43122", + "accept": false, + "reason": "Both are tokenizer regressions, but one is wrong decoder type and the other is changed tokenization output." + }, + { + "left": "issue:44964", + "right": "issue:45278", + "accept": false, + "reason": "A single model-loading failure is not the same as the broad multi-import regression report." + }, + { + "left": "issue:43012", + "right": "issue:43577", + "accept": false, + "reason": "Bfloat16 compilation warning and BLIP2 dtype remaining float32 are different dtype-related bugs." + }, + { + "left": "issue:38617", + "right": "issue:44220", + "accept": false, + "reason": "Missing import symbol and _torch_extract_fbank_features failure are unrelated." + }, + { + "left": "issue:42371", + "right": "issue:44261", + "accept": false, + "reason": "TF32 behavior warning and MLA rms_norm_eps precision mismatch are unrelated." + }, + { + "left": "issue:42222", + "right": "issue:45459", + "accept": false, + "reason": "Vitpose model breakage and protobuf-hidden tokenizer errors do not share the same underlying bug." + }, + { + "left": "issue:44410", + "right": "issue:45478", + "accept": false, + "reason": "Both mention Qwen models, but they are different loading failures with different model paths." + }, + { + "left": "issue:44484", + "right": "issue:45478", + "accept": false, + "reason": "Chat API request handling and Qwen3.5 MoE model loading are unrelated." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 39, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:38617", + "issue:41628", + "issue:42371", + "issue:42673", + "issue:42757", + "issue:42831", + "issue:43012", + "issue:43064", + "issue:43097", + "issue:43335", + "issue:43408", + "issue:43441", + "issue:43452", + "issue:43502", + "issue:43577", + "issue:43873", + "issue:43874", + "issue:43906", + "issue:43931", + "issue:43976", + "issue:43986", + "issue:43992", + "issue:44038", + "issue:44062", + "issue:44117", + "issue:44261", + "issue:44263", + "issue:44315", + "issue:44351", + "issue:44368", + "issue:44410", + "issue:44464", + "issue:44484", + "issue:44485", + "issue:44493", + "issue:44704", + "issue:44829", + "issue:44945", + "issue:44964", + "issue:44977", + "issue:44987", + "issue:45071", + "issue:45103", + "issue:45245", + "issue:45341", + "issue:45375", + "issue:45542", + "issue:45588" + ], + "soft_pairs": [ + "issue:43097|issue:44493", + "issue:43874|issue:44261", + "issue:44263|issue:44485", + "issue:43012|issue:43408", + "issue:43441|issue:44829", + "issue:41628|issue:45071", + "issue:45071|issue:45542", + "issue:42673|issue:44977", + "issue:42371|issue:42831", + "issue:42673|issue:43906", + "issue:43986|issue:44484", + "issue:44117|issue:44987", + "issue:43097|issue:43452", + "issue:43931|issue:44038", + "issue:43335|issue:45071", + "issue:43408|issue:44368", + "issue:43976|issue:44945", + "issue:43874|issue:44263", + "issue:44829|issue:45245", + "issue:38617|issue:42757", + "issue:43064|issue:44315", + "issue:44464|issue:45071", + "issue:44410|issue:45375", + "issue:44062|issue:44117", + "issue:43097|issue:45588", + "issue:42673|issue:43577", + "issue:38617|issue:44351", + "issue:44704|issue:45103", + "issue:44038|issue:44410", + "issue:43931|issue:44410", + "issue:43986|issue:44117", + "issue:44261|issue:44485", + "issue:44964|issue:45341", + "issue:43502|issue:43992", + "issue:43097|issue:43873" + ], + "prepared_review_unit_hash": "05aff5baaaf86eb57dcafb34aa43f2b5ba3838cc3897b3275706e720027949a6", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 35, + "serialized_chars": 38166, + "estimated_input_tokens": 9542, + "estimated_eval_tokens": 19340 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Loose similarity cluster of many unrelated Transformers issues (import errors, model-specific load bugs, warnings, and training regressions). None of the soft pairs look like true duplicates.", + "confidence": 0.94, + "canonical_issue_reason": "No single canonical issue: the cluster spans multiple unrelated failure modes and model families, so one issue does not cleanly subsume the others.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 43097 is the closest representative anchor because it is a broad v5 regression and has inbound references, but it still does not explain the rest of the cluster.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43097", + "right": "issue:44493", + "accept": false, + "reason": "Removed API warning vs unexpected position-id key warning; different regression and code path." + }, + { + "left": "issue:43874", + "right": "issue:44261", + "accept": false, + "reason": "Different models and failures: missing image-patch method vs missing rms_norm_eps in config." + }, + { + "left": "issue:44263", + "right": "issue:44485", + "accept": false, + "reason": "GlmMoeDsaIndexer split behavior and GLM-5 RoPE implementation are distinct bugs." + }, + { + "left": "issue:43012", + "right": "issue:43408", + "accept": false, + "reason": "Float32 precision warning during compile is unrelated to the sam3 tracker/video model-type warning." + }, + { + "left": "issue:43441", + "right": "issue:44829", + "accept": false, + "reason": "Both mention flash attention, but one is a load-time failure and the other is degenerate training; not the same bug." + }, + { + "left": "issue:41628", + "right": "issue:45071", + "accept": false, + "reason": "Missing AutoImageProcessor import and PretrainedConfig type-checking breakage are unrelated issues." + }, + { + "left": "issue:45071", + "right": "issue:45542", + "accept": false, + "reason": "Config type-checking regression vs undefined TF backend from tensorboard-only install; different causes." + }, + { + "left": "issue:42673", + "right": "issue:44977", + "accept": false, + "reason": "VRAM leak in multi-threaded dataloading vs generation failure with flash-attention; distinct symptoms and paths." + }, + { + "left": "issue:42371", + "right": "issue:42831", + "accept": false, + "reason": "TF32 API-setting guidance and FineGrainedFP8 accuracy loss are different concerns." + }, + { + "left": "issue:42673", + "right": "issue:43906", + "accept": false, + "reason": "VRAM leak report vs reproduction of another issue; no concrete shared defect." + }, + { + "left": "issue:43986", + "right": "issue:44484", + "accept": false, + "reason": "AutoProcessor/video-model crash without torchvision is unrelated to save_pretrained shard sizing." + }, + { + "left": "issue:44117", + "right": "issue:44987", + "accept": false, + "reason": "TOKENIZER_MAPPING_NAMES None handling and a model load failure for physical-intelligence/fast are unrelated." + }, + { + "left": "issue:43097", + "right": "issue:43452", + "accept": false, + "reason": "Config removal warning and gguf_file loading breakage are different failures." + }, + { + "left": "issue:43931", + "right": "issue:44038", + "accept": false, + "reason": "Qwen3-VL weight mismatch and Qwen3-VL-Moe loading bug are different model-specific issues." + }, + { + "left": "issue:43335", + "right": "issue:45071", + "accept": false, + "reason": "SwitchTransformers sparse-layer creation bug and config type-checking regression are unrelated." + }, + { + "left": "issue:43408", + "right": "issue:44368", + "accept": false, + "reason": "Different warnings on different models: sam3 tracker/type mismatch vs tie_word_embeddings guidance." + }, + { + "left": "issue:43976", + "right": "issue:44945", + "accept": false, + "reason": "Python-version compatibility problem vs incorrect output under pipeline parallelism; not the same defect." + }, + { + "left": "issue:43874", + "right": "issue:44263", + "accept": false, + "reason": "Missing image-patch method and torch.split return-value issue are separate GLM bugs." + }, + { + "left": "issue:44829", + "right": "issue:45245", + "accept": false, + "reason": "FlashAttention training degeneration and category-count runtime limit are unrelated." + }, + { + "left": "issue:38617", + "right": "issue:42757", + "accept": false, + "reason": "Both are import errors, but they involve different missing symbols from different packages." + }, + { + "left": "issue:43064", + "right": "issue:44315", + "accept": false, + "reason": "FSDP2/PEFT optimizer-state mismatch and Liger Kernel not applied with model_init are separate training-stack issues." + }, + { + "left": "issue:44464", + "right": "issue:45071", + "accept": false, + "reason": "Compiled-forward chunked generation inconsistency is unrelated to PretrainedConfig type checking." + }, + { + "left": "issue:44410", + "right": "issue:45375", + "accept": false, + "reason": "Missing projections in qwen3next and missing deepstack_visual_indexes in config are distinct model/config bugs." + }, + { + "left": "issue:44062", + "right": "issue:44117", + "accept": false, + "reason": "AddedToken keyword duplication and TOKENIZER_MAPPING_NAMES None are unrelated tokenizer problems." + }, + { + "left": "issue:43097", + "right": "issue:45588", + "accept": false, + "reason": "Embedding-tying warning and flash_attention.py s_aux=None crash are different code paths." + }, + { + "left": "issue:42673", + "right": "issue:43577", + "accept": false, + "reason": "VRAM leak under dataloader threads and Blip2 dtype staying float32 are unrelated." + }, + { + "left": "issue:38617", + "right": "issue:44351", + "accept": false, + "reason": "Different missing imports from different modules; same broad category only, not the same bug." + }, + { + "left": "issue:44704", + "right": "issue:45103", + "accept": false, + "reason": "Kwarg forwarding to cached_file and auto_docstring AttributeError are unrelated utility bugs." + }, + { + "left": "issue:44038", + "right": "issue:44410", + "accept": false, + "reason": "Qwen3-VL-Moe loading issue and qwen3next missing projections are different model failures." + }, + { + "left": "issue:43931", + "right": "issue:44410", + "accept": false, + "reason": "Weight-shape mismatch for Qwen3-VL-30B and missing projections in qwen3next are not the same defect." + }, + { + "left": "issue:43986", + "right": "issue:44117", + "accept": false, + "reason": "Video-model torchvision crash and tokenizer-mapping None handling are unrelated." + }, + { + "left": "issue:44261", + "right": "issue:44485", + "accept": false, + "reason": "Missing rms_norm_eps in MLA q_a_layernorm and GLM-5 RoPE implementation are different issues." + }, + { + "left": "issue:44964", + "right": "issue:45341", + "accept": false, + "reason": "Phi-4 multimodal load failure and a testing_utils bug have no shared underlying cause." + }, + { + "left": "issue:43502", + "right": "issue:43992", + "accept": false, + "reason": "Unexpected API requests with local_files_only and missing embed_tokens.weight on UMT5Encoder are unrelated." + }, + { + "left": "issue:43097", + "right": "issue:43873", + "accept": false, + "reason": "Embedding-tying deprecation warning and quantization offloading behavior are different problems." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 40, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:36683", + "issue:38617", + "issue:42617", + "issue:42757", + "issue:42886", + "issue:42907", + "issue:42994", + "issue:43012", + "issue:43064", + "issue:43066", + "issue:43097", + "issue:43208", + "issue:43299", + "issue:43352", + "issue:43381", + "issue:43475", + "issue:43502", + "issue:43526", + "issue:43531", + "issue:43550", + "issue:43618", + "issue:43761", + "issue:43824", + "issue:43825", + "issue:43867", + "issue:43901", + "issue:43906", + "issue:43937", + "issue:43986", + "issue:43992", + "issue:44038", + "issue:44220", + "issue:44230", + "issue:44368", + "issue:44380", + "issue:44483", + "issue:44484", + "issue:44514", + "issue:44655", + "issue:44683", + "issue:45216", + "issue:45290", + "issue:45341", + "issue:45372", + "issue:45381", + "issue:45561", + "issue:45588", + "issue:45753" + ], + "soft_pairs": [ + "issue:43066|issue:43906", + "issue:43986|issue:45372", + "issue:43825|issue:44655", + "issue:43986|issue:45290", + "issue:45341|issue:45561", + "issue:43986|issue:45341", + "issue:43097|issue:43901", + "issue:42617|issue:45341", + "issue:43992|issue:44484", + "issue:45341|issue:45588", + "issue:43012|issue:44368", + "issue:42886|issue:43381", + "issue:43012|issue:43550", + "issue:43208|issue:43550", + "issue:42907|issue:45753", + "issue:42907|issue:44655", + "issue:43531|issue:45341", + "issue:43208|issue:43526", + "issue:44483|issue:44514", + "issue:43012|issue:44683", + "issue:42757|issue:43824", + "issue:43618|issue:43761", + "issue:44483|issue:45290", + "issue:43064|issue:44380", + "issue:44038|issue:44230", + "issue:43299|issue:43867", + "issue:38617|issue:43502", + "issue:44483|issue:45381", + "issue:43901|issue:44655", + "issue:36683|issue:43475", + "issue:44220|issue:44655", + "issue:43867|issue:45216", + "issue:43352|issue:43937", + "issue:42994|issue:44655", + "issue:43066|issue:45381" + ], + "prepared_review_unit_hash": "702b219bd17b29be1bf6a7c13d7b0c1a056e0507f0ce983b37cffe09e6e2460c", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 35, + "serialized_chars": 37893, + "estimated_input_tokens": 9474, + "estimated_eval_tokens": 19204 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly a loose collection of unrelated Transformers bug reports. The strongest duplicate candidate is the pair of CLIP v5 regression issues, but the titles alone do not prove they are the same underlying bug, so it should remain unmatched unless the bodies show the same regression path.", + "confidence": 0.58, + "canonical_issue_reason": "issue:43761 is the cleaner representative of the CLIP regression because it states the concrete `CLIPVisionModel.forward` symptom and the v5 hidden-state failure more explicitly; issue:43618 reads like a narrower symptom report.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43761 is the best issue to keep as the representative bug report because it is more specific and directly describes the observed regression path, while issue:43618 is less explicit.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43066", + "right": "issue:43906", + "accept": false, + "reason": "Both are model-loading/tokenization-related, but they point to different failures and different code paths." + }, + { + "left": "issue:43986", + "right": "issue:45372", + "accept": false, + "reason": "Different root causes: missing `torchvision` versus a `ReasoningEffort` import failure." + }, + { + "left": "issue:43825", + "right": "issue:44655", + "accept": false, + "reason": "`pipeline()` error messaging and `save_pretrained()` for Pipeline objects are different functionalities." + }, + { + "left": "issue:43986", + "right": "issue:45290", + "accept": false, + "reason": "Processor loading without torchvision is unrelated to `apply_chat_template` crashing on tool-call messages." + }, + { + "left": "issue:45341", + "right": "issue:45561", + "accept": false, + "reason": "A vague testing-utils bug and an xdist race on `captured_info.txt` are not clearly the same bug." + }, + { + "left": "issue:43986", + "right": "issue:45341", + "accept": false, + "reason": "Only superficial overlap in testing/utils wording; the symptoms and code paths differ." + }, + { + "left": "issue:43097", + "right": "issue:43901", + "accept": false, + "reason": "A v5 API removal complaint and a docs mismatch for `return_all_scores` are not the same issue." + }, + { + "left": "issue:42617", + "right": "issue:45341", + "accept": false, + "reason": "A 3D parallel script failure and a testing-utils race are unrelated." + }, + { + "left": "issue:43992", + "right": "issue:44484", + "accept": false, + "reason": "Missing checkpoint weights on load and the `max_shard_size` default in save_pretrained are different bugs." + }, + { + "left": "issue:45341", + "right": "issue:45588", + "accept": false, + "reason": "Testing-utils race and a flash-attention crash in sink-less models are different code paths." + }, + { + "left": "issue:43012", + "right": "issue:44368", + "accept": false, + "reason": "A PyTorch precision warning during compilation is not the same as a `tie_word_embeddings` config warning." + }, + { + "left": "issue:42886", + "right": "issue:43381", + "accept": false, + "reason": "Offline cache loading and gradient-checkpointing eval-mode behavior are unrelated." + }, + { + "left": "issue:43012", + "right": "issue:43550", + "accept": false, + "reason": "One is a warning, the other is a `torch.compile` SDPA failure; different bugs." + }, + { + "left": "issue:43208", + "right": "issue:43550", + "accept": false, + "reason": "xLSTM training bugs and Bamba SDPA compile failures do not match." + }, + { + "left": "issue:42907", + "right": "issue:45753", + "accept": false, + "reason": "Saving dequantized models and an infinite-loop image bug are different problems." + }, + { + "left": "issue:42907", + "right": "issue:44655", + "accept": false, + "reason": "Quantized/dequantized model saving is unrelated to Pipeline serialization." + }, + { + "left": "issue:43531", + "right": "issue:45341", + "accept": false, + "reason": "Qwen3-MoE sliding-window behavior and testing-utils bugs do not share an underlying failure." + }, + { + "left": "issue:43208", + "right": "issue:43526", + "accept": false, + "reason": "xLSTM training issues and BEiT label-reduction behavior are unrelated." + }, + { + "left": "issue:44483", + "right": "issue:44514", + "accept": false, + "reason": "Chat-completions API validation and batched `apply_chat_template` crashes are different failures." + }, + { + "left": "issue:43012", + "right": "issue:44683", + "accept": false, + "reason": "A bfloat16 warning is unrelated to compiled flex-attention failures on newer torch versions." + }, + { + "left": "issue:42757", + "right": "issue:43824", + "accept": false, + "reason": "Both are import errors, but they involve different missing symbols and different packages." + }, + { + "left": "issue:43618", + "right": "issue:43761", + "accept": false, + "reason": "They are likely related CLIP v5 regressions, but the titles alone do not establish they are definitely the same bug." + }, + { + "left": "issue:44483", + "right": "issue:45290", + "accept": false, + "reason": "Chat-completions request validation and `apply_chat_template` tool-call crashes are different issues." + }, + { + "left": "issue:43064", + "right": "issue:44380", + "accept": false, + "reason": "Trainer/FSDP optimizer-state corruption is unrelated to GPT2 attention scaling being ignored." + }, + { + "left": "issue:44038", + "right": "issue:44230", + "accept": false, + "reason": "A general Qwen3-VL-MoE breakage and an fp8 support request are not the same bug." + }, + { + "left": "issue:43299", + "right": "issue:43867", + "accept": false, + "reason": "Qwen3VL-MoE loading breakage and sorted-state_dict load errors are different issues." + }, + { + "left": "issue:38617", + "right": "issue:43502", + "accept": false, + "reason": "An import failure and accidental API requests despite `local_files_only=True` are unrelated." + }, + { + "left": "issue:44483", + "right": "issue:45381", + "accept": false, + "reason": "Different Qwen2.5-VL failures with different symptoms and code paths." + }, + { + "left": "issue:43901", + "right": "issue:44655", + "accept": false, + "reason": "Pipeline docs mismatch and Pipeline save_pretrained failure are different problems." + }, + { + "left": "issue:36683", + "right": "issue:43475", + "accept": false, + "reason": "Gemma3 config missing `vocab_size` and SAM3 output missing `fpn_position_embeddings` are unrelated attribute errors." + }, + { + "left": "issue:44220", + "right": "issue:44655", + "accept": false, + "reason": "Feature-extraction helper behavior and Pipeline serialization do not share an underlying bug." + }, + { + "left": "issue:43867", + "right": "issue:45216", + "accept": false, + "reason": "A load error tied to state_dict ordering is not the same as a Qwen3.5 save_pretrained checkpoint regression." + }, + { + "left": "issue:43352", + "right": "issue:43937", + "accept": false, + "reason": "Nemotron Flash Attention support and GLM-5 generation-config validation are unrelated." + }, + { + "left": "issue:42994", + "right": "issue:44655", + "accept": false, + "reason": "Quantized model saving and Pipeline saving are different serialization paths and bugs." + }, + { + "left": "issue:43066", + "right": "issue:45381", + "accept": false, + "reason": "Wrong tokenizer decoder type and Qwen2.5-VL video position-id issues are unrelated." + } + ] + }, + "evaluator_result": { + "accept": false, + "feedback": "Most of the rejects are reasonably conservative, but the one accepted edge is a bit too strong given only the packet titles: issue:43618 (\u2018CLIPOutput attentions is no longer assigned\u2019) and issue:43761 (\u2018CLIPVisionModel.forward returns hidden_states=None even when output_hidden_states=True\u2019) may be related, but the output does not establish they are definitely the same bug. I\u2019d want that verdict softened to \u2018likely related\u2019 or left unmatched unless the bodies show the same regression path. The cluster summary also slightly overcommits by calling this the \u2018only credible duplicate edge\u2019 instead of \u2018the strongest possible duplicate candidate.\u2019" + }, + "evaluator_used": true, + "retried": true, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 41, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:28282", + "issue:33453", + "issue:36010", + "issue:36296", + "issue:36683", + "issue:41669", + "issue:42757", + "issue:42890", + "issue:43064", + "issue:43097", + "issue:43299", + "issue:43352", + "issue:43441", + "issue:43525", + "issue:43575", + "issue:43618", + "issue:43756", + "issue:43867", + "issue:43881", + "issue:43906", + "issue:43976", + "issue:43992", + "issue:44077", + "issue:44117", + "issue:44242", + "issue:44246", + "issue:44263", + "issue:44355", + "issue:44360", + "issue:44380", + "issue:44484", + "issue:44485", + "issue:44492", + "issue:44655", + "issue:44855", + "issue:44857", + "issue:44938", + "issue:44977", + "issue:45216", + "issue:45310", + "issue:45357", + "issue:45375", + "issue:45440", + "issue:45478", + "issue:45479", + "issue:45561", + "issue:45588", + "issue:45753" + ], + "soft_pairs": [ + "issue:36010|issue:42757", + "issue:44263|issue:45440", + "issue:43756|issue:44485", + "issue:36683|issue:43525", + "issue:43618|issue:45216", + "issue:43906|issue:45753", + "issue:36683|issue:44242", + "issue:36683|issue:43881", + "issue:43299|issue:45753", + "issue:43352|issue:43976", + "issue:41669|issue:44246", + "issue:43441|issue:44380", + "issue:43992|issue:45310", + "issue:43352|issue:44380", + "issue:44380|issue:45479", + "issue:43867|issue:45375", + "issue:43064|issue:44492", + "issue:43618|issue:44380", + "issue:43352|issue:45478", + "issue:44355|issue:44855", + "issue:36683|issue:44117", + "issue:42890|issue:45561", + "issue:43097|issue:43352", + "issue:43867|issue:45357", + "issue:43976|issue:44938", + "issue:43992|issue:45478", + "issue:44360|issue:45440", + "issue:43064|issue:44655", + "issue:36296|issue:43575", + "issue:44077|issue:44380", + "issue:43618|issue:44857", + "issue:43867|issue:44484", + "issue:43618|issue:45588", + "issue:28282|issue:33453", + "issue:43618|issue:44977" + ], + "prepared_review_unit_hash": "6e9ce916a630d98b7cc23a92a8105b929410b3cc234f90d5f37625c08c679928", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 35, + "serialized_chars": 37796, + "estimated_input_tokens": 9449, + "estimated_eval_tokens": 19154 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is largely noisy: several issues share broad keywords like configs, loading, flash-attention, or import/runtime failures, but they do not look like the same underlying bug or change. No pull requests are present.", + "confidence": 0.59, + "canonical_issue_reason": "issue:36683 is the best representative of the small config-attribute/load-path subtheme and is the most connected issue in the set, but the overall cluster is still heterogeneous.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:36683 is the strongest single issue to keep as the cluster representative because it is active and central to the few genuinely related config-attribute reports.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:36010", + "right": "issue:42757", + "accept": false, + "reason": "Both are import errors, but they involve different missing symbols from different packages and different failure modes." + }, + { + "left": "issue:44263", + "right": "issue:45440", + "accept": false, + "reason": "These are unrelated model-path bugs: one is a torch.split/indexer issue, the other is a DeepSeekV3 implementation divergence." + }, + { + "left": "issue:43756", + "right": "issue:44485", + "accept": false, + "reason": "Both mention RoPE, but they concern different models and distinct implementation problems." + }, + { + "left": "issue:36683", + "right": "issue:43525", + "accept": false, + "reason": "Both are missing-config-attribute reports, but on different models and different fields, so they are not the same bug." + }, + { + "left": "issue:43618", + "right": "issue:45216", + "accept": false, + "reason": "One is about CLIPOutput attentions no longer being assigned; the other is a Qwen3.5 save_pretrained regression." + }, + { + "left": "issue:43906", + "right": "issue:45753", + "accept": false, + "reason": "The first is an isolated reproduction of another issue; the second is a specific-image infinite loop in Qwen3_5, which is a different symptom." + }, + { + "left": "issue:36683", + "right": "issue:44242", + "accept": false, + "reason": "A missing Gemma3Config attribute is unrelated to load-balancing loss handling when output_router_logits=False." + }, + { + "left": "issue:36683", + "right": "issue:43881", + "accept": false, + "reason": "Different models and different loading failures; only the generic \"loading failed\" theme overlaps." + }, + { + "left": "issue:43299", + "right": "issue:45753", + "accept": false, + "reason": "Both involve Qwen variants, but one is a v5.dev0 loading break and the other is an inference infinite-loop bug." + }, + { + "left": "issue:43352", + "right": "issue:43976", + "accept": false, + "reason": "A FlashAttention support limitation is unrelated to the Python-version compatibility failure." + }, + { + "left": "issue:41669", + "right": "issue:44246", + "accept": false, + "reason": "Import-time performance and occasional import slowness are too broad here; these do not describe the same code path or bug." + }, + { + "left": "issue:43441", + "right": "issue:44380", + "accept": false, + "reason": "Both touch attention backends, but they are different model-specific bugs with different root causes." + }, + { + "left": "issue:43992", + "right": "issue:45310", + "accept": false, + "reason": "Missing embed_tokens.weight during UMT5Encoder loading is unrelated to the Qwen3.5 MoE from_pretrained regression." + }, + { + "left": "issue:43352", + "right": "issue:44380", + "accept": false, + "reason": "FlashAttention support in Nemotron and GPT2 attention scaling being ignored are separate backend/model issues." + }, + { + "left": "issue:44380", + "right": "issue:45479", + "accept": false, + "reason": "Attention scaling behavior and a degenerate zero-loss classification bug are unrelated." + }, + { + "left": "issue:43867", + "right": "issue:45375", + "accept": false, + "reason": "A sorted-state_dict loading error is not the same as a missing deepstack_visual_indexes config field." + }, + { + "left": "issue:43064", + "right": "issue:44492", + "accept": false, + "reason": "Optimizer-state corruption during FSDP2/PEFT training is unrelated to a cache-strategy typo." + }, + { + "left": "issue:43618", + "right": "issue:44380", + "accept": false, + "reason": "These affect different models and different attention-related behaviors." + }, + { + "left": "issue:43352", + "right": "issue:45478", + "accept": false, + "reason": "Both are Qwen/Nemotron-era model bugs, but the concrete failure modes are unrelated." + }, + { + "left": "issue:44355", + "right": "issue:44855", + "accept": false, + "reason": "Compiled Python file errors and a Python 3.13 torch.jit.script parsing issue are distinct import/runtime problems." + }, + { + "left": "issue:36683", + "right": "issue:44117", + "accept": false, + "reason": "A missing model config attribute is not the same as TOKENIZER_MAPPING_NAMES returning None." + }, + { + "left": "issue:42890", + "right": "issue:45561", + "accept": false, + "reason": "Both are test flakiness, but one is missing set_seed and the other is an xdist race on a captured file." + }, + { + "left": "issue:43097", + "right": "issue:43352", + "accept": false, + "reason": "A removed config option and a FlashAttention support issue are unrelated." + }, + { + "left": "issue:43867", + "right": "issue:45357", + "accept": false, + "reason": "These both concern saving/loading, but one is a state_dict ordering problem and the other is incorrect visual encoder keys." + }, + { + "left": "issue:43976", + "right": "issue:44938", + "accept": false, + "reason": "Both are Python-version compatibility reports, but they target different versions and likely different breakages." + }, + { + "left": "issue:43992", + "right": "issue:45478", + "accept": false, + "reason": "A missing embedding weight during load is unrelated to a Qwen3.5 MoE from_pretrained error." + }, + { + "left": "issue:44360", + "right": "issue:45440", + "accept": false, + "reason": "A missing ReLU in the DSA indexer is unrelated to DeepSeekV3 implementation divergence." + }, + { + "left": "issue:43064", + "right": "issue:44655", + "accept": false, + "reason": "Training optimizer-state bugs and saving Pipeline objects are different code paths." + }, + { + "left": "issue:36296", + "right": "issue:43575", + "accept": false, + "reason": "Tensor-parallel training bugs and TP-related OOM during model load are not the same defect." + }, + { + "left": "issue:44077", + "right": "issue:44380", + "accept": false, + "reason": "An optional post_init restriction in patchtsmixer has nothing to do with GPT2 attention scaling." + }, + { + "left": "issue:43618", + "right": "issue:44857", + "accept": false, + "reason": "Missing CLIPOutput attentions and a float16 AMP CUDA crash are unrelated bugs." + }, + { + "left": "issue:43867", + "right": "issue:44484", + "accept": false, + "reason": "State_dict sorting failures and the max_shard_size default question are not the same issue." + }, + { + "left": "issue:43618", + "right": "issue:45588", + "accept": false, + "reason": "The CLIPOutput attentions regression and the s_aux=None flash_attention crash affect different code paths." + }, + { + "left": "issue:28282", + "right": "issue:33453", + "accept": false, + "reason": "AutoModel missing PyTorch and tokenizer-loading regression are completely different failures." + }, + { + "left": "issue:43618", + "right": "issue:44977", + "accept": false, + "reason": "These are unrelated: one concerns a missing attention field, the other is a Qwen3.5 flash-attention generation bug." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 42, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:30333", + "issue:30990", + "issue:31795", + "issue:33453", + "issue:34689", + "issue:36296", + "issue:38617", + "issue:42548", + "issue:42757", + "issue:42831", + "issue:42890", + "issue:43010", + "issue:43299", + "issue:43502", + "issue:43576", + "issue:43723", + "issue:43824", + "issue:43827", + "issue:43874", + "issue:43901", + "issue:43976", + "issue:44062", + "issue:44075", + "issue:44112", + "issue:44222", + "issue:44230", + "issue:44315", + "issue:44351", + "issue:44355", + "issue:44360", + "issue:44368", + "issue:44387", + "issue:44485", + "issue:44556", + "issue:44655", + "issue:44749", + "issue:44938", + "issue:44961", + "issue:44964", + "issue:45071", + "issue:45125", + "issue:45278", + "issue:45341", + "issue:45375", + "issue:45468", + "issue:45478", + "issue:45600", + "issue:45656" + ], + "soft_pairs": [ + "issue:43901|issue:44938", + "issue:42890|issue:44964", + "issue:44655|issue:44938", + "issue:42890|issue:44112", + "issue:42890|issue:45600", + "issue:43502|issue:44062", + "issue:42890|issue:45341", + "issue:43010|issue:44075", + "issue:42548|issue:44351", + "issue:44222|issue:45125", + "issue:43502|issue:44485", + "issue:44355|issue:44749", + "issue:45125|issue:45478", + "issue:30990|issue:31795", + "issue:45125|issue:45375", + "issue:43010|issue:45278", + "issue:44387|issue:44556", + "issue:30333|issue:33453", + "issue:43723|issue:44556", + "issue:43010|issue:44368", + "issue:43502|issue:44222", + "issue:44315|issue:44360", + "issue:43502|issue:44961", + "issue:44355|issue:44655", + "issue:42831|issue:43874", + "issue:30333|issue:36296", + "issue:44355|issue:45071", + "issue:44222|issue:44230", + "issue:43827|issue:44961", + "issue:43010|issue:45656", + "issue:33453|issue:34689", + "issue:43299|issue:44556", + "issue:42757|issue:43576", + "issue:33453|issue:36296", + "issue:43576|issue:45071", + "issue:43576|issue:43824", + "issue:43010|issue:43976", + "issue:44961|issue:45468", + "issue:31795|issue:38617" + ], + "prepared_review_unit_hash": "0dfd9c2b3bc8ef5417dc950ce3be516006b9d6124d60a8d03096a8fd7b799427", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 39, + "serialized_chars": 38861, + "estimated_input_tokens": 9716, + "estimated_eval_tokens": 19688 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "Mostly a mixed bag of unrelated bug reports and documentation regressions around Transformers v5; only one soft edge looks like a plausible duplicate pair. I\u2019d use the broad v5 import-regression report as the representative issue, but there is no PR in this cluster.", + "confidence": 0.42, + "canonical_issue_reason": "issue:45278 is the broadest umbrella report here ('Many import errors after update from 4.57.0 to 5.5.0') and best represents the general v5 import/load regression theme, even though the cluster is not a set of true duplicates.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:45278 is the most general and representative issue among the mostly unrelated reports; it captures the shared version-upgrade breakage theme better than the narrower model-specific or doc-only items.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43901", + "right": "issue:44938", + "accept": false, + "reason": "Docs about a removed pipeline argument and a Python 3.14 import/load failure are different problems." + }, + { + "left": "issue:42890", + "right": "issue:44964", + "accept": false, + "reason": "Seed-setting flakiness in a SAM HQ test is unrelated to Phi-4 multimodal model loading." + }, + { + "left": "issue:44655", + "right": "issue:44938", + "accept": false, + "reason": "Saving Pipeline objects and Python 3.14 import failure are not the same code-path bug." + }, + { + "left": "issue:42890", + "right": "issue:44112", + "accept": false, + "reason": "Test nondeterminism from missing set_seed() is separate from a GraniteSpeech CI device-override failure." + }, + { + "left": "issue:42890", + "right": "issue:45600", + "accept": false, + "reason": "A flaky integration test is unrelated to stale auto_mappings references causing CI breakage." + }, + { + "left": "issue:43502", + "right": "issue:44062", + "accept": false, + "reason": "local_files_only leaking network requests and AddedToken keyword duplication are distinct loader/tokenizer issues." + }, + { + "left": "issue:42890", + "right": "issue:45341", + "accept": false, + "reason": "Both are test-related, but they describe different failures and different fixes." + }, + { + "left": "issue:43010", + "right": "issue:44075", + "accept": false, + "reason": "Cache update/no_grad behavior is unrelated to SGD optimizer arguments being ignored." + }, + { + "left": "issue:42548", + "right": "issue:44351", + "accept": false, + "reason": "Both are import errors, but for different missing symbols and likely different fixes." + }, + { + "left": "issue:44222", + "right": "issue:45125", + "accept": false, + "reason": "FP8 save_pretrained on MoE models is not the same as missing _tp_plan for tensor parallelism." + }, + { + "left": "issue:43502", + "right": "issue:44485", + "accept": false, + "reason": "local_files_only network leakage and GLM-5 RoPE discussion are unrelated." + }, + { + "left": "issue:44355", + "right": "issue:44749", + "accept": false, + "reason": "Compiled Python file errors are unrelated to a slowdown after upgrading Transformers." + }, + { + "left": "issue:45125", + "right": "issue:45478", + "accept": true, + "reason": "Both point to the same Qwen3.5 MoE loading failure path in v5, plausibly caused by missing tensor-parallel support." + }, + { + "left": "issue:30990", + "right": "issue:31795", + "accept": false, + "reason": "SentenceTransformer loading issues and model.forward documentation confusion are unrelated." + }, + { + "left": "issue:45125", + "right": "issue:45375", + "accept": false, + "reason": "Missing _tp_plan and a missing deepstack_visual_indexes config field are different model-configuration bugs." + }, + { + "left": "issue:43010", + "right": "issue:45278", + "accept": false, + "reason": "Cache no_grad/update behavior is unrelated to broad import errors after upgrading to v5.5.0." + }, + { + "left": "issue:44387", + "right": "issue:44556", + "accept": false, + "reason": "Int4 quantization memory regression and checkpoint reload failure after upgrade are different bugs." + }, + { + "left": "issue:30333", + "right": "issue:33453", + "accept": false, + "reason": "MLflow reporting on failed training jobs and tokenizer loading regression are unrelated." + }, + { + "left": "issue:43723", + "right": "issue:44556", + "accept": false, + "reason": "Tokenizer loading in v5 and reloading old checkpoints after upgrade are related broadly, but not the same concrete bug." + }, + { + "left": "issue:43010", + "right": "issue:44368", + "accept": false, + "reason": "Cache update/no_grad behavior is unrelated to a tie_word_embeddings warning during Qwen LoRA fine-tuning." + }, + { + "left": "issue:43502", + "right": "issue:44222", + "accept": false, + "reason": "local_files_only network calls and FP8 MoE save_pretrained are different issues." + }, + { + "left": "issue:44315", + "right": "issue:44360", + "accept": false, + "reason": "Liger Kernel not applying with model_init is unrelated to the DSA indexer ReLU discussion." + }, + { + "left": "issue:43502", + "right": "issue:44961", + "accept": false, + "reason": "local_files_only network leakage and a one-word issue title do not describe the same bug." + }, + { + "left": "issue:44355", + "right": "issue:44655", + "accept": false, + "reason": "Compiled Python file errors and inability to save Pipeline objects are different failure modes." + }, + { + "left": "issue:42831", + "right": "issue:43874", + "accept": false, + "reason": "FP8 accuracy regression and missing image-processor method are unrelated." + }, + { + "left": "issue:30333", + "right": "issue:36296", + "accept": false, + "reason": "MLflow reporting failure and tensor-parallel training bug are different training-time issues." + }, + { + "left": "issue:44355", + "right": "issue:45071", + "accept": false, + "reason": "Compiled file loading issues and PretrainedConfig type-checking breakage are unrelated." + }, + { + "left": "issue:44222", + "right": "issue:44230", + "accept": false, + "reason": "Both involve FP8 MoE, but one is save_pretrained and the other is inference support; not the same concrete bug." + }, + { + "left": "issue:43827", + "right": "issue:44961", + "accept": false, + "reason": "Pipeline docs references and a one-word title are unrelated." + }, + { + "left": "issue:43010", + "right": "issue:45656", + "accept": false, + "reason": "Cache update decoration and optimizer double-step behavior under deepspeed are different code paths." + }, + { + "left": "issue:33453", + "right": "issue:34689", + "accept": false, + "reason": "Tokenizer loading regression and Llama 3.2 Vision model loading failure are separate model-loading bugs." + }, + { + "left": "issue:43299", + "right": "issue:44556", + "accept": false, + "reason": "Qwen3VL MoE loading failure and old-checkpoint reload failure are both upgrade-related, but not the same underlying bug." + }, + { + "left": "issue:42757", + "right": "issue:43576", + "accept": false, + "reason": "Missing huggingface_hub export and a broken transformers env command are unrelated." + }, + { + "left": "issue:43576", + "right": "issue:45071", + "accept": false, + "reason": "env command breakage and PretrainedConfig type checking are unrelated." + }, + { + "left": "issue:43576", + "right": "issue:43824", + "accept": false, + "reason": "Broken env command and missing Qwen2_5_VLForConditionalGeneration import are different regressions." + }, + { + "left": "issue:43010", + "right": "issue:43976", + "accept": false, + "reason": "Cache update behavior and Python 3.9 compatibility failure are unrelated." + }, + { + "left": "issue:44961", + "right": "issue:45468", + "accept": false, + "reason": "Uninformative title and Gemma-4 audio positional encoding bug are unrelated." + }, + { + "left": "issue:31795", + "right": "issue:38617", + "accept": false, + "reason": "Documentation confusion in model.forward and a missing import symbol are unrelated." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "Grounded and conservative overall. The representative issue choice is supported by the packet, the summary correctly notes this is a mixed cluster rather than true duplicates, and the one accepted soft-edge pair looks plausibly mergeable as the same Qwen3.5 MoE loading regression." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 43, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:31795", + "issue:33453", + "issue:36296", + "issue:38617", + "issue:41628", + "issue:41669", + "issue:42548", + "issue:42890", + "issue:43010", + "issue:43232", + "issue:43352", + "issue:43408", + "issue:43526", + "issue:43576", + "issue:43673", + "issue:43827", + "issue:43901", + "issue:43976", + "issue:44038", + "issue:44077", + "issue:44246", + "issue:44261", + "issue:44273", + "issue:44295", + "issue:44297", + "issue:44315", + "issue:44351", + "issue:44355", + "issue:44360", + "issue:44373", + "issue:44393", + "issue:44492", + "issue:44556", + "issue:44623", + "issue:44704", + "issue:44829", + "issue:44861", + "issue:44908", + "issue:44961", + "issue:45003", + "issue:45071", + "issue:45092", + "issue:45125", + "issue:45230", + "issue:45341", + "issue:45478", + "issue:45542", + "issue:45753" + ], + "soft_pairs": [ + "issue:41628|issue:42548", + "issue:38617|issue:43576", + "issue:43576|issue:44351", + "issue:44908|issue:44961", + "issue:43010|issue:43232", + "issue:36296|issue:43901", + "issue:41628|issue:44355", + "issue:36296|issue:43526", + "issue:44261|issue:44360", + "issue:44273|issue:45003", + "issue:44623|issue:45230", + "issue:44273|issue:45542", + "issue:44393|issue:45753", + "issue:43576|issue:44556", + "issue:41669|issue:43673", + "issue:44393|issue:45071", + "issue:44038|issue:45125", + "issue:44556|issue:45092", + "issue:44373|issue:44961", + "issue:44704|issue:44908", + "issue:42890|issue:44908", + "issue:44246|issue:44273", + "issue:43408|issue:44315", + "issue:43976|issue:44393", + "issue:43352|issue:44393", + "issue:43576|issue:44908", + "issue:44297|issue:45230", + "issue:43673|issue:44492", + "issue:44908|issue:45071", + "issue:44295|issue:45230", + "issue:41669|issue:44351", + "issue:41669|issue:45341", + "issue:41669|issue:44492", + "issue:44393|issue:45478", + "issue:44315|issue:44829", + "issue:44077|issue:44861", + "issue:43827|issue:44908", + "issue:31795|issue:33453", + "issue:41669|issue:44908" + ], + "prepared_review_unit_hash": "10a6187d61c96c91a4cd9b4ea586188c5acc338f5507daced519f4dd45865222", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 48, + "item_count": 48, + "soft_pair_count": 39, + "serialized_chars": 38746, + "estimated_input_tokens": 9687, + "estimated_eval_tokens": 19630 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly a set of separate Transformers issues spanning imports, lazy loading, generation/cache behavior, docs, and model-specific regressions. No soft pair is strongly supported as the same underlying bug; the lazy-loading pair in particular lacks enough evidence to merge confidently.", + "confidence": 0.62, + "canonical_issue_reason": "issue:44273 is the broadest open report in the cluster and the safest representative for the lazy-loading subgroup, but the evidence is not strong enough to treat it as the same bug as issue:45003.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:44273 is the most central open issue and the best cluster representative by scope, even though the duplicate linkage to issue:45003 is unproven.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:41628", + "right": "issue:42548", + "accept": false, + "reason": "Both are import failures, but they involve different missing symbols and different export paths; not the same concrete bug." + }, + { + "left": "issue:38617", + "right": "issue:43576", + "accept": false, + "reason": "Missing internal import vs broken env command; different features and failure modes." + }, + { + "left": "issue:43576", + "right": "issue:44351", + "accept": false, + "reason": "An env command regression and a missing top-level cache export are unrelated." + }, + { + "left": "issue:44908", + "right": "issue:44961", + "accept": false, + "reason": "Scheduler kwargs bug and an unrelated placeholder issue; no shared code-path." + }, + { + "left": "issue:43010", + "right": "issue:43232", + "accept": false, + "reason": "Both touch generation/cache code, but they describe different defects and different call sites." + }, + { + "left": "issue:36296", + "right": "issue:43901", + "accept": false, + "reason": "Tensor-parallel training bug vs documentation mismatch; not the same issue." + }, + { + "left": "issue:41628", + "right": "issue:44355", + "accept": false, + "reason": "Top-level import failure for a symbol vs compiled-file execution error; different problems." + }, + { + "left": "issue:36296", + "right": "issue:43526", + "accept": false, + "reason": "Tensor parallel training bug and a BeitImageProcessorFast label-return bug are unrelated." + }, + { + "left": "issue:44261", + "right": "issue:44360", + "accept": false, + "reason": "Layer norm precision/config issue vs DSA indexer missing ReLU; different subsystems." + }, + { + "left": "issue:44273", + "right": "issue:45003", + "accept": false, + "reason": "The titles are loosely related around lazy loading/import behavior, but there is not enough evidence here to say they are the same concrete bug." + }, + { + "left": "issue:44623", + "right": "issue:45230", + "accept": false, + "reason": "Missing files on processor save vs a vague bug report; no evidence of the same defect." + }, + { + "left": "issue:44273", + "right": "issue:45542", + "accept": false, + "reason": "Lazy-loading failure and TensorFlow backend detection error are different bugs." + }, + { + "left": "issue:44393", + "right": "issue:45753", + "accept": false, + "reason": "Same model family, but different symptoms and likely different root causes; too broad to merge." + }, + { + "left": "issue:43576", + "right": "issue:44556", + "accept": false, + "reason": "Broken env command vs checkpoint reload incompatibility after v5 upgrade; different user flows." + }, + { + "left": "issue:41669", + "right": "issue:43673", + "accept": false, + "reason": "Import-star performance regression vs missing generation cache in chunked prefill; not the same code-path." + }, + { + "left": "issue:44393", + "right": "issue:45071", + "accept": false, + "reason": "Qwen3-VL output bug and PretrainedConfig type-checking regression are unrelated." + }, + { + "left": "issue:44038", + "right": "issue:45125", + "accept": false, + "reason": "Both are Qwen3-related, but one is a general bug and the other is a tensor-parallel missing-plan issue." + }, + { + "left": "issue:44556", + "right": "issue:45092", + "accept": false, + "reason": "Both concern v5 checkpoint incompatibilities, but they target different models and different failure causes." + }, + { + "left": "issue:44373", + "right": "issue:44961", + "accept": false, + "reason": "Docstring issue vs an unrelated placeholder issue; no overlap." + }, + { + "left": "issue:44704", + "right": "issue:44908", + "accept": false, + "reason": "Processor kwargs forwarding bug and scheduler kwargs bug are different APIs." + }, + { + "left": "issue:42890", + "right": "issue:44908", + "accept": false, + "reason": "Flaky test due to missing seeds vs scheduler configuration bug; unrelated." + }, + { + "left": "issue:44246", + "right": "issue:44273", + "accept": false, + "reason": "Slow import sometimes and lazy loading not working properly are related in theme, but the reports do not describe the same concrete defect." + }, + { + "left": "issue:43408", + "right": "issue:44315", + "accept": false, + "reason": "Model-type warning vs model_init/Liger integration issue; different code paths." + }, + { + "left": "issue:43976", + "right": "issue:44393", + "accept": false, + "reason": "Python version compatibility and a Qwen3-VL inference bug are unrelated." + }, + { + "left": "issue:43352", + "right": "issue:44393", + "accept": false, + "reason": "Flash-Attention-2 support error and Qwen3-VL bounding-box hallucination are different issues." + }, + { + "left": "issue:43576", + "right": "issue:44908", + "accept": false, + "reason": "Broken env command and scheduler kwargs regression are unrelated." + }, + { + "left": "issue:44297", + "right": "issue:45230", + "accept": false, + "reason": "Tokenizer save metadata mismatch vs a vague bug report; no same underlying change." + }, + { + "left": "issue:43673", + "right": "issue:44492", + "accept": false, + "reason": "Generation cache bug and a typo in cache strategies are not the same defect." + }, + { + "left": "issue:44908", + "right": "issue:45071", + "accept": false, + "reason": "Scheduler kwargs handling and config type checking are unrelated." + }, + { + "left": "issue:44295", + "right": "issue:45230", + "accept": false, + "reason": "position_ids buffer access error vs vague bug report; no evidence of duplication." + }, + { + "left": "issue:41669", + "right": "issue:44351", + "accept": false, + "reason": "Import-star cleanup/performance issue vs missing HybridCache export; related to imports, but not the same bug." + }, + { + "left": "issue:41669", + "right": "issue:45341", + "accept": false, + "reason": "Import performance regression and a testing_utils bug are unrelated." + }, + { + "left": "issue:41669", + "right": "issue:44492", + "accept": false, + "reason": "Import-star performance problem and a cache-strategy typo are different issues." + }, + { + "left": "issue:44393", + "right": "issue:45478", + "accept": false, + "reason": "Different Qwen3.5/Moe bugs with different symptoms and no shared concrete failure mode." + }, + { + "left": "issue:44315", + "right": "issue:44829", + "accept": false, + "reason": "Liger-kernel application bug vs flash_attention_3 degenerate training; different paths." + }, + { + "left": "issue:44077", + "right": "issue:44861", + "accept": false, + "reason": "Optional post_init validation issue vs tied-weight-key AttributeError; unrelated." + }, + { + "left": "issue:43827", + "right": "issue:44908", + "accept": false, + "reason": "Docs still referencing pipeline() vs scheduler kwargs bug; different kinds of changes." + }, + { + "left": "issue:31795", + "right": "issue:33453", + "accept": false, + "reason": "Documentation confusion in model.forward vs tokenizer loading regression; unrelated." + }, + { + "left": "issue:41669", + "right": "issue:44908", + "accept": false, + "reason": "Import-star performance regression and scheduler kwargs bug are unrelated." + } + ] + }, + "evaluator_result": { + "accept": false, + "feedback": "The cluster summary is broadly reasonable, but it overstates the evidence for the one accepted soft pair. In the packet, issue:44273 and issue:45003 only have titles and a modest similarity score; there are no explicit links or body details proving they are the same bug. Describing 45003 as a concrete cause/root-cause for 44273 is too strong on this evidence. The rest of the soft-edge rejections look conservative, but this accepted duplicate verdict is not sufficiently grounded, so I can\u2019t accept the analyst output." + }, + "evaluator_used": true, + "retried": true, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 2, + "component_count": 16, + "review_unit_index": 44, + "review_unit_count": 44, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-15354-384", + "nodes": [ + "issue:30333", + "issue:31515", + "issue:31795", + "issue:33453", + "issue:34689", + "issue:36296", + "issue:37428", + "issue:38617", + "issue:41669", + "issue:42371", + "issue:42913", + "issue:43066", + "issue:43298", + "issue:43502", + "issue:43519", + "issue:43525", + "issue:43572", + "issue:43576", + "issue:43606", + "issue:43704", + "issue:43827", + "issue:44230", + "issue:44261", + "issue:44263", + "issue:44273", + "issue:44315", + "issue:44360", + "issue:44485", + "issue:44492", + "issue:44623", + "issue:44655", + "issue:44861", + "issue:44908", + "issue:44998", + "issue:45092", + "issue:45230", + "issue:45290", + "issue:45310", + "issue:45341", + "issue:45468" + ], + "soft_pairs": [ + "issue:42371|issue:44492", + "issue:43576|issue:44273", + "issue:44655|issue:44861", + "issue:44360|issue:45468", + "issue:45230|issue:45310", + "issue:43827|issue:44492", + "issue:44273|issue:44908", + "issue:43519|issue:44485", + "issue:41669|issue:43704", + "issue:43519|issue:44908", + "issue:31795|issue:34689", + "issue:31795|issue:37428", + "issue:44315|issue:45092", + "issue:43704|issue:44485", + "issue:44485|issue:45468", + "issue:42371|issue:43704", + "issue:43502|issue:43519", + "issue:43519|issue:44623", + "issue:43519|issue:43704", + "issue:44263|issue:45468", + "issue:43606|issue:45468", + "issue:44261|issue:45468", + "issue:43704|issue:45341", + "issue:42913|issue:43066", + "issue:44908|issue:44998", + "issue:44861|issue:45290", + "issue:41669|issue:44998", + "issue:43525|issue:44861", + "issue:43704|issue:44998", + "issue:43066|issue:44998", + "issue:43572|issue:44861", + "issue:31515|issue:38617", + "issue:44360|issue:44998", + "issue:44230|issue:44623", + "issue:30333|issue:31515", + "issue:31515|issue:36296", + "issue:31515|issue:31795", + "issue:44230|issue:45310", + "issue:31515|issue:33453", + "issue:43298|issue:44230" + ], + "prepared_review_unit_hash": "6d99e37ee71e9b1c63b7091e95ed0a0e99b3cb77020de6f8a3a9afb3445b46b5", + "component_budget": { + "node_count": 384, + "item_count": 384, + "soft_pair_count": 1441, + "serialized_chars": 742684, + "estimated_input_tokens": 185671, + "estimated_eval_tokens": 371598 + }, + "budget": { + "node_count": 40, + "item_count": 40, + "soft_pair_count": 40, + "serialized_chars": 34369, + "estimated_input_tokens": 8593, + "estimated_eval_tokens": 17442 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "The cluster is highly heterogeneous: the soft pairs span unrelated bugs in tokenizers, loading, docs, schedulers, model internals, and utilities. None of the candidate pairs look like the same underlying issue/change.", + "confidence": 0.96, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:42371", + "right": "issue:44492", + "accept": false, + "reason": "TF32 behavior/settings API vs a cache-strategy typo; unrelated symptoms and code paths." + }, + { + "left": "issue:43576", + "right": "issue:44273", + "accept": false, + "reason": "Broken env command in v5 vs lazy-loading failure; different feature areas." + }, + { + "left": "issue:44655", + "right": "issue:44861", + "accept": false, + "reason": "Saving Pipeline objects vs tied-weights key handling; both involve save/load but not the same bug." + }, + { + "left": "issue:44360", + "right": "issue:45468", + "accept": false, + "reason": "DSA indexer ReLU issue vs Gemma4 audio positional encoding; different model internals." + }, + { + "left": "issue:45230", + "right": "issue:45310", + "accept": false, + "reason": "Generic bug report vs a specific Qwen3.5 MoE from_pretrained regression; not the same concrete issue." + }, + { + "left": "issue:43827", + "right": "issue:44492", + "accept": false, + "reason": "Docs still referencing pipeline() vs cache-strategy typo; unrelated." + }, + { + "left": "issue:44273", + "right": "issue:44908", + "accept": false, + "reason": "Lazy loading bug vs inverse_sqrt scheduler kwargs bug; different subsystems." + }, + { + "left": "issue:43519", + "right": "issue:44485", + "accept": false, + "reason": "Qwen3VL timestamp calculation vs GLM-5 RoPE implementation; both model-specific but distinct defects." + }, + { + "left": "issue:41669", + "right": "issue:43704", + "accept": false, + "reason": "Import-star cleanup/performance issue vs VRAM leak in Qwen3ForCausalLM; unrelated." + }, + { + "left": "issue:43519", + "right": "issue:44908", + "accept": false, + "reason": "Timestamp math in processor vs scheduler kwargs handling; no overlap in code path." + }, + { + "left": "issue:31795", + "right": "issue:34689", + "accept": false, + "reason": "Documentation confusion in model.forward vs a model-loading regression; unrelated." + }, + { + "left": "issue:31795", + "right": "issue:37428", + "accept": false, + "reason": "Docs around input_ids/past_key_values vs flash-attention import error; different topics." + }, + { + "left": "issue:44315", + "right": "issue:45092", + "accept": false, + "reason": "Liger kernel not applied with model_init vs meta-init incompatibility for old remote-code checkpoints; not the same bug." + }, + { + "left": "issue:43704", + "right": "issue:44485", + "accept": false, + "reason": "VRAM leak in dataloader-threaded causal LM use vs GLM-5 RoPE implementation; distinct issues." + }, + { + "left": "issue:44485", + "right": "issue:45468", + "accept": false, + "reason": "GLM-5 RoPE math vs Gemma4 audio positional encoding; different models and failures." + }, + { + "left": "issue:42371", + "right": "issue:43704", + "accept": false, + "reason": "TF32 configuration warning vs VRAM leak; unrelated." + }, + { + "left": "issue:43502", + "right": "issue:43519", + "accept": false, + "reason": "local_files_only still making API calls vs timestamp miscalculation; different code paths." + }, + { + "left": "issue:43519", + "right": "issue:44623", + "accept": false, + "reason": "Processor timestamp bug vs processor.save_pretrained missing files; both processor-related but not the same defect." + }, + { + "left": "issue:43519", + "right": "issue:43704", + "accept": false, + "reason": "Timestamp calculation vs VRAM leak; unrelated." + }, + { + "left": "issue:44263", + "right": "issue:45468", + "accept": false, + "reason": "torch.split return handling in GlmMoeDsaIndexer vs Gemma4 audio positional encoding; different components." + }, + { + "left": "issue:43606", + "right": "issue:45468", + "accept": false, + "reason": "CPU offload device mismatch for bark-small vs Gemma4 audio positional encoding; different models/problems." + }, + { + "left": "issue:44261", + "right": "issue:45468", + "accept": false, + "reason": "Missing rms_norm_eps in MLA q_a_layernorm vs Gemma4 audio positional encoding; unrelated." + }, + { + "left": "issue:43704", + "right": "issue:45341", + "accept": false, + "reason": "VRAM leak in model use vs a testing_utils bug; no concrete overlap." + }, + { + "left": "issue:42913", + "right": "issue:43066", + "accept": false, + "reason": "General tokenizer behavior difference across v4/v5 vs wrong decoder type in v5; related area but not clearly the same change." + }, + { + "left": "issue:44908", + "right": "issue:44998", + "accept": false, + "reason": "Scheduler kwarg bug vs unrelated issue titled 'Unemployment'; no relation." + }, + { + "left": "issue:44861", + "right": "issue:45290", + "accept": false, + "reason": "Tied-weight-key crash during loading vs chat-template crash with tool calls; different APIs and failure modes." + }, + { + "left": "issue:41669", + "right": "issue:44998", + "accept": false, + "reason": "Import-star cleanup/performance vs unrelated issue; no shared bug." + }, + { + "left": "issue:43525", + "right": "issue:44861", + "accept": false, + "reason": "Missing pad_token_id on Llama4Config vs tied-weight-key AttributeError; unrelated." + }, + { + "left": "issue:43704", + "right": "issue:44998", + "accept": false, + "reason": "VRAM leak vs unrelated issue; no overlap." + }, + { + "left": "issue:43066", + "right": "issue:44998", + "accept": false, + "reason": "Tokenizer decoder type bug vs unrelated issue; no relation." + }, + { + "left": "issue:43572", + "right": "issue:44861", + "accept": false, + "reason": "StableLmConfig pad_token_idx regression vs tied-weight-key crash; different config and code paths." + }, + { + "left": "issue:31515", + "right": "issue:38617", + "accept": false, + "reason": "Slow from_pretrained checkpoint loading vs ImportError for layer_type_validation; unrelated loader problems." + }, + { + "left": "issue:44360", + "right": "issue:44998", + "accept": false, + "reason": "DSA indexer math issue vs unrelated issue; no common bug." + }, + { + "left": "issue:44230", + "right": "issue:44623", + "accept": false, + "reason": "FP8 inference support for Qwen3-VL/Qwen3.5 MoE vs processor.save_pretrained file omission; unrelated." + }, + { + "left": "issue:30333", + "right": "issue:31515", + "accept": false, + "reason": "MLFlow reporting on failed training jobs vs slow checkpoint loading; unrelated." + }, + { + "left": "issue:31515", + "right": "issue:36296", + "accept": false, + "reason": "Slow checkpoint loading vs tensor parallel training bug; different subsystems." + }, + { + "left": "issue:31515", + "right": "issue:31795", + "accept": false, + "reason": "Loading performance regression vs documentation confusion in model.forward; unrelated." + }, + { + "left": "issue:44230", + "right": "issue:45310", + "accept": false, + "reason": "FP8 support request for Qwen3-VL/Qwen3.5 MoE infer vs a from_pretrained error on transformers>=5.4.0; different concrete problems." + }, + { + "left": "issue:31515", + "right": "issue:33453", + "accept": false, + "reason": "Slow from_pretrained loading vs tokenizer-loading regression; not the same bug." + }, + { + "left": "issue:43298", + "right": "issue:44230", + "accept": false, + "reason": "Continuous batching not supporting audio models vs FP8 support for Qwen3-VL/Qwen3.5 MoE; different features and models." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 3, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-26350-2", + "nodes": [ + "issue:26350", + "issue:33260" + ], + "soft_pairs": [ + "issue:26350|issue:33260" + ], + "prepared_review_unit_hash": "b16df79519325d4d44019b581c01e8b7983539a7c1b09b039b5d16e5f1202548", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1590, + "estimated_input_tokens": 398, + "estimated_eval_tokens": 1052 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1590, + "estimated_input_tokens": 398, + "estimated_eval_tokens": 1052 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are two separate feature requests for different model/runtime integration work: one asks for Flash Attention 2 support across more architectures, the other for GGUF support across more architectures. They share the same contribution framing but not the same underlying change.", + "confidence": 0.98, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:26350", + "right": "issue:33260", + "accept": false, + "reason": "Different features and code paths: Flash Attention 2 support vs GGUF support. Similar wording about architecture coverage does not make them duplicates." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 4, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-27088-4", + "nodes": [ + "issue:27088", + "issue:43834", + "issue:44365", + "issue:44370" + ], + "soft_pairs": [ + "issue:27088|issue:44370", + "issue:27088|issue:44365", + "issue:27088|issue:43834", + "issue:43834|issue:44370", + "issue:43834|issue:44365", + "issue:44365|issue:44370" + ], + "prepared_review_unit_hash": "536033348200ca16ef36127a4729cd4bce6a19ee705e4804a5339a55471f796e", + "component_budget": { + "node_count": 4, + "item_count": 4, + "soft_pair_count": 6, + "serialized_chars": 4454, + "estimated_input_tokens": 1114, + "estimated_eval_tokens": 2484 + }, + "budget": { + "node_count": 4, + "item_count": 4, + "soft_pair_count": 6, + "serialized_chars": 4454, + "estimated_input_tokens": 1114, + "estimated_eval_tokens": 2484 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "This cluster looks like repeated i18n-doc translation request templates, not true duplicates. The only concrete request is Turkish; the others are generic placeholder-language variants that likely refer to different languages, so they should stay separate.", + "confidence": 0.87, + "canonical_issue_reason": "issue:27088 is the most concrete and identifiable request: it names Turkish explicitly, while the others are generic template-style placeholders with no specific language shown.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:27088 is the best representative to keep because it is specific, human-readable, and clearly scoped to one translation request; the others appear to be less concrete template instances.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:27088", + "right": "issue:44370", + "accept": false, + "reason": "Same translation-request template, but no evidence they target the same language; placeholder titles suggest different underlying requests." + }, + { + "left": "issue:27088", + "right": "issue:44365", + "accept": false, + "reason": "Both are i18n-doc translation requests, but the specific language is different/unspecified, so this is template similarity rather than a duplicate." + }, + { + "left": "issue:27088", + "right": "issue:43834", + "accept": false, + "reason": "Turkish is a concrete language request; the other issue is a generic placeholder. They do not appear to be the same underlying issue." + }, + { + "left": "issue:43834", + "right": "issue:44370", + "accept": false, + "reason": "Highly similar template wording, but nothing indicates the same language target or the same exact translation task." + }, + { + "left": "issue:43834", + "right": "issue:44365", + "accept": false, + "reason": "These look like separate instances of the same template for different languages, not one duplicated issue." + }, + { + "left": "issue:44365", + "right": "issue:44370", + "accept": false, + "reason": "Only template-level similarity is visible; without a shared concrete language or task, they should not be merged as duplicates." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 5, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-36817-14", + "nodes": [ + "issue:36817", + "issue:37015", + "issue:41115", + "issue:41211", + "issue:42503", + "issue:42738", + "issue:43671", + "issue:43837", + "issue:43909", + "issue:44121", + "issue:44205", + "issue:45027", + "issue:45177", + "issue:45322" + ], + "soft_pairs": [ + "issue:36817|issue:41211", + "issue:36817|issue:43671", + "issue:37015|issue:43837", + "issue:37015|issue:44121", + "issue:37015|issue:43909", + "issue:37015|issue:42503", + "issue:37015|issue:45027", + "issue:41115|issue:43837", + "issue:41115|issue:44121", + "issue:41115|issue:42503", + "issue:41115|issue:45027", + "issue:41211|issue:43671", + "issue:42503|issue:43837", + "issue:42503|issue:43671", + "issue:42503|issue:42738", + "issue:42738|issue:45027", + "issue:43671|issue:43837", + "issue:43671|issue:45027", + "issue:43671|issue:45177", + "issue:43671|issue:44121", + "issue:43837|issue:45322", + "issue:43837|issue:43909", + "issue:43909|issue:44205", + "issue:43671|issue:43909", + "issue:44121|issue:45322", + "issue:43837|issue:44121", + "issue:43837|issue:45027", + "issue:43837|issue:45177" + ], + "prepared_review_unit_hash": "168ed3567329b21338d6486f8c4ce8b16cb462c8b4212c63d7c878503e11e042", + "component_budget": { + "node_count": 14, + "item_count": 14, + "soft_pair_count": 28, + "serialized_chars": 17649, + "estimated_input_tokens": 4413, + "estimated_eval_tokens": 9082 + }, + "budget": { + "node_count": 14, + "item_count": 14, + "soft_pair_count": 28, + "serialized_chars": 17649, + "estimated_input_tokens": 4413, + "estimated_eval_tokens": 9082 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are all model-request issues in the same broad transformers ecosystem, but they describe different concrete models/features rather than duplicate bugs or the same change.", + "confidence": 0.28, + "canonical_issue_reason": "Issue 41115 is the most representative broad architecture-add request in the set, with stronger activity than most others and a generic title that fits the cluster\u2019s overall theme.", + "canonical_pr_reason": null, + "best_issue_reason": "41115 is the best standalone issue to anchor the cluster because it is concrete, broadly phrased, and closest to the shared pattern of requesting new model support.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:36817", + "right": "issue:41211", + "accept": false, + "reason": "Different model families (EuroBERT vs DEIMv2); same add-model pattern only." + }, + { + "left": "issue:36817", + "right": "issue:43671", + "accept": false, + "reason": "EuroBERT config request and Qwen3-TTS support are different models and modalities." + }, + { + "left": "issue:37015", + "right": "issue:43837", + "accept": false, + "reason": "NeoBERT and Qwen3-ASR are unrelated concrete additions." + }, + { + "left": "issue:37015", + "right": "issue:44121", + "accept": false, + "reason": "NeoBERT support and weight-sparse transformer support are different code/model changes." + }, + { + "left": "issue:37015", + "right": "issue:43909", + "accept": false, + "reason": "NeoBERT vs LFM2.5 Audio are different model requests." + }, + { + "left": "issue:37015", + "right": "issue:42503", + "accept": false, + "reason": "NeoBERT and ModernVBERT are distinct model additions." + }, + { + "left": "issue:37015", + "right": "issue:45027", + "accept": false, + "reason": "NeoBERT and Voxtral TTS are different model families and tasks." + }, + { + "left": "issue:41115", + "right": "issue:43837", + "accept": false, + "reason": "MiniCPM3 architecture work and Qwen3-ASR support are not the same underlying change." + }, + { + "left": "issue:41115", + "right": "issue:44121", + "accept": false, + "reason": "MiniCPM3 support and the sparse-transformer request target different implementations." + }, + { + "left": "issue:41115", + "right": "issue:42503", + "accept": false, + "reason": "MiniCPM3 and ModernVBERT are separate model-support requests." + }, + { + "left": "issue:41115", + "right": "issue:45027", + "accept": false, + "reason": "MiniCPM3 architecture support and Voxtral-TTS support are unrelated changes." + }, + { + "left": "issue:41211", + "right": "issue:43671", + "accept": false, + "reason": "DEIMv2 and Qwen3-TTS are different model additions." + }, + { + "left": "issue:42503", + "right": "issue:43837", + "accept": false, + "reason": "ModernVBERT and Qwen3-ASR are distinct model requests." + }, + { + "left": "issue:42503", + "right": "issue:43671", + "accept": false, + "reason": "ModernVBERT and Qwen3-TTS are different concrete supports." + }, + { + "left": "issue:42503", + "right": "issue:42738", + "accept": false, + "reason": "One is a specific model request, the other is a broad BERT/RoPE feature proposal." + }, + { + "left": "issue:42738", + "right": "issue:45027", + "accept": false, + "reason": "BERT-like models with RoPE and Voxtral-TTS are unrelated requests." + }, + { + "left": "issue:43671", + "right": "issue:43837", + "accept": false, + "reason": "Both are Qwen3 audio requests, but TTS and ASR are separate concrete changes." + }, + { + "left": "issue:43671", + "right": "issue:45027", + "accept": false, + "reason": "Qwen3-TTS and Voxtral-TTS are different model integrations." + }, + { + "left": "issue:43671", + "right": "issue:43909", + "accept": false, + "reason": "TTS support for Qwen3 and LFM2.5 Audio 1.5B are different model additions." + }, + { + "left": "issue:43671", + "right": "issue:45177", + "accept": false, + "reason": "Qwen3-TTS support and DeepSeek-OCR2 are different modalities and code paths." + }, + { + "left": "issue:43671", + "right": "issue:44121", + "accept": false, + "reason": "Audio model support and sparse-transformer support are unrelated." + }, + { + "left": "issue:43837", + "right": "issue:45322", + "accept": false, + "reason": "Qwen3-ASR and EUPE are different model requests." + }, + { + "left": "issue:43837", + "right": "issue:43909", + "accept": false, + "reason": "ASR support and LFM2.5 Audio support are related only at a broad audio level, not the same change." + }, + { + "left": "issue:43909", + "right": "issue:44205", + "accept": false, + "reason": "LFM2.5 Audio and SAM3-LiteText are different model families." + }, + { + "left": "issue:44121", + "right": "issue:45322", + "accept": false, + "reason": "Sparse-transformer support and EUPE support are unrelated changes." + }, + { + "left": "issue:43837", + "right": "issue:44121", + "accept": false, + "reason": "Qwen3-ASR and the sparse-transformer request target different code paths." + }, + { + "left": "issue:43837", + "right": "issue:45027", + "accept": false, + "reason": "Qwen3-ASR and Voxtral-TTS are different concrete audio-model integrations." + }, + { + "left": "issue:43837", + "right": "issue:45177", + "accept": false, + "reason": "Qwen3-ASR and DeepSeek-OCR2 are different tasks and models." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 6, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-37934-6", + "nodes": [ + "issue:37934", + "issue:42936", + "issue:43240", + "issue:43668", + "issue:43813", + "issue:44031" + ], + "soft_pairs": [ + "issue:37934|issue:43668", + "issue:42936|issue:43240", + "issue:42936|issue:44031", + "issue:42936|issue:43813", + "issue:42936|issue:43668", + "issue:43240|issue:43668", + "issue:43668|issue:44031", + "issue:43668|issue:43813" + ], + "prepared_review_unit_hash": "0ab1a9b3f48581e5ce6637fec5422564f2d1d4e7401a049d1bdd9149206412c1", + "component_budget": { + "node_count": 6, + "item_count": 6, + "soft_pair_count": 8, + "serialized_chars": 6291, + "estimated_input_tokens": 1573, + "estimated_eval_tokens": 3402 + }, + "budget": { + "node_count": 6, + "item_count": 6, + "soft_pair_count": 8, + "serialized_chars": 6291, + "estimated_input_tokens": 1573, + "estimated_eval_tokens": 3402 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These issues are all superficially similar only at the wording level, but they describe different bugs across unrelated components (norm/config typing, Mask2Former loss handling, tokenizer regex warnings, and a typo in a quantization attribute). I do not see a true duplicate cluster here.", + "confidence": 0.93, + "canonical_issue_reason": "No single issue clearly represents a duplicate set; the items span distinct subsystems and problem types, so there is no good canonical duplicate issue.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 43668 is the cleanest representative only in the sense that it is a narrowly scoped, concrete bug report, but it is not a cluster canonical because the other issues are unrelated.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:37934", + "right": "issue:43668", + "accept": false, + "reason": "Different problems: Llama4TextL2Norm/RMS norm semantics vs ModernBERTConfig type-hint error. No shared code path or concrete duplicate evidence." + }, + { + "left": "issue:42936", + "right": "issue:43240", + "accept": false, + "reason": "Mask2Former ignore_value handling and kwargs not reaching loss calculation are different bugs, even though both touch loss-related code." + }, + { + "left": "issue:42936", + "right": "issue:44031", + "accept": false, + "reason": "Mask2Former loss ignore_value bug is unrelated to tokenizer regex warning behavior." + }, + { + "left": "issue:42936", + "right": "issue:43813", + "accept": false, + "reason": "Different subsystems and failure modes: model loss parameter use vs a typo in an attribute name." + }, + { + "left": "issue:42936", + "right": "issue:43668", + "accept": false, + "reason": "Mask2Former loss issue and ModernBERT config type hint issue are unrelated." + }, + { + "left": "issue:43240", + "right": "issue:43668", + "accept": false, + "reason": "Loss kwargs forwarding bug is not the same as an incorrect config type hint." + }, + { + "left": "issue:43668", + "right": "issue:44031", + "accept": false, + "reason": "ModernBERTConfig typing issue and tokenizer regex warning issue do not describe the same underlying defect." + }, + { + "left": "issue:43668", + "right": "issue:43813", + "accept": false, + "reason": "A type-hint mistake in ModernBERTConfig is unrelated to a typo in quantization-operation naming." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 7, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-41084-2", + "nodes": [ + "issue:41084", + "issue:43037" + ], + "soft_pairs": [ + "issue:41084|issue:43037" + ], + "prepared_review_unit_hash": "d32cdff39c4e1d0b88ebeb119b69fc9c0518059ad9a72492dbebb8435786e533", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1480, + "estimated_input_tokens": 370, + "estimated_eval_tokens": 996 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1480, + "estimated_input_tokens": 370, + "estimated_eval_tokens": 996 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "The two issues are unrelated: one requests Block Decoding support, the other requests DeepSeek v3.2 support. There is no shared codepath, filename, or explicit linkage suggesting duplication.", + "confidence": 0.97, + "canonical_issue_reason": "Issue 41084 is the more concrete, feature-oriented request (\u201cSet Block Decoding\u201d) and is not dependent on a model-specific support request.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 41084 is the better canonical anchor because it names a specific functionality, whereas 43037 is a broader model-support request.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:41084", + "right": "issue:43037", + "accept": false, + "reason": "Different underlying requests: block decoding vs DeepSeek v3.2 support. The similarity is only superficial and not enough to treat them as duplicates." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 8, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43598-2", + "nodes": [ + "issue:43598", + "issue:43600" + ], + "soft_pairs": [ + "issue:43598|issue:43600" + ], + "prepared_review_unit_hash": "4103cf883f3a49d9259d06e6f1f5bc9e9d370f32c18573bdf5bd7eede20b75a0", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1582, + "estimated_input_tokens": 396, + "estimated_eval_tokens": 1048 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1582, + "estimated_input_tokens": 396, + "estimated_eval_tokens": 1048 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "The two issues are related only in that both mention custom trainers, but they describe different problems: one is about `remove_unused_column` behavior in `Trainer`, the other is about FLOPS estimation. They should not be treated as duplicates.", + "confidence": 0.93, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": "Issue 43598 is the more concrete Trainer behavior/customizability report, but it still does not subsume 43600; the latter is a separate FLOPS-accounting concern.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43598", + "right": "issue:43600", + "accept": false, + "reason": "Same broad area (custom trainers), but different underlying bugs: `remove_unused_column` customizability versus FLOPS estimation. They would not be merged as one issue." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 9, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43684-2", + "nodes": [ + "issue:43684", + "issue:44405" + ], + "soft_pairs": [ + "issue:43684|issue:44405" + ], + "prepared_review_unit_hash": "718ff7d60e2025fb7923a1f824c9e1fc0d0b79990e64c22f00360e466157782a", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1611, + "estimated_input_tokens": 403, + "estimated_eval_tokens": 1062 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1611, + "estimated_input_tokens": 403, + "estimated_eval_tokens": 1062 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are not duplicates: one requests AutoModel/AutoModelForConditionalGeneration registration for Qwen3-Omni, while the other requests AutoModelForSequenceClassification support for Qwen3.5. They touch different model families and different auto-class entry points.", + "confidence": 0.94, + "canonical_issue_reason": "issue:43684 is the broader, more central registration request and best represents the cluster\u2019s general theme of adding missing AutoModel support.", + "canonical_pr_reason": null, + "best_issue_reason": "issue:43684 is the best representative because it is the broader core-model registration task; issue:44405 is a narrower, different head-specific support request.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43684", + "right": "issue:44405", + "accept": false, + "reason": "Different model families and different features: Qwen3-Omni registration for AutoModel/AutoModelForConditionalGeneration vs Qwen3.5 sequence-classification support. Not the same underlying bug/change." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 10, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43835-4", + "nodes": [ + "issue:43835", + "issue:43845", + "issue:43846", + "issue:43859" + ], + "soft_pairs": [ + "issue:43835|issue:43845", + "issue:43835|issue:43846", + "issue:43835|issue:43859", + "issue:43845|issue:43846", + "issue:43845|issue:43859", + "issue:43846|issue:43859" + ], + "prepared_review_unit_hash": "2e8c6528fa82edb81e5cb9233bd1f2fc53677184b19d347f9d3fc994a0e0a9d5", + "component_budget": { + "node_count": 4, + "item_count": 4, + "soft_pair_count": 6, + "serialized_chars": 4264, + "estimated_input_tokens": 1066, + "estimated_eval_tokens": 2388 + }, + "budget": { + "node_count": 4, + "item_count": 4, + "soft_pair_count": 6, + "serialized_chars": 4264, + "estimated_input_tokens": 1066, + "estimated_eval_tokens": 2388 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "The cluster is mostly low-signal: three nearly empty issues titled \"huggingface\" and one unrelated short issue titled \"Fett\". The title match alone is not enough to confirm duplicate underlying bugs, so I would not merge any of the soft edges.", + "confidence": 0.58, + "canonical_issue_reason": "Issue 43845 is the earliest of the three near-identical \"huggingface\" reports, but the bodies are too sparse to establish a concrete duplicate target with confidence.", + "canonical_pr_reason": null, + "best_issue_reason": "If one issue must represent the cluster, 43845 is the safest pick because it is the earliest repeated title; however, the content is too thin to treat it as a strong canonical bug report.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:43835", + "right": "issue:43845", + "accept": false, + "reason": "Different titles and no body/detail evidence tying them to the same underlying bug." + }, + { + "left": "issue:43835", + "right": "issue:43846", + "accept": false, + "reason": "Different titles and no concrete shared issue description beyond weak similarity." + }, + { + "left": "issue:43835", + "right": "issue:43859", + "accept": false, + "reason": "Different titles and no explicit target or technical detail indicating the same problem." + }, + { + "left": "issue:43845", + "right": "issue:43846", + "accept": false, + "reason": "Same short title, but the reports are too empty to confirm they describe the same concrete bug." + }, + { + "left": "issue:43845", + "right": "issue:43859", + "accept": false, + "reason": "Same short title, but there is not enough detail to safely treat these as duplicate reports of one bug." + }, + { + "left": "issue:43846", + "right": "issue:43859", + "accept": false, + "reason": "Same short title, but the available evidence is too thin to confirm a shared underlying issue." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 11, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-44371-5", + "nodes": [ + "issue:44371", + "issue:44418", + "issue:44572", + "issue:44573", + "issue:44937" + ], + "soft_pairs": [ + "issue:44371|issue:44572", + "issue:44371|issue:44573", + "issue:44371|issue:44418", + "issue:44371|issue:44937", + "issue:44418|issue:44572", + "issue:44418|issue:44573", + "issue:44572|issue:44573", + "issue:44572|issue:44937", + "issue:44573|issue:44937" + ], + "prepared_review_unit_hash": "616eca8ac804bb082d3b2a9b45fd5f7259c8ef66031a1e7cb6561c55445a91d7", + "component_budget": { + "node_count": 5, + "item_count": 5, + "soft_pair_count": 9, + "serialized_chars": 5788, + "estimated_input_tokens": 1447, + "estimated_eval_tokens": 3150 + }, + "budget": { + "node_count": 5, + "item_count": 5, + "soft_pair_count": 9, + "serialized_chars": 5788, + "estimated_input_tokens": 1447, + "estimated_eval_tokens": 3150 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are low-information spam/documentation-like issues with no concrete shared bug or change, so they should not be deduplicated into one canonical thread.", + "confidence": 0.93, + "canonical_issue_reason": "No clear canonical duplicate exists: the issues do not describe the same problem, and the available text is too generic/incoherent to establish a single underlying report.", + "canonical_pr_reason": null, + "best_issue_reason": "No issue is a good global canonical because the cluster does not form a real duplicate set; the titles point to unrelated spam/marketing-style reports rather than one defect.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:44371", + "right": "issue:44572", + "accept": false, + "reason": "Both are generic spam markers, but there is no concrete shared bug or change to prove they are the same underlying issue." + }, + { + "left": "issue:44371", + "right": "issue:44573", + "accept": false, + "reason": "Identical-looking spam titles are not enough to confirm a duplicate; no substantive issue details are present." + }, + { + "left": "issue:44371", + "right": "issue:44418", + "accept": false, + "reason": "One is spam-like and the other is a documentation suggestion; they do not describe the same problem." + }, + { + "left": "issue:44371", + "right": "issue:44937", + "accept": false, + "reason": "The titles refer to different topics, and there is no evidence of a shared underlying defect or request." + }, + { + "left": "issue:44418", + "right": "issue:44572", + "accept": false, + "reason": "Documentation enhancement and spam are not the same user problem." + }, + { + "left": "issue:44418", + "right": "issue:44573", + "accept": false, + "reason": "Documentation enhancement and spam are not the same user problem." + }, + { + "left": "issue:44572", + "right": "issue:44573", + "accept": false, + "reason": "Both are generic spam titles, but the content is too empty to safely treat them as the same underlying report." + }, + { + "left": "issue:44572", + "right": "issue:44937", + "accept": false, + "reason": "Spam and a Google LLC mention are different topics with no concrete shared bug/change." + }, + { + "left": "issue:44573", + "right": "issue:44937", + "accept": false, + "reason": "Spam and a Google LLC mention are different topics with no concrete shared bug/change." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 12, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45145-2", + "nodes": [ + "issue:45145", + "issue:45151" + ], + "soft_pairs": [ + "issue:45145|issue:45151" + ], + "prepared_review_unit_hash": "edc315d7efaaef003d33479330b683c205cf7fccfa2e14cedd97f52410a3c64c", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1646, + "estimated_input_tokens": 412, + "estimated_eval_tokens": 1080 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1646, + "estimated_input_tokens": 412, + "estimated_eval_tokens": 1080 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "The two closed issues appear to be duplicates of the same request/problem: identical titles, nearly identical body sizes, and strong text similarity. Issue 45145 is the better canonical choice because it was created earlier and has an inbound reference.", + "confidence": 0.98, + "canonical_issue_reason": "Issue 45145 is the best canonical issue: same title/topic as 45151, created earlier, and it already has an inbound reference.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 45145 is the strongest representative of the duplicate set because it is earlier and has slightly more link activity.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:45145", + "right": "issue:45151", + "accept": true, + "reason": "Identical titles and very similar bodies strongly indicate the same underlying issue/request." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet: both issues have identical titles, similar body sizes, and the pair is marked with soft_similarity. Choosing 45145 as canonical is also supported by its earlier creation time and inbound reference. The soft-edge verdict is conservative enough given the evidence available." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 13, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45295-2", + "nodes": [ + "issue:45295", + "issue:45373" + ], + "soft_pairs": [], + "prepared_review_unit_hash": "f5fd95fada2f530ab0999e343c17c16266cde0c7182bb22509972ca44827cd49", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 0, + "serialized_chars": 1286, + "estimated_input_tokens": 322, + "estimated_eval_tokens": 900 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 0, + "serialized_chars": 1286, + "estimated_input_tokens": 322, + "estimated_eval_tokens": 900 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These two issues describe the same missing Gemma 4 sequence-classification support; the later, more specific open issue is the better cluster representative.", + "confidence": 0.94, + "canonical_issue_reason": "issue #45373 is the clearest canonical tracker: it is open, explicitly names the missing class `Gemma4ForSequenceClassification`, and states the module-level gap relative to Gemma 2/3.", + "canonical_pr_reason": null, + "best_issue_reason": "issue #45373 best captures the concrete request and likely remaining work, while #45295 is a shorter earlier duplicate with less detail and is already closed.", + "best_pr_reason": null, + "soft_edge_verdicts": [] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 14, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45672-2", + "nodes": [ + "issue:45672", + "issue:45674" + ], + "soft_pairs": [ + "issue:45672|issue:45674" + ], + "prepared_review_unit_hash": "48cb58b1df25447b25d967ba80fc36ec4ed4c1cd1fdcaff98a33700102be2c0c", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1669, + "estimated_input_tokens": 418, + "estimated_eval_tokens": 1092 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1669, + "estimated_input_tokens": 418, + "estimated_eval_tokens": 1092 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "The two issues both concern 4-bit inference regressions, but they describe different root causes and different failure paths: one is Gemma4/audio-module specific torch.finfo() on uint8 weights, the other is a BitsAndBytesConfig skip_modules regression that drops the default lm_head exclusion.", + "confidence": 0.95, + "canonical_issue_reason": "Issue 45674 is the better cluster representative because it describes a self-contained configuration regression with a clear trigger and failure mode; issue 45672 is a separate, model-specific Gemma4 bug.", + "canonical_pr_reason": null, + "best_issue_reason": "Issue 45674 is the strongest standalone issue to keep because it is more general and actionable, while 45672 appears to be a narrower, unrelated model-specific defect.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:45672", + "right": "issue:45674", + "accept": false, + "reason": "Reject: both mention 4-bit inference, but the symptoms and causes differ. 45672 is about torch.finfo() on uint8 weights in Gemma4 audio modules; 45674 is about BitsAndBytesConfig llm_int8_skip_modules clearing the default lm_head exclusion and triggering an AssertionError. These are not the same underlying bug." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 15, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45762-2", + "nodes": [ + "issue:45762", + "issue:45763" + ], + "soft_pairs": [ + "issue:45762|issue:45763" + ], + "prepared_review_unit_hash": "984d883ec9e367730dd9ec8978243c3136573b858c38d78562b05b793fd65f76", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1476, + "estimated_input_tokens": 369, + "estimated_eval_tokens": 994 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1476, + "estimated_input_tokens": 369, + "estimated_eval_tokens": 994 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "The two closed issues are just person-name/date entries with no body or shared context, so they do not look like duplicates or the same underlying problem.", + "confidence": 0.99, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:45762", + "right": "issue:45763", + "accept": false, + "reason": "Different names and dates, with no description, references, or shared context indicating the same underlying issue." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "issue", + "component_index": 16, + "component_count": 16, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45764-2", + "nodes": [ + "issue:45764", + "issue:45765" + ], + "soft_pairs": [ + "issue:45764|issue:45765" + ], + "prepared_review_unit_hash": "55ee6551cd9d69cb5bc35db67c2536ee0d9ce188d65a7369315c1bf7ddb30caf", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1481, + "estimated_input_tokens": 371, + "estimated_eval_tokens": 998 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 1481, + "estimated_input_tokens": 371, + "estimated_eval_tokens": 998 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These two closed issues are not duplicates or part of the same bug/change; they are unrelated title-only entries with different names and dates.", + "confidence": 0.99, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": "No issue is a good canonical representative here; both are unrelated, body-less placeholder-style issues, so neither should be preferred as a cluster canonical.", + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "issue:45764", + "right": "issue:45765", + "accept": false, + "reason": "Different personal names and dates, no body or shared problem description; not the same underlying issue." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 1, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-39895-13", + "nodes": [ + "pull_request:39895", + "pull_request:41855", + "pull_request:42668", + "pull_request:43251", + "pull_request:43254", + "pull_request:43424", + "pull_request:43823", + "pull_request:43995", + "pull_request:44827", + "pull_request:45170", + "pull_request:45213", + "pull_request:45655", + "pull_request:45770" + ], + "soft_pairs": [ + "pull_request:39895|pull_request:43424", + "pull_request:41855|pull_request:43823", + "pull_request:41855|pull_request:43424", + "pull_request:42668|pull_request:43823", + "pull_request:43251|pull_request:43823", + "pull_request:43251|pull_request:44827", + "pull_request:43424|pull_request:44827", + "pull_request:43424|pull_request:43823", + "pull_request:43823|pull_request:45770", + "pull_request:43823|pull_request:45655", + "pull_request:43823|pull_request:43995", + "pull_request:43995|pull_request:45770", + "pull_request:43995|pull_request:45213", + "pull_request:43995|pull_request:44827", + "pull_request:44827|pull_request:45170", + "pull_request:44827|pull_request:45655", + "pull_request:43823|pull_request:44827", + "pull_request:44827|pull_request:45213", + "pull_request:43823|pull_request:45213", + "pull_request:45655|pull_request:45770", + "pull_request:44827|pull_request:45770", + "pull_request:43251|pull_request:43254" + ], + "prepared_review_unit_hash": "25829ef5c0f193de131203f2bb21e58e34de0d326dc6ca44f01760b74bcb2194", + "component_budget": { + "node_count": 13, + "item_count": 13, + "soft_pair_count": 22, + "serialized_chars": 36416, + "estimated_input_tokens": 9104, + "estimated_eval_tokens": 18464 + }, + "budget": { + "node_count": 13, + "item_count": 13, + "soft_pair_count": 22, + "serialized_chars": 36416, + "estimated_input_tokens": 9104, + "estimated_eval_tokens": 18464 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "This cluster is mostly heterogeneous: model additions, auto/processing refactors, tests, and docs are mixed together. The only true duplicate-like pair is the two fixed_cross_entropy PRs (43251 and 43254); all other soft edges look unrelated and should stay separate.", + "confidence": 0.96, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "pull_request:39895", + "right": "pull_request:43424", + "accept": false, + "reason": "VideoPrism model addition vs an Executorch dynamic-shape test; different code paths and goals." + }, + { + "left": "pull_request:41855", + "right": "pull_request:43823", + "accept": false, + "reason": "Mistral tokenizer helper methods vs adding a new MobileLLM model; unrelated areas." + }, + { + "left": "pull_request:41855", + "right": "pull_request:43424", + "accept": false, + "reason": "Tokenizer API changes vs Executorch export test coverage; not the same bug or change." + }, + { + "left": "pull_request:42668", + "right": "pull_request:43823", + "accept": false, + "reason": "Processor auto-loading robustness vs a new model implementation; separate problems." + }, + { + "left": "pull_request:43251", + "right": "pull_request:43823", + "accept": false, + "reason": "Cross-entropy kwargs fix vs MobileLLM model addition; no shared underlying change." + }, + { + "left": "pull_request:43251", + "right": "pull_request:44827", + "accept": false, + "reason": "Loss-function API fix vs Mistral4 test fixes; unrelated code paths." + }, + { + "left": "pull_request:43424", + "right": "pull_request:44827", + "accept": false, + "reason": "Executorch export test vs Mistral4 tests; both are tests, but not the same concrete issue." + }, + { + "left": "pull_request:43424", + "right": "pull_request:43823", + "accept": false, + "reason": "Executorch dynamic-shape export test vs model addition; unrelated." + }, + { + "left": "pull_request:43823", + "right": "pull_request:45770", + "accept": false, + "reason": "New MobileLLM model files vs AutoModel text_config unwrapping; different subsystems." + }, + { + "left": "pull_request:43823", + "right": "pull_request:45655", + "accept": false, + "reason": "MobileLLM model addition vs config resolution-order fix; not the same change." + }, + { + "left": "pull_request:43823", + "right": "pull_request:43995", + "accept": false, + "reason": "New model implementation vs Falcon output-collection refactor; unrelated." + }, + { + "left": "pull_request:43995", + "right": "pull_request:45770", + "accept": false, + "reason": "Falcon model refactor vs AutoFactory config unwrapping; different code paths." + }, + { + "left": "pull_request:43995", + "right": "pull_request:45213", + "accept": false, + "reason": "Falcon refactor vs model-creation skill documentation; clearly unrelated." + }, + { + "left": "pull_request:43995", + "right": "pull_request:44827", + "accept": false, + "reason": "Falcon refactor vs Mistral4 test fixes; not the same bug or patch." + }, + { + "left": "pull_request:44827", + "right": "pull_request:45170", + "accept": false, + "reason": "Mistral4 test fixes vs a layernorm typo rename sweep; unrelated changes." + }, + { + "left": "pull_request:44827", + "right": "pull_request:45655", + "accept": false, + "reason": "Mistral4 tests vs cls.config resolution order fix; different issues." + }, + { + "left": "pull_request:43823", + "right": "pull_request:44827", + "accept": false, + "reason": "MobileLLM model addition vs Mistral4 test fixes; unrelated." + }, + { + "left": "pull_request:44827", + "right": "pull_request:45213", + "accept": false, + "reason": "Mistral4 tests vs AI skill/docs for model creation; not a duplicate." + }, + { + "left": "pull_request:43823", + "right": "pull_request:45213", + "accept": false, + "reason": "Model implementation PR vs workflow skill/documentation PR; unrelated." + }, + { + "left": "pull_request:43823", + "right": "pull_request:45770", + "accept": false, + "reason": "MobileLLM model addition vs AutoFactory text_config unwrap; different areas." + }, + { + "left": "pull_request:43995", + "right": "pull_request:45655", + "accept": false, + "reason": "Falcon refactor vs Python 3.14 config-resolution fix; separate concerns." + }, + { + "left": "pull_request:45655", + "right": "pull_request:45770", + "accept": false, + "reason": "Both touch config handling, but one fixes __init_subclass__ resolution order and the other unwraps text_config in AutoModelFor.from_config; not the same underlying bug." + }, + { + "left": "pull_request:43251", + "right": "pull_request:43254", + "accept": true, + "reason": "Both PRs patch fixed_cross_entropy to forward supported nn.functional.cross_entropy kwargs (weight/label_smoothing) in the same function and code path; they are effectively the same fix." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet: the cluster is heterogeneous, and the only clearly duplicate-like pair is 43251/43254, which both patch `fixed_cross_entropy` in the same file to forward the same supported `nn.functional.cross_entropy` kwargs. The other soft-edge rejections are conservative and do not overclaim a shared bug or mergeable single-PR fix." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 2, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-40546-2", + "nodes": [ + "pull_request:40546", + "pull_request:42781" + ], + "soft_pairs": [ + "pull_request:40546|pull_request:42781" + ], + "prepared_review_unit_hash": "5aa00e3b995dedc48a6f9df50c5a085ee1f1dc814b444063a0e42ad708a976e3", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 7226, + "estimated_input_tokens": 1807, + "estimated_eval_tokens": 3870 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 7226, + "estimated_input_tokens": 1807, + "estimated_eval_tokens": 3870 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are related VibeVoice PRs, but they are not duplicates: PR 40546 is the base VibeVoice integration, while PR 42781 adds VibeVoice Realtime plus an acoustic tokenizer and other realtime-specific changes. The overlap is mostly model-family plumbing, not the same concrete fix/change.", + "confidence": 0.95, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 40546 is the better canonical PR because it is the main, broader VibeVoice implementation and has substantial review activity; PR 42781 is a separate realtime-focused extension with additional components and a different scope.", + "best_issue_reason": null, + "best_pr_reason": "PR 40546 is the best representative of the cluster since it appears to be the foundational VibeVoice addition. PR 42781 is draft and targets a different feature set, so it should not be treated as the same PR.", + "soft_edge_verdicts": [ + { + "left": "pull_request:40546", + "right": "pull_request:42781", + "accept": false, + "reason": "Shared VibeVoice files indicate the same model family, but the changes are not the same underlying code-path problem. 40546 implements VibeVoice generally; 42781 adds realtime support and an acoustic tokenizer, which is a separate extension and not plausibly mergeable as one duplicate PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 3, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-40755-2", + "nodes": [ + "pull_request:40755", + "pull_request:41224" + ], + "soft_pairs": [ + "pull_request:40755|pull_request:41224" + ], + "prepared_review_unit_hash": "29ac0ff1a0017db82c84920c2cfc934dcc5d17751ac02c8874fddd77931d5134", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4675, + "estimated_input_tokens": 1169, + "estimated_eval_tokens": 2594 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4675, + "estimated_input_tokens": 1169, + "estimated_eval_tokens": 2594 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "The two pull requests target completely different models and code paths: TimesFM covariate forecasting vs. DINOv3 image classification support. They should not be clustered as duplicates.", + "confidence": 0.98, + "canonical_issue_reason": null, + "canonical_pr_reason": "[TimesFM] Add support for forecasting with covariates is the broader, more self-contained PR in this cluster and is the better representative if a single PR must be chosen.", + "best_issue_reason": null, + "best_pr_reason": "[TimesFM] Add support for forecasting with covariates is the best overall representative because it has the larger, more substantive model change and clearer feature scope.", + "soft_edge_verdicts": [ + { + "left": "[TimesFM] Add support for forecasting with covariates", + "right": "Add DINOv3ViTForImageClassification support", + "accept": false, + "reason": "Different model families, different tasks, and different implementation paths; they do not fix the same underlying bug or add the same feature." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 4, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-41356-2", + "nodes": [ + "pull_request:41356", + "pull_request:44339" + ], + "soft_pairs": [ + "pull_request:41356|pull_request:44339" + ], + "prepared_review_unit_hash": "aead1308c167b0bae6ff11bf9358340baa5e76f816618383418b505a10a9cd01", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 6050, + "estimated_input_tokens": 1513, + "estimated_eval_tokens": 3282 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 6050, + "estimated_input_tokens": 1513, + "estimated_eval_tokens": 3282 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Two DEIMv2 pull requests target the same model family, but they are not the same change: PR 41356 is an early/basic model + image processor add, while PR 44339 is a much broader follow-up that adds loss code, auto-mappings, conversion tooling, modular files, and more integration work.", + "confidence": 0.91, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 44339 is the more complete and representative DEIMv2 implementation; PR 41356 looks like an earlier partial version of the same model bring-up.", + "best_issue_reason": null, + "best_pr_reason": "PR 44339 is the strongest single artifact for the cluster because it subsumes the basic model work and includes the broader integration needed for a full Transformers addition.", + "soft_edge_verdicts": [ + { + "left": "pull_request:41356", + "right": "pull_request:44339", + "accept": false, + "reason": "Same model area, but not the same concrete change: 44339 expands the implementation substantially and adds new code paths beyond the basic PR 41356. They look like related iterations, not mergeable duplicates." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 5, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-41797-2", + "nodes": [ + "pull_request:41797", + "pull_request:45186" + ], + "soft_pairs": [ + "pull_request:41797|pull_request:45186" + ], + "prepared_review_unit_hash": "dd632b20348ef0297b4280709455a2d6ca01c645f51e1ca897e30a61eae44643", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 6722, + "estimated_input_tokens": 1681, + "estimated_eval_tokens": 3618 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 6722, + "estimated_input_tokens": 1681, + "estimated_eval_tokens": 3618 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are two separate model-addition PRs (DeepSeek OCR vs Isaac) that share framework registration plumbing but do not fix the same code-path or bug.", + "confidence": 0.97, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": "Neither PR is a duplicate of the other. If forced to pick one representative, 45186 is the cleaner model-addition PR, but it is not a canonical merge target for this cluster because the changes are unrelated.", + "soft_edge_verdicts": [ + { + "left": "pull_request:41797", + "right": "pull_request:45186", + "accept": false, + "reason": "Different model families and separate integrations. The overlap is only in shared Transformers scaffolding (auto registries, docs, repo checks), not the same concrete bug or change, so they should not be merged as duplicates." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 6, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-42975-2", + "nodes": [ + "pull_request:42975", + "pull_request:42976" + ], + "soft_pairs": [ + "pull_request:42975|pull_request:42976" + ], + "prepared_review_unit_hash": "4134162635f963596c7da92e60ed426224de5a5084e030abfa8d2d12d1a4b273", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 5935, + "estimated_input_tokens": 1484, + "estimated_eval_tokens": 3224 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 5935, + "estimated_input_tokens": 1484, + "estimated_eval_tokens": 3224 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Two workflow-maintenance PRs overlap in several GitHub Actions files, but they make different upgrades: PR 42975 is a broad Node 24 compatibility bump for core actions, while PR 42976 is a separate latest-version update centered on docker/build-push-action and related workflows.", + "confidence": 0.94, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 42975 is the clearest representative of the broader GitHub Actions compatibility upgrade; PR 42976 is a separate maintenance change focused on different action versions, so they are not duplicates.", + "best_issue_reason": null, + "best_pr_reason": "PR 42975 is the better representative for the cluster because it has a clear, specific rationale (Node 24 compatibility) and the broader workflow surface; PR 42976 is a distinct action-version bump rather than the same change.", + "soft_edge_verdicts": [ + { + "left": "pull_request:42975", + "right": "pull_request:42976", + "accept": false, + "reason": "Shared workflow filenames alone are not enough: 42975 updates actions for Node 24 compatibility, while 42976 upgrades docker/build-push-action and other actions to newer versions. Different concrete maintenance changes, so not the same underlying PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 7, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-42978-5", + "nodes": [ + "pull_request:42978", + "pull_request:43448", + "pull_request:43451", + "pull_request:45490", + "pull_request:45630" + ], + "soft_pairs": [ + "pull_request:42978|pull_request:43448", + "pull_request:42978|pull_request:43451", + "pull_request:42978|pull_request:45490", + "pull_request:43448|pull_request:43451", + "pull_request:43448|pull_request:45490", + "pull_request:43451|pull_request:45490", + "pull_request:45490|pull_request:45630" + ], + "prepared_review_unit_hash": "5ef69bba952317c3a888a426673ef7d4d9f91a1826e2112882f73ebb15d3da71", + "component_budget": { + "node_count": 5, + "item_count": 5, + "soft_pair_count": 7, + "serialized_chars": 17616, + "estimated_input_tokens": 4404, + "estimated_eval_tokens": 9064 + }, + "budget": { + "node_count": 5, + "item_count": 5, + "soft_pair_count": 7, + "serialized_chars": 17616, + "estimated_input_tokens": 4404, + "estimated_eval_tokens": 9064 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are separate model-addition PRs, not duplicates. They only overlap on shared boilerplate areas like docs and auto-registry updates, but each targets a different model/code path (ViT NEPA, Molmo, Molmo2, CTSM, Kimi2-6).", + "confidence": 0.99, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": null, + "soft_edge_verdicts": [ + { + "left": "pull_request:42978", + "right": "pull_request:43448", + "accept": false, + "reason": "Different model additions (ViT NEPA vs Molmo); shared docs/auto wiring is boilerplate, not the same fix." + }, + { + "left": "pull_request:42978", + "right": "pull_request:43451", + "accept": false, + "reason": "ViT NEPA and Molmo2 are distinct architectures with different files and tests; not the same underlying change." + }, + { + "left": "pull_request:42978", + "right": "pull_request:45490", + "accept": false, + "reason": "These add different models (ViT NEPA vs CTSM) and do not fix one shared bug or implement one shared feature." + }, + { + "left": "pull_request:43448", + "right": "pull_request:43451", + "accept": false, + "reason": "Molmo and Molmo2 are separate model families; overlap is limited to common integration scaffolding." + }, + { + "left": "pull_request:43448", + "right": "pull_request:45490", + "accept": false, + "reason": "Molmo and CTSM are unrelated model additions; similar registry/docs edits are not enough to merge." + }, + { + "left": "pull_request:43451", + "right": "pull_request:45490", + "accept": false, + "reason": "Molmo2 and CTSM touch different model implementations and test suites, so they are not duplicates." + }, + { + "left": "pull_request:45490", + "right": "pull_request:45630", + "accept": false, + "reason": "CTSM and Kimi2-6 are different model additions with different code paths; they should remain separate PRs." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 8, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43291-5", + "nodes": [ + "pull_request:43291", + "pull_request:43665", + "pull_request:44815", + "pull_request:45747", + "pull_request:45749" + ], + "soft_pairs": [ + "pull_request:43291|pull_request:43665", + "pull_request:43291|pull_request:44815", + "pull_request:43665|pull_request:44815", + "pull_request:43665|pull_request:45747", + "pull_request:43665|pull_request:45749", + "pull_request:44815|pull_request:45747" + ], + "prepared_review_unit_hash": "71667fba30a134eafec6910aed568191e2ede3cd51942216d6011056408ec2eb", + "component_budget": { + "node_count": 5, + "item_count": 5, + "soft_pair_count": 6, + "serialized_chars": 10367, + "estimated_input_tokens": 2592, + "estimated_eval_tokens": 5440 + }, + "budget": { + "node_count": 5, + "item_count": 5, + "soft_pair_count": 6, + "serialized_chars": 10367, + "estimated_input_tokens": 2592, + "estimated_eval_tokens": 5440 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These PRs are only superficially similar (mostly small fixes with generic titles). They do not appear to address the same underlying bug or change, so the soft links should be rejected. Among them, the FP8/dequant work is the strongest standalone code fix.", + "confidence": 0.9, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR #44815 is the most substantive, concrete code fix in the set: it changes loading and FP8 integration logic across multiple files, with tests, and has a clear bug-fix scope compared with the others.", + "best_issue_reason": null, + "best_pr_reason": "PR #44815 is the best representative of the cluster because it has the clearest technical problem statement and the most meaningful implementation, while the other PRs are either test-only, cosmetic, or too underspecified.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43291", + "right": "pull_request:43665", + "accept": false, + "reason": "Different subsystems and different behavior: Whisper tokenizer decoding vs CLIP/ViT tests. Similarity is only in being small fixes, not the same bug." + }, + { + "left": "pull_request:43291", + "right": "pull_request:44815", + "accept": false, + "reason": "Unrelated changes: Whisper decoding logic/tests versus FP8 dequant/loading code. No shared code path or concrete defect." + }, + { + "left": "pull_request:43665", + "right": "pull_request:44815", + "accept": false, + "reason": "CLIP/ViT test edits are not the same underlying issue as FP8 loading/dequant changes. Shared 'fix' wording is too weak to merge." + }, + { + "left": "pull_request:43665", + "right": "pull_request:45747", + "accept": false, + "reason": "No evidence of a shared bug or code path; one is test-file changes, the other is an unspecified batch-size fix." + }, + { + "left": "pull_request:43665", + "right": "pull_request:45749", + "accept": false, + "reason": "Test edits in vision models and a docstring spelling correction are clearly unrelated changes." + }, + { + "left": "pull_request:44815", + "right": "pull_request:45747", + "accept": false, + "reason": "FP8 loading/dequant changes and a split-batch-size fix are different problems with no concrete overlap." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 9, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43488-2", + "nodes": [ + "pull_request:43488", + "pull_request:43532" + ], + "soft_pairs": [ + "pull_request:43488|pull_request:43532" + ], + "prepared_review_unit_hash": "bc87674bcfacdf3a6004bbc7ad3e7a28853a54b7424532c30efc671742a872c0", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4563, + "estimated_input_tokens": 1141, + "estimated_eval_tokens": 2538 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4563, + "estimated_input_tokens": 1141, + "estimated_eval_tokens": 2538 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These two PRs are not duplicates: #43488 is a small bot-format/versioning test change around setup.py and Vit test scaffolding, while #43532 rewires several GitHub workflows plus helper/notification code. The shared Vit test filename is incidental, not evidence of the same underlying fix.", + "confidence": 0.97, + "canonical_issue_reason": null, + "canonical_pr_reason": "No canonical PR: the PRs target different code paths and different maintenance purposes, so they should not be merged into one duplicate cluster.", + "best_issue_reason": null, + "best_pr_reason": "If one PR must represent the cluster, #43532 is the broader and more substantive change, but it is still not a duplicate of #43488.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43488", + "right": "pull_request:43532", + "accept": false, + "reason": "Different underlying changes: #43488 is a setup/versioning and test-bot check, while #43532 changes CI/workflow behavior and notification utilities. They do not fix the same concrete problem and would not be merged as one PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 10, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43612-2", + "nodes": [ + "pull_request:43612", + "pull_request:43613" + ], + "soft_pairs": [ + "pull_request:43612|pull_request:43613" + ], + "prepared_review_unit_hash": "650d45821801a3666ffaa3cfe5b0bb77aad4f96932887409a164d4257e43d21b", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 6082, + "estimated_input_tokens": 1521, + "estimated_eval_tokens": 3298 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 6082, + "estimated_input_tokens": 1521, + "estimated_eval_tokens": 3298 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Two open PRs add similarly named but distinct pipelines: one for promptable concept segmentation and one for promptable visual segmentation. They share repository scaffolding, but they target different feature sets and model integrations, so they should not be treated as duplicates.", + "confidence": 0.96, + "canonical_issue_reason": null, + "canonical_pr_reason": "No canonical PR: the two PRs implement different pipelines with different model backends and task docs, not the same underlying change.", + "best_issue_reason": null, + "best_pr_reason": "No single best PR for deduplication; each PR is the primary candidate for its own distinct feature area.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43612", + "right": "pull_request:43613", + "accept": false, + "reason": "Reject: these are separate pipeline additions. One introduces Promptable Concept Segmentation (SAM3-focused), the other Promptable Visual Segmentation (SAM/SAM2/EdgeTam/tracker-focused). Shared docs/registry edits do not make them the same code-path fix or something that could be merged into one PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 11, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43651-2", + "nodes": [ + "pull_request:43651", + "pull_request:43663" + ], + "soft_pairs": [ + "pull_request:43651|pull_request:43663" + ], + "prepared_review_unit_hash": "e69588df2d965c49574fbf07583655a432c49b0a7145660ae1d99720ff7458ba", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4292, + "estimated_input_tokens": 1073, + "estimated_eval_tokens": 2402 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4292, + "estimated_input_tokens": 1073, + "estimated_eval_tokens": 2402 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both items are open PRs in `trainer.py`, but they address different Trainer extensibility points: one changes gradient-accumulation loss scaling, the other adds an override hook for signature-column filtering. They are not the same fix and should not be clustered as duplicates.", + "confidence": 0.96, + "canonical_issue_reason": null, + "canonical_pr_reason": "No single canonical PR stands out: `#43651` and `#43663` modify different code paths and solve unrelated Trainer customization problems.", + "best_issue_reason": null, + "best_pr_reason": "Neither PR is a good global representative for the other; each targets a separate behavior change in `Trainer` rather than a shared underlying bug.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43651", + "right": "pull_request:43663", + "accept": false, + "reason": "Reject: same file, but different functionality and code paths. `#43651` is about loss normalization under gradient accumulation; `#43663` is about extracting signature columns for filtering. They are not plausibly merged as one fix." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 12, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43656-3", + "nodes": [ + "pull_request:43656", + "pull_request:43836", + "pull_request:43842" + ], + "soft_pairs": [ + "pull_request:43656|pull_request:43842", + "pull_request:43836|pull_request:43842" + ], + "prepared_review_unit_hash": "bc99fb03682d89a03d7b4e6e342bd0d78c4215353840669a655441578aacfafc", + "component_budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 2, + "serialized_chars": 5992, + "estimated_input_tokens": 1498, + "estimated_eval_tokens": 3252 + }, + "budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 2, + "serialized_chars": 5992, + "estimated_input_tokens": 1498, + "estimated_eval_tokens": 3252 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "All three PRs target the same `src/transformers/cli/serve.py` runtime annotation/TypeAdapter NameError in the CLI server. PR 43842 is the cleanest, most targeted fix; the other two are broader or alternate annotation-workaround variants.", + "confidence": 0.86, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 43842 is the most precise fix: it guards `TypeAdapter` behind `TYPE_CHECKING` and quotes the affected annotation, directly addressing the reported pydantic-missing NameError with minimal scope.", + "best_issue_reason": null, + "best_pr_reason": "PR 43842 best matches the concrete crash path and has the smallest, clearest patch while still fixing the underlying runtime NameError.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43656", + "right": "pull_request:43842", + "accept": true, + "reason": "Both are fixes for the same CLI server annotation NameError around `TypeAdapter`/forward references in `serve.py`; they address the same underlying runtime problem and could be merged into one PR." + }, + { + "left": "pull_request:43836", + "right": "pull_request:43842", + "accept": true, + "reason": "Both patch the same `serve.py` `TypeAdapter` NameError path with annotation-only changes; they are alternative implementations of the same bug fix and could plausibly be combined." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet: all three PRs touch `src/transformers/cli/serve.py` and are aimed at preventing annotation-related `NameError`s, with PR 43842 being the narrowest fix. The soft-edge pair verdicts are also conservative enough: 43836 and 43842 clearly address the same `TypeAdapter` annotation issue, and 43656 is plausibly another implementation of the same underlying forward-reference/runtime annotation problem in the same file." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 13, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43779-2", + "nodes": [ + "pull_request:43779", + "pull_request:43816" + ], + "soft_pairs": [ + "pull_request:43779|pull_request:43816" + ], + "prepared_review_unit_hash": "e7c674adcc031aa8883109a5dfc678b4407fb20b10764302092e384d031ba83a", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4362, + "estimated_input_tokens": 1091, + "estimated_eval_tokens": 2438 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4362, + "estimated_input_tokens": 1091, + "estimated_eval_tokens": 2438 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both pull requests appear to implement the same SwanLab integration change: adding run id/resume support in `integration_utils.py` for issue #43698, with one variant also expanding the docs. They are strong duplicate PR candidates.", + "confidence": 0.94, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR #43816 is the cleaner representative: it carries the same SwanLab id/resume support plus the accompanying documentation updates, making it the more complete version of the shared change.", + "best_issue_reason": null, + "best_pr_reason": "PR #43816 is the best global representative because it appears to include the full feature change and docs for SwanLab run id/resume support, while #43779 is the narrower implementation-focused variant.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43779", + "right": "pull_request:43816", + "accept": true, + "reason": "Same underlying SwanLab callback enhancement for id/resume support in the same file and same issue target; the doc differences do not change the core code-path change, so these are duplicate PRs." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet: both PRs target issue #43698, touch the same file, and the previews indicate the same SwanLab id/resume support with #43816 adding docs. The duplicate/mergeable verdict is conservative enough given the shared target and overlapping code change." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 14, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 2, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43996-28", + "nodes": [ + "pull_request:43996", + "pull_request:43997", + "pull_request:44007", + "pull_request:44010", + "pull_request:44013", + "pull_request:44017", + "pull_request:44018", + "pull_request:44019", + "pull_request:44024", + "pull_request:44044", + "pull_request:44056", + "pull_request:44059", + "pull_request:44066", + "pull_request:44068", + "pull_request:44071", + "pull_request:44072", + "pull_request:44073", + "pull_request:44074", + "pull_request:44076", + "pull_request:44085", + "pull_request:44086", + "pull_request:44098", + "pull_request:44101", + "pull_request:44116", + "pull_request:44129", + "pull_request:44154", + "pull_request:44161", + "pull_request:44722" + ], + "soft_pairs": [ + "pull_request:43996|pull_request:44085", + "pull_request:43996|pull_request:44044", + "pull_request:44066|pull_request:44085", + "pull_request:44007|pull_request:44072", + "pull_request:44072|pull_request:44722", + "pull_request:44066|pull_request:44072", + "pull_request:44013|pull_request:44044", + "pull_request:44066|pull_request:44086", + "pull_request:44018|pull_request:44068", + "pull_request:44066|pull_request:44071", + "pull_request:44066|pull_request:44068", + "pull_request:44018|pull_request:44066", + "pull_request:44019|pull_request:44722", + "pull_request:44086|pull_request:44722", + "pull_request:44019|pull_request:44085", + "pull_request:44018|pull_request:44019", + "pull_request:44019|pull_request:44071", + "pull_request:44068|pull_request:44116", + "pull_request:44068|pull_request:44722", + "pull_request:44018|pull_request:44116", + "pull_request:44071|pull_request:44722", + "pull_request:44018|pull_request:44086", + "pull_request:44018|pull_request:44085", + "pull_request:44068|pull_request:44086", + "pull_request:44071|pull_request:44085", + "pull_request:44059|pull_request:44086", + "pull_request:44085|pull_request:44086", + "pull_request:44098|pull_request:44101", + "pull_request:44059|pull_request:44722", + "pull_request:44018|pull_request:44071", + "pull_request:44017|pull_request:44161", + "pull_request:44017|pull_request:44722", + "pull_request:44017|pull_request:44073", + "pull_request:44068|pull_request:44071", + "pull_request:44071|pull_request:44086", + "pull_request:43997|pull_request:44073", + "pull_request:44017|pull_request:44071", + "pull_request:44073|pull_request:44101", + "pull_request:44059|pull_request:44071", + "pull_request:44059|pull_request:44068", + "pull_request:44056|pull_request:44101", + "pull_request:44018|pull_request:44059", + "pull_request:44073|pull_request:44161", + "pull_request:44101|pull_request:44161", + "pull_request:43997|pull_request:44017", + "pull_request:43997|pull_request:44056", + "pull_request:44071|pull_request:44076", + "pull_request:44074|pull_request:44101", + "pull_request:43997|pull_request:44018", + "pull_request:44056|pull_request:44161", + "pull_request:43997|pull_request:44161", + "pull_request:44074|pull_request:44161", + "pull_request:43997|pull_request:44071", + "pull_request:44010|pull_request:44071", + "pull_request:44010|pull_request:44073", + "pull_request:44018|pull_request:44129", + "pull_request:43997|pull_request:44074", + "pull_request:43997|pull_request:44129", + "pull_request:44010|pull_request:44018", + "pull_request:44071|pull_request:44129", + "pull_request:43996|pull_request:44161", + "pull_request:43996|pull_request:44074", + "pull_request:44010|pull_request:44044", + "pull_request:44044|pull_request:44074", + "pull_request:44066|pull_request:44722", + "pull_request:44085|pull_request:44722", + "pull_request:44085|pull_request:44129", + "pull_request:43996|pull_request:44007", + "pull_request:43996|pull_request:44129", + "pull_request:44044|pull_request:44129", + "pull_request:44044|pull_request:44154", + "pull_request:44024|pull_request:44076" + ], + "prepared_review_unit_hash": "7ae479abfc365a6763527dbfc1f0619b7ea1781d5a75fa633416ca225bff432b", + "component_budget": { + "node_count": 28, + "item_count": 28, + "soft_pair_count": 77, + "serialized_chars": 84521, + "estimated_input_tokens": 21131, + "estimated_eval_tokens": 42518 + }, + "budget": { + "node_count": 28, + "item_count": 28, + "soft_pair_count": 72, + "serialized_chars": 76955, + "estimated_input_tokens": 19239, + "estimated_eval_tokens": 38734 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "This cluster is mostly a set of independent model-specific output-tracing refactors tied to the same tracking issue, not one duplicate bug. The only strong duplicate/superseding pairs are within the same model family and concrete code path (notably GPT-Neo, and GPT-J/CodeGen).", + "confidence": 0.9, + "canonical_issue_reason": null, + "canonical_pr_reason": "44722 is the best canonical PR for the cluster\u2019s main theme: it is a later, concrete GPT-J/CodeGen output-tracing refactor with the same issue target and the clearest, most complete code-path change among the overlapping GPT-J work.", + "best_issue_reason": null, + "best_pr_reason": "44722 is the strongest representative PR: it has explicit issue linkage, touches the actual implementation files, and most clearly reflects the standardized output-tracing refactor pattern seen across the cluster.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43996", + "right": "pull_request:44085", + "accept": false, + "reason": "Different models/code paths (CVT/FNet vs GPT-J/CodeGen); same tracking issue is not enough." + }, + { + "left": "pull_request:43996", + "right": "pull_request:44044", + "accept": false, + "reason": "Different model families and different implementation paths; only the output-tracing theme overlaps." + }, + { + "left": "pull_request:44066", + "right": "pull_request:44085", + "accept": true, + "reason": "Same GPT-J output-tracing refactor in the same file family; one is effectively a variant/superset of the other." + }, + { + "left": "pull_request:44007", + "right": "pull_request:44072", + "accept": false, + "reason": "ResNet vs EfficientNet are separate model implementations, so not the same underlying change." + }, + { + "left": "pull_request:44072", + "right": "pull_request:44722", + "accept": false, + "reason": "Different models and code paths; shared issue target alone does not make them duplicates." + }, + { + "left": "pull_request:44066", + "right": "pull_request:44072", + "accept": false, + "reason": "GPT-J/CodeGen refactor vs EfficientNet refactor; different underlying change." + }, + { + "left": "pull_request:44013", + "right": "pull_request:44044", + "accept": false, + "reason": "MobileNetV2 vs DeBERTa V2 are unrelated implementations." + }, + { + "left": "pull_request:44066", + "right": "pull_request:44086", + "accept": false, + "reason": "GPT-J/CodeGen output tracing vs MGP-STR output tracing; different code paths." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44068", + "accept": true, + "reason": "Both are GPT-Neo output-tracing refactors over the same file and tests; 44068 looks like a later superseding version." + }, + { + "left": "pull_request:44066", + "right": "pull_request:44071", + "accept": false, + "reason": "GPT-J/CodeGen vs MPT; separate model implementations." + }, + { + "left": "pull_request:44066", + "right": "pull_request:44068", + "accept": false, + "reason": "Different models (GPT-J/CodeGen vs GPT-Neo), so not the same concrete change." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44066", + "accept": false, + "reason": "Different model families; only the output-tracing migration pattern is shared." + }, + { + "left": "pull_request:44019", + "right": "pull_request:44722", + "accept": false, + "reason": "ResNet output tracing is unrelated to GPT-J/CodeGen output tracing." + }, + { + "left": "pull_request:44086", + "right": "pull_request:44722", + "accept": false, + "reason": "MGP-STR vs GPT-J/CodeGen are different model-specific changes." + }, + { + "left": "pull_request:44019", + "right": "pull_request:44085", + "accept": false, + "reason": "ResNet vs GPT-J; separate code paths." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44019", + "accept": false, + "reason": "GPT-Neo and ResNet are unrelated model implementations." + }, + { + "left": "pull_request:44019", + "right": "pull_request:44071", + "accept": false, + "reason": "ResNet vs MPT; different underlying refactors." + }, + { + "left": "pull_request:44068", + "right": "pull_request:44116", + "accept": false, + "reason": "GPT-Neo vs Flaubert/XLM output tracing are different model changes." + }, + { + "left": "pull_request:44068", + "right": "pull_request:44722", + "accept": false, + "reason": "GPT-Neo is a different model/code path from GPT-J/CodeGen." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44116", + "accept": false, + "reason": "GPT-Neo vs Flaubert/XLM are unrelated implementations." + }, + { + "left": "pull_request:44071", + "right": "pull_request:44722", + "accept": false, + "reason": "MPT vs GPT-J/CodeGen are different concrete changes." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44086", + "accept": false, + "reason": "GPT-Neo vs MGP-STR; different model code paths." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44085", + "accept": false, + "reason": "GPT-Neo vs GPT-J/CodeGen are not the same underlying change." + }, + { + "left": "pull_request:44068", + "right": "pull_request:44086", + "accept": false, + "reason": "GPT-Neo and MGP-STR are different model-specific refactors." + }, + { + "left": "pull_request:44059", + "right": "pull_request:44086", + "accept": false, + "reason": "GPT-2 vs MGP-STR; same output-tracing theme but different code paths." + }, + { + "left": "pull_request:44085", + "right": "pull_request:44086", + "accept": false, + "reason": "GPT-J/CodeGen vs MGP-STR are unrelated model changes." + }, + { + "left": "pull_request:44098", + "right": "pull_request:44101", + "accept": false, + "reason": "ViLT vs Flaubert/XLM are different model families." + }, + { + "left": "pull_request:44059", + "right": "pull_request:44722", + "accept": false, + "reason": "GPT-2 refactor is separate from GPT-J/CodeGen." + }, + { + "left": "pull_request:44071", + "right": "pull_request:44085", + "accept": false, + "reason": "MPT vs GPT-J; different underlying changes." + }, + { + "left": "pull_request:43997", + "right": "pull_request:44017", + "accept": false, + "reason": "RegNet vs Segformer are unrelated model implementations." + }, + { + "left": "pull_request:43997", + "right": "44056", + "accept": false, + "reason": "RegNet vs MPNet are separate code paths." + }, + { + "left": "pull_request:44071", + "right": "pull_request:44076", + "accept": false, + "reason": "MPT vs ImageGPT are different models; shared output-tracing pattern is too broad." + }, + { + "left": "pull_request:44074", + "right": "pull_request:44101", + "accept": false, + "reason": "TextNet vs Flaubert/XLM are unrelated changes." + }, + { + "left": "pull_request:43997", + "right": "pull_request:44018", + "accept": false, + "reason": "RegNet vs GPT-Neo are different model-specific refactors." + }, + { + "left": "pull_request:44071", + "right": "pull_request:44129", + "accept": false, + "reason": "MPT vs SpeechT5 are separate code paths." + }, + { + "left": "pull_request:43996", + "right": "pull_request:44161", + "accept": false, + "reason": "CVT/FNet vs LongT5 are unrelated model changes." + }, + { + "left": "pull_request:43996", + "right": "pull_request:44074", + "accept": false, + "reason": "CVT/FNet vs TextNet are different underlying changes." + }, + { + "left": "pull_request:44010", + "right": "pull_request:44044", + "accept": false, + "reason": "SqueezeBert vs DeBERTa V2 are separate model implementations." + }, + { + "left": "pull_request:44044", + "right": "pull_request:44074", + "accept": false, + "reason": "DeBERTa V2 vs TextNet are unrelated changes." + }, + { + "left": "pull_request:44066", + "right": "pull_request:44722", + "accept": true, + "reason": "Same GPT-J/CodeGen output-tracing refactor family; 44722 is a broader later version of the same change." + }, + { + "left": "pull_request:44085", + "right": "pull_request:44722", + "accept": true, + "reason": "Both are GPT-J output-tracing refactors on the same code path; 44722 appears to supersede/extend the earlier PR." + }, + { + "left": "pull_request:44085", + "right": "pull_request:44129", + "accept": false, + "reason": "GPT-J vs SpeechT5 are unrelated model changes." + }, + { + "left": "pull_request:43996", + "right": "pull_request:44007", + "accept": false, + "reason": "CVT/FNet vs ResNet are different model-specific refactors." + }, + { + "left": "pull_request:43996", + "right": "pull_request:44129", + "accept": false, + "reason": "CVT/FNet vs SpeechT5 are different underlying changes." + }, + { + "left": "pull_request:44044", + "right": "pull_request:44129", + "accept": false, + "reason": "DeBERTa V2 vs SpeechT5 are unrelated model changes." + }, + { + "left": "pull_request:44044", + "right": "pull_request:44154", + "accept": false, + "reason": "DeBERTa V2 vs Vits are separate model-specific refactors." + }, + { + "left": "pull_request:44024", + "right": "pull_request:44076", + "accept": false, + "reason": "FocalNet vs ImageGPT are different models; same refactor style does not make them duplicates." + }, + { + "left": "pull_request:44017", + "right": "pull_request:44161", + "accept": false, + "reason": "Segformer vs LongT5 are unrelated code paths." + }, + { + "left": "pull_request:44019", + "right": "pull_request:44101", + "accept": false, + "reason": "ResNet vs Flaubert/XLM are different model implementations." + }, + { + "left": "pull_request:44018", + "right": "pull_request:44068", + "accept": true, + "reason": "Same GPT-Neo output-tracing refactor with the same file/test surface; likely a later duplicate/superseding PR." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The summary is grounded in the packet, and the conservative theme distinction is appropriate: most PRs are separate model-specific output-tracing refactors sharing only the tracking issue. The accepted soft-edge pairs are limited to same-family, same-code-path overlaps (GPT-Neo and GPT-J/CodeGen), which do look mergeable as one bugfix PR. I did not see any clear overstatement of evidence." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 14, + "component_count": 26, + "review_unit_index": 2, + "review_unit_count": 2, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43996-28", + "nodes": [ + "pull_request:43997", + "pull_request:44010", + "pull_request:44056", + "pull_request:44073", + "pull_request:44074", + "pull_request:44129" + ], + "soft_pairs": [ + "pull_request:44073|pull_request:44074", + "pull_request:44056|pull_request:44073", + "pull_request:44056|pull_request:44074", + "pull_request:43997|pull_request:44010", + "pull_request:44010|pull_request:44129" + ], + "prepared_review_unit_hash": "52fdf146c1522d186152b3265164bfc207c5c6a49734db5fae454e7c6b8a1ab2", + "component_budget": { + "node_count": 28, + "item_count": 28, + "soft_pair_count": 77, + "serialized_chars": 84521, + "estimated_input_tokens": 21131, + "estimated_eval_tokens": 42518 + }, + "budget": { + "node_count": 6, + "item_count": 6, + "soft_pair_count": 5, + "serialized_chars": 13066, + "estimated_input_tokens": 3267, + "estimated_eval_tokens": 6790 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": true, + "analyst_result": { + "summary": "A cluster of similar-sounding but model-specific refactors to migrate output tracing/capture across different Transformer implementations. They share a pattern, not the same concrete bug or change.", + "confidence": 0.69, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 44129 is the best representative: it has an explicit issue link, a clear title, and a substantive output-capture refactor in one model.", + "best_issue_reason": null, + "best_pr_reason": "PR 44129 is the strongest single PR to represent the cluster because it is the most concretely scoped and best tracked, but it is still model-specific rather than duplicate of the others.", + "soft_edge_verdicts": [ + { + "left": "pull_request:44073", + "right": "pull_request:44074", + "accept": false, + "reason": "Both are output-tracing refactors, but for different models (VisualBert vs TextNet) and different code paths; not the same change." + }, + { + "left": "pull_request:44056", + "right": "pull_request:44073", + "accept": false, + "reason": "MPNet and VisualBert are separate implementations with independent decorator migrations, not one shared fix." + }, + { + "left": "pull_request:44056", + "right": "pull_request:44074", + "accept": false, + "reason": "MPNet vs TextNet are unrelated model code paths; this is only a thematic similarity." + }, + { + "left": "pull_request:43997", + "right": "pull_request:44010", + "accept": false, + "reason": "RegNet and SqueezeBert refactor different model internals; they do not fix the same concrete issue." + }, + { + "left": "pull_request:44010", + "right": "pull_request:44129", + "accept": false, + "reason": "These touch different models and output APIs; they are separate migrations, not mergeable duplicate PRs." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 15, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-43998-13", + "nodes": [ + "pull_request:43998", + "pull_request:43999", + "pull_request:44000", + "pull_request:44001", + "pull_request:44002", + "pull_request:44003", + "pull_request:44004", + "pull_request:44025", + "pull_request:44026", + "pull_request:44027", + "pull_request:44028", + "pull_request:44029", + "pull_request:44030" + ], + "soft_pairs": [ + "pull_request:43998|pull_request:44002", + "pull_request:43998|pull_request:44000", + "pull_request:43998|pull_request:44001", + "pull_request:43998|pull_request:44003", + "pull_request:43998|pull_request:44029", + "pull_request:43999|pull_request:44001", + "pull_request:43999|pull_request:44002", + "pull_request:43999|pull_request:44025", + "pull_request:43999|pull_request:44026", + "pull_request:43999|pull_request:44027", + "pull_request:44000|pull_request:44001", + "pull_request:44000|pull_request:44002", + "pull_request:44000|pull_request:44003", + "pull_request:44000|pull_request:44025", + "pull_request:44001|pull_request:44002", + "pull_request:44001|pull_request:44003", + "pull_request:44001|pull_request:44025", + "pull_request:44002|pull_request:44029", + "pull_request:44002|pull_request:44003", + "pull_request:44003|pull_request:44028", + "pull_request:44003|pull_request:44004", + "pull_request:43998|pull_request:44004", + "pull_request:44001|pull_request:44004", + "pull_request:44002|pull_request:44004", + "pull_request:44000|pull_request:44004", + "pull_request:44025|pull_request:44026", + "pull_request:44025|pull_request:44027", + "pull_request:44025|pull_request:44028", + "pull_request:44025|pull_request:44029", + "pull_request:44025|pull_request:44030", + "pull_request:44026|pull_request:44029", + "pull_request:44026|pull_request:44027", + "pull_request:44026|pull_request:44028", + "pull_request:44026|pull_request:44030", + "pull_request:44027|pull_request:44028", + "pull_request:44027|pull_request:44029", + "pull_request:44027|pull_request:44030", + "pull_request:44028|pull_request:44029", + "pull_request:44028|pull_request:44030", + "pull_request:44029|pull_request:44030" + ], + "prepared_review_unit_hash": "c3919d02f3fbf8f474c9c504e44a56d4d295f4af20d359637347970b2e3071b6", + "component_budget": { + "node_count": 13, + "item_count": 13, + "soft_pair_count": 40, + "serialized_chars": 40161, + "estimated_input_tokens": 10041, + "estimated_eval_tokens": 20338 + }, + "budget": { + "node_count": 13, + "item_count": 13, + "soft_pair_count": 40, + "serialized_chars": 40161, + "estimated_input_tokens": 10041, + "estimated_eval_tokens": 20338 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "This cluster is a set of similar-but-separate output-tracing refactors across different model families. They share a theme, but they are not duplicates of the same underlying bug/change.", + "confidence": 0.88, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 44025 is the cleanest representative of the cluster: it applies the same output-tracing refactor across two closely related Depth Anything files and is a focused, coherent change.", + "best_issue_reason": null, + "best_pr_reason": "PR 44025 is the best overall representative for triage because it is narrow, self-contained, and clearly exemplifies the cluster\u2019s output-tracing modernization.", + "soft_edge_verdicts": [ + { + "left": "pull_request:43998", + "right": "pull_request:44002", + "accept": false, + "reason": "Different model families and files (timm_backbone vs upernet); same refactor pattern, not the same concrete change." + }, + { + "left": "pull_request:43998", + "right": "pull_request:44000", + "accept": false, + "reason": "Different model families (timm_backbone vs vision_text_dual_encoder); only the output-tracing theme matches." + }, + { + "left": "pull_request:43998", + "right": "pull_request:44001", + "accept": false, + "reason": "Different architectures and files; both are refactors of return handling, not one underlying fix." + }, + { + "left": "pull_request:43998", + "right": "pull_request:44003", + "accept": false, + "reason": "timm_backbone and mamba/falcon_mamba are unrelated code paths; shared refactor motif is too broad." + }, + { + "left": "pull_request:43998", + "right": "pull_request:44029", + "accept": false, + "reason": "Different models and execution paths (timm_backbone vs rwkv); not mergeable as one concrete PR." + }, + { + "left": "pull_request:43999", + "right": "pull_request:44001", + "accept": false, + "reason": "MobileNetV1 and UnivNet are unrelated models; same tracing cleanup pattern only." + }, + { + "left": "pull_request:43999", + "right": "pull_request:44002", + "accept": false, + "reason": "Different subsystems (MobileNetV1 vs UperNet); no shared bug or code path." + }, + { + "left": "pull_request:43999", + "right": "pull_request:44025", + "accept": false, + "reason": "MobileNetV1 and Depth Anything are separate model families; both refactor outputs but do not fix the same thing." + }, + { + "left": "pull_request:43999", + "right": "pull_request:44026", + "accept": false, + "reason": "MobileNetV1 and VisionEncoderDecoder are unrelated code paths; only the refactor style is similar." + }, + { + "left": "pull_request:43999", + "right": "pull_request:44027", + "accept": false, + "reason": "MobileNetV1 and SpeechEncoderDecoder are different architectures; not the same concrete change." + }, + { + "left": "pull_request:44000", + "right": "pull_request:44001", + "accept": false, + "reason": "Vision-Text Dual Encoder and UnivNet are unrelated models; same output-tracing modernization, not a duplicate." + }, + { + "left": "pull_request:44000", + "right": "pull_request:44002", + "accept": false, + "reason": "Different model families and file scopes; no shared underlying bug." + }, + { + "left": "pull_request:44000", + "right": "pull_request:44003", + "accept": false, + "reason": "Vision-Text Dual Encoder vs Mamba/FalconMamba are distinct code paths; too broad to combine." + }, + { + "left": "pull_request:44000", + "right": "pull_request:44025", + "accept": false, + "reason": "Different model families and behaviors; only the tracing refactor is common." + }, + { + "left": "pull_request:44001", + "right": "pull_request:44002", + "accept": false, + "reason": "UnivNet and UperNet are unrelated; same refactor theme does not make them duplicates." + }, + { + "left": "pull_request:44001", + "right": "pull_request:44003", + "accept": false, + "reason": "UnivNet and Mamba/FalconMamba fix different model code paths; not mergeable as one PR." + }, + { + "left": "pull_request:44001", + "right": "pull_request:44025", + "accept": false, + "reason": "UnivNet and Depth Anything are separate model implementations; only the API cleanup is shared." + }, + { + "left": "pull_request:44002", + "right": "pull_request:44029", + "accept": false, + "reason": "UperNet and RWKV are unrelated architectures; not the same underlying issue." + }, + { + "left": "pull_request:44002", + "right": "pull_request:44003", + "accept": false, + "reason": "UperNet and Mamba/FalconMamba are different code paths; shared pattern is insufficient." + }, + { + "left": "pull_request:44003", + "right": "pull_request:44028", + "accept": false, + "reason": "Mamba/FalconMamba and SuperPoint are unrelated model families; these are separate refactors." + }, + { + "left": "pull_request:44003", + "right": "pull_request:44004", + "accept": false, + "reason": "Mamba/FalconMamba and CodeGen are different implementations with different output-shape changes; not one concrete fix." + }, + { + "left": "pull_request:43998", + "right": "pull_request:44004", + "accept": false, + "reason": "Timm Backbone and CodeGen are unrelated model paths; same tracing cleanup only." + }, + { + "left": "pull_request:44001", + "right": "pull_request:44004", + "accept": false, + "reason": "UnivNet and CodeGen are different model types and outputs; not a duplicate change." + }, + { + "left": "pull_request:44002", + "right": "pull_request:44004", + "accept": false, + "reason": "UperNet and CodeGen are unrelated; the refactor pattern alone is not enough." + }, + { + "left": "pull_request:44000", + "right": "pull_request:44004", + "accept": false, + "reason": "Vision-Text Dual Encoder and CodeGen address different forward paths; no shared concrete bug." + }, + { + "left": "pull_request:44025", + "right": "pull_request:44026", + "accept": false, + "reason": "Depth Anything and VisionEncoderDecoder are unrelated model families; only the output-tracing refactor pattern overlaps." + }, + { + "left": "pull_request:44025", + "right": "pull_request:44027", + "accept": false, + "reason": "Depth Anything and SpeechEncoderDecoder are different code paths; not the same underlying change." + }, + { + "left": "pull_request:44025", + "right": "pull_request:44028", + "accept": false, + "reason": "Depth Anything and SuperPoint are distinct models; same modernized return handling, not a duplicate." + }, + { + "left": "pull_request:44025", + "right": "pull_request:44029", + "accept": false, + "reason": "Depth Anything and RWKV are unrelated implementations; too broad to merge as one PR." + }, + { + "left": "pull_request:44025", + "right": "pull_request:44030", + "accept": false, + "reason": "Depth Anything and DPR are separate model families; same refactor theme only." + }, + { + "left": "pull_request:44026", + "right": "pull_request:44029", + "accept": false, + "reason": "VisionEncoderDecoder and RWKV are unrelated model paths; not the same concrete problem." + }, + { + "left": "pull_request:44026", + "right": "pull_request:44027", + "accept": false, + "reason": "VisionEncoderDecoder and SpeechEncoderDecoder are different wrapper models; no shared bug." + }, + { + "left": "pull_request:44026", + "right": "pull_request:44028", + "accept": false, + "reason": "VisionEncoderDecoder and SuperPoint are unrelated; both just adopt the same tracing API changes." + }, + { + "left": "pull_request:44026", + "right": "pull_request:44030", + "accept": false, + "reason": "VisionEncoderDecoder and DPR are distinct code paths; not duplicates." + }, + { + "left": "pull_request:44027", + "right": "pull_request:44028", + "accept": false, + "reason": "SpeechEncoderDecoder and SuperPoint are unrelated model families; same output-tracing theme only." + }, + { + "left": "pull_request:44027", + "right": "pull_request:44029", + "accept": false, + "reason": "SpeechEncoderDecoder and RWKV are different architectures; not one underlying change." + }, + { + "left": "pull_request:44027", + "right": "pull_request:44030", + "accept": false, + "reason": "SpeechEncoderDecoder and DPR are unrelated; no shared concrete fix." + }, + { + "left": "pull_request:44028", + "right": "pull_request:44029", + "accept": false, + "reason": "SuperPoint and RWKV are different model paths; not the same bug or change." + }, + { + "left": "pull_request:44028", + "right": "pull_request:44030", + "accept": false, + "reason": "SuperPoint and DPR are unrelated implementations; only the refactor style matches." + }, + { + "left": "pull_request:44029", + "right": "pull_request:44030", + "accept": false, + "reason": "RWKV and DPR are different model families; same output-tracing motif is too broad to treat as duplicate." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 16, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-44662-3", + "nodes": [ + "pull_request:44662", + "pull_request:45075", + "pull_request:45401" + ], + "soft_pairs": [ + "pull_request:44662|pull_request:45401", + "pull_request:44662|pull_request:45075", + "pull_request:45075|pull_request:45401" + ], + "prepared_review_unit_hash": "9ed0a4d6e72151d80cc757f8aabedde7367ea672ef04932995acd51146e3c693", + "component_budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 3, + "serialized_chars": 10914, + "estimated_input_tokens": 2729, + "estimated_eval_tokens": 5714 + }, + "budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 3, + "serialized_chars": 10914, + "estimated_input_tokens": 2729, + "estimated_eval_tokens": 5714 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are three separate model-addition PRs (PenguinVL, DeepSeek-OCR-2, Voxtral TTS). They share common Transformers scaffolding and docs/auto-registry edits, but they do not address the same model or concrete code-path bug, so they are not duplicates.", + "confidence": 0.97, + "canonical_issue_reason": null, + "canonical_pr_reason": null, + "best_issue_reason": null, + "best_pr_reason": "No single PR is a duplicate anchor here; each introduces a different model implementation with its own files and scope.", + "soft_edge_verdicts": [ + { + "left": "pull_request:44662", + "right": "pull_request:45401", + "accept": false, + "reason": "Different models and different code paths; only generic model-registration/docs overlap." + }, + { + "left": "pull_request:44662", + "right": "pull_request:45075", + "accept": false, + "reason": "Separate model implementations (PenguinVL vs DeepSeek-OCR-2); shared auto/docs plumbing is not enough to merge as one PR." + }, + { + "left": "pull_request:45075", + "right": "pull_request:45401", + "accept": false, + "reason": "Different model families and functionality; overlap is limited to shared Transformers boilerplate." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 17, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-44979-2", + "nodes": [ + "pull_request:44979", + "pull_request:45363" + ], + "soft_pairs": [ + "pull_request:44979|pull_request:45363" + ], + "prepared_review_unit_hash": "f6c5b2dbc9d5006d4549d38c3eef47bad39c7fc3b50f36d1d92a952477e0a08a", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4407, + "estimated_input_tokens": 1102, + "estimated_eval_tokens": 2460 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4407, + "estimated_input_tokens": 1102, + "estimated_eval_tokens": 2460 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These PRs are related conceptually around module/kernel fusion, but they implement different layers of the stack: #44979 adds a new generic Module Fusion API, while #45363 adds KernelConfig-based n-to-1 kernel fusion plumbing and an example. They are not the same underlying change and should stay separate.", + "confidence": 0.86, + "canonical_issue_reason": null, + "canonical_pr_reason": "#44979 is the better canonical PR because it is the more foundational, self-contained introduction of the fusion API and tests; #45363 is a broader, more specific follow-on around KernelConfig integration and example usage.", + "best_issue_reason": null, + "best_pr_reason": "#44979 is the best PR representative for the cluster since it defines the core abstraction being introduced. #45363 is a downstream implementation/integration PR rather than the same concrete code-path fix or mergeable change set.", + "soft_edge_verdicts": [ + { + "left": "pull_request:44979", + "right": "pull_request:45363", + "accept": false, + "reason": "Related topic, but not the same change: #44979 introduces the generic Module Fusion API, while #45363 wires in KernelConfig-based n-to-1 kernel fusion and example code. They touch different code paths and would not plausibly be merged as one PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 18, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45086-2", + "nodes": [ + "pull_request:45086", + "pull_request:45317" + ], + "soft_pairs": [ + "pull_request:45086|pull_request:45317" + ], + "prepared_review_unit_hash": "07996f7c637634ea9322ce3656527843f98c8799c76a290ca3866d560dd48483", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4413, + "estimated_input_tokens": 1104, + "estimated_eval_tokens": 2464 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4413, + "estimated_input_tokens": 1104, + "estimated_eval_tokens": 2464 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both pull requests address the same Mistral regex AttributeError in `_patch_mistral_regex`; 45317 is the more complete version because it includes the fix plus tests, so it is the best representative of the cluster.", + "confidence": 0.97, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 45317 is the stronger canonical PR: it implements the same core code-path fix as 45086 and adds test coverage, making it the more complete representative of the duplicate pair.", + "best_issue_reason": null, + "best_pr_reason": "PR 45317 is the best PR to keep because it contains the same bug fix as 45086 plus validation, so it is the most useful single record for this cluster.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45086", + "right": "pull_request:45317", + "accept": true, + "reason": "Same underlying AttributeError fix in `_patch_mistral_regex` on the same file and shared issue target; the second PR appears to be a superset with tests, so they are mergeable duplicates." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 19, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45115-2", + "nodes": [ + "pull_request:45115", + "pull_request:45569" + ], + "soft_pairs": [ + "pull_request:45115|pull_request:45569" + ], + "prepared_review_unit_hash": "3e7e75e904d3a0d100ab4c19578907fbba5e573c92ee02f7e68e716d82a02cd1", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 6066, + "estimated_input_tokens": 1517, + "estimated_eval_tokens": 3290 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 6066, + "estimated_input_tokens": 1517, + "estimated_eval_tokens": 3290 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are related NemotronH refactors, but they are not the same change: one is a narrow inheritance/mapping refactor, the other is a broader model split introducing NemotronHDense/NemotronHSparse plus docs and tests.", + "confidence": 0.89, + "canonical_issue_reason": null, + "canonical_pr_reason": "pull_request:45569 is the broader, more complete implementation: it adds the dedicated dense/sparse model packages, updates auto-mapping/docs, and includes tests. pull_request:45115 is a narrower support refactor for conversion/mapping and inheritance.", + "best_issue_reason": null, + "best_pr_reason": "pull_request:45569 is the best representative PR because it captures the full architectural change and user-visible packaging/docs updates, whereas pull_request:45115 is only a smaller internal refactor.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45115", + "right": "pull_request:45569", + "accept": false, + "reason": "Shared NemotronH filenames do not imply duplicate scope: 45115 focuses on conversion mapping and inheritance refactoring, while 45569 introduces separate dense/sparse model implementations, docs, auto-mapping, and tests. They address related but different code paths and would not plausibly collapse into one PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 20, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45332-2", + "nodes": [ + "pull_request:45332", + "pull_request:45333" + ], + "soft_pairs": [ + "pull_request:45332|pull_request:45333" + ], + "prepared_review_unit_hash": "7351ab2a7b013fcc34fe7972ac6f0c83bba09ce692e7b20e3ca17b177b19f36d", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 5398, + "estimated_input_tokens": 1350, + "estimated_eval_tokens": 2956 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 5398, + "estimated_input_tokens": 1350, + "estimated_eval_tokens": 2956 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These are related but distinct PRs: one adds per-layer heterogeneous config plumbing, the other adds broader heterogeneous model/runtime support. They are not the same concrete change, so the soft edge should be rejected.", + "confidence": 0.92, + "canonical_issue_reason": null, + "canonical_pr_reason": "No true canonical duplicate here; PR 45333 is the more foundational per-layer config piece, while PR 45332 builds additional modeling/runtime support on top of that theme.", + "best_issue_reason": null, + "best_pr_reason": "PR 45333 is the cleaner representative for the shared heterogeneity/configuration thread because it introduces the core per-layer config abstraction used by the broader support work.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45332", + "right": "pull_request:45333", + "accept": false, + "reason": "Related but not duplicate: 45333 is config-layer plumbing, while 45332 is a larger heterogeneous model/runtime change. They touch the same area but do not appear to fix the same concrete code-path problem, and would not naturally collapse into one PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 21, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45415-2", + "nodes": [ + "pull_request:45415", + "pull_request:45425" + ], + "soft_pairs": [ + "pull_request:45415|pull_request:45425" + ], + "prepared_review_unit_hash": "088834af3a3151c52c465b2346ff99f7e3d7474af9b58464bc4a4702a7dc72a4", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 5512, + "estimated_input_tokens": 1378, + "estimated_eval_tokens": 3012 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 5512, + "estimated_input_tokens": 1378, + "estimated_eval_tokens": 3012 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both pull requests are type-checking related, but they address different changes: #45415 is a broad pass adding type checking and suppressing many type-ignore warnings across numerous files, while #45425 introduces typing protocols/helpers and adjusts a smaller set of call sites. They are not the same underlying fix and should not be merged as duplicates.", + "confidence": 0.94, + "canonical_issue_reason": null, + "canonical_pr_reason": "#45425 is the better canonical PR for this cluster because it is narrower and more concrete, centered on new typing abstractions and targeted call-site updates rather than a wide repository-wide type-check sweep.", + "best_issue_reason": null, + "best_pr_reason": "#45425 is the better representative PR: it has a clearer, more specific change set and a tighter code-path focus than #45415's broad type-checking cleanup.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45415", + "right": "pull_request:45425", + "accept": false, + "reason": "They share a typing theme and a few filenames, but the changes are substantively different: #45415 is a broad type-checking/config cleanup across many modules, while #45425 adds typing protocols and targeted adjustments for specific interfaces. Not the same concrete bug/change." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 22, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45487-2", + "nodes": [ + "pull_request:45487", + "pull_request:45615" + ], + "soft_pairs": [ + "pull_request:45487|pull_request:45615" + ], + "prepared_review_unit_hash": "4d8f8e6fc38d36dde1c43d759efc5af5b24d0e92cb6567d7d9803e6468bbc7d8", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 5160, + "estimated_input_tokens": 1290, + "estimated_eval_tokens": 2836 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 5160, + "estimated_input_tokens": 1290, + "estimated_eval_tokens": 2836 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "These two PRs are unrelated: one fixes a real device-placement bug in model code for AltCLIP/ChineseClip, while the other only updates Qianfan OCR XPU test expectations.", + "confidence": 0.97, + "canonical_issue_reason": null, + "canonical_pr_reason": "pull_request:45487 is the canonical PR because it changes the actual forward path/device handling in model implementation files across AltCLIP and ChineseClip, addressing a concrete runtime bug.", + "best_issue_reason": null, + "best_pr_reason": "pull_request:45487 is the best PR candidate for the cluster since it contains the substantive code fix; pull_request:45615 is test-only and does not fix the underlying code-path problem.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45487", + "right": "pull_request:45615", + "accept": false, + "reason": "Reject: these do not fix the same bug. 45487 patches model code to move token_type_ids to the correct device; 45615 only adds XPU test expectations for a different model." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 23, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45548-2", + "nodes": [ + "pull_request:45548", + "pull_request:45662" + ], + "soft_pairs": [ + "pull_request:45548|pull_request:45662" + ], + "prepared_review_unit_hash": "01b3cde4d662b289184fb928740237ba651cce03d593a2190748ec164bf777ae", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4507, + "estimated_input_tokens": 1127, + "estimated_eval_tokens": 2510 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4507, + "estimated_input_tokens": 1127, + "estimated_eval_tokens": 2510 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both PRs are about expert parallelism, but they fix different concrete failure modes: one is EP + ZeRO-3/accelerate config loading, the other is EP + FSDP2 expert parameters being overwritten during sharded loading. They are related, not duplicates.", + "confidence": 0.92, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 45662 is the better representative because it targets the concrete expert-overwrite path in sharded loading/runtime, with broader EP/FSDP2 handling changes. PR 45548 is a separate accelerate/ZeRO-3 loading/config serialization fix.", + "best_issue_reason": null, + "best_pr_reason": "PR 45662 is the strongest standalone candidate: it addresses a specific silent corruption bug for experts and includes the relevant loading and DTensor/sharding changes. PR 45548 looks like a different EP-related loading bug.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45548", + "right": "pull_request:45662", + "accept": false, + "reason": "Different bug paths and fixes: 45548 is about EP + DeepSpeed ZeRO-3 loading via accelerate launch and config serialization; 45662 is about EP + FSDP2 experts being overwritten by rank-0 broadcast during sharded loading. Shared EP context is not enough to merge them as one PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 24, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45639-3", + "nodes": [ + "pull_request:45639", + "pull_request:45645", + "pull_request:45778" + ], + "soft_pairs": [ + "pull_request:45639|pull_request:45645", + "pull_request:45645|pull_request:45778", + "pull_request:45639|pull_request:45778" + ], + "prepared_review_unit_hash": "0e6802102eca3044722f28bf7c0c1ac2815887d2f6f30ab0f23d6410c4412efb", + "component_budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 3, + "serialized_chars": 7470, + "estimated_input_tokens": 1868, + "estimated_eval_tokens": 3992 + }, + "budget": { + "node_count": 3, + "item_count": 3, + "soft_pair_count": 3, + "serialized_chars": 7470, + "estimated_input_tokens": 1868, + "estimated_eval_tokens": 3992 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "All three PRs address the same pytest-xdist/captured_info collision in testing_utils, with PR 45645 providing the broadest end-to-end fix because it also updates CI artifact collection. The relationships are best treated as overlapping fixes for the same bug, not as a strict superset claim.", + "confidence": 0.89, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR 45645 is the best canonical PR because it fixes the captured_info collision in the test helper and also updates the CI workflow that consumes those logs, making it the most complete standalone change in the cluster.", + "best_issue_reason": null, + "best_pr_reason": "PR 45645 is the strongest representative of the cluster because it covers the core xdist-safe logging fix plus the downstream artifact handling needed for CI, whereas the others are narrower variants of the same bugfix.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45639", + "right": "pull_request:45645", + "accept": true, + "reason": "Same underlying xdist captured_info collision in the same testing code path; 45645 is an overlapping, more complete fix that also adjusts CI log collection." + }, + { + "left": "pull_request:45645", + "right": "pull_request:45778", + "accept": true, + "reason": "Both target the same captured_info xdist race in testing_utils; they use different implementation details, but they are still overlapping fixes for the same bug." + }, + { + "left": "pull_request:45639", + "right": "pull_request:45778", + "accept": true, + "reason": "Both PRs fix the same concrete captured_info collision under pytest-xdist in src/transformers/testing_utils.py, so they look like alternative implementations for the same bug." + } + ] + }, + "evaluator_result": { + "accept": false, + "feedback": "The cluster is plausible, but one soft-edge verdict is too strong for the evidence. In particular, describing PR 45645 as a \"superset\" of PR 45639 is not directly supported by the packet: both touch the same xdist/captured_info area, but they appear to use different implementation details, so the relationship should be stated more conservatively as \"overlapping fix for the same bug.\"" + }, + "evaluator_used": true, + "retried": true, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 25, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "llm", + "cache_hit": false, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45683-2", + "nodes": [ + "pull_request:45683", + "pull_request:45734" + ], + "soft_pairs": [ + "pull_request:45683|pull_request:45734" + ], + "prepared_review_unit_hash": "eba4ca0f26b2d3be0c00e316270a9341d31af2208b44dbab31b1d3a8e81e7538", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4469, + "estimated_input_tokens": 1118, + "estimated_eval_tokens": 2492 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4469, + "estimated_input_tokens": 1118, + "estimated_eval_tokens": 2492 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Two open PRs touch the same quantization helper, but they fix different bugs: one special-cases audio submodules, the other changes how default skip lists are combined with user-supplied skips. They should not be clustered as duplicates.", + "confidence": 0.93, + "canonical_issue_reason": null, + "canonical_pr_reason": "45734 is the better representative PR because it fixes the core skip-module composition logic in `get_modules_to_not_convert`, which is broader and more central than the narrow audio-module special case in 45683.", + "best_issue_reason": null, + "best_pr_reason": "45734 is the strongest standalone PR in the set: it addresses a concrete quantization failure mode in the main code path and its fix is directly about the helper\u2019s behavior rather than a model-specific workaround.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45683", + "right": "pull_request:45734", + "accept": false, + "reason": "Both touch quantization skip logic, but 45683 adds audio-tower/embed_audio exclusions while 45734 changes additive handling of `skip_modules`; these are different underlying bugs and would not naturally merge into one PR." + } + ] + }, + "evaluator_result": null, + "evaluator_used": false, + "retried": false, + "accepted_nontrivial_soft_edge": false, + "error_kind": null, + "error_message": null + }, + { + "label": "pull_request", + "component_index": 26, + "component_count": 26, + "review_unit_index": 1, + "review_unit_count": 1, + "status": "reviewed", + "reason": null, + "source": "cache", + "cache_hit": true, + "model": "gpt-5.4-mini?service_tier=flex", + "cluster_id": "cluster-45737-2", + "nodes": [ + "pull_request:45737", + "pull_request:45738" + ], + "soft_pairs": [ + "pull_request:45737|pull_request:45738" + ], + "prepared_review_unit_hash": "cc2ade73a49ec6a8883a7a24c433988645588bcd049e8aad2bce56b1b24daa1c", + "component_budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4131, + "estimated_input_tokens": 1033, + "estimated_eval_tokens": 2322 + }, + "budget": { + "node_count": 2, + "item_count": 2, + "soft_pair_count": 1, + "serialized_chars": 4131, + "estimated_input_tokens": 1033, + "estimated_eval_tokens": 2322 + }, + "overflow_policy": "truncate_then_skip", + "trimmed": false, + "aggressively_trimmed": false, + "split": false, + "analyst_result": { + "summary": "Both PRs address the same MusicgenMelody cache initialization bug in `forward()`: replacing `EncoderDecoderCache(DynamicCache, DynamicCache)` with a single `DynamicCache`. PR #45737 is the more complete duplicate because it bundles the fix with a regression test, while #45738 is the fix-only variant.", + "confidence": 0.97, + "canonical_issue_reason": null, + "canonical_pr_reason": "PR #45737 is the best canonical PR because it contains the same code fix as #45738 plus a regression test, making it the most complete representation of the bug fix.", + "best_issue_reason": null, + "best_pr_reason": "PR #45737 is the strongest overall candidate: it fixes the concrete cache bug and adds a targeted regression test, whereas #45738 only changes the implementation.", + "soft_edge_verdicts": [ + { + "left": "pull_request:45737", + "right": "pull_request:45738", + "accept": true, + "reason": "Same underlying code-path bug in `musicgen_melody` cache setup; the diffs match on the core fix and the test-only addition in #45737 could plausibly live in the same PR as #45738." + } + ] + }, + "evaluator_result": { + "accept": true, + "feedback": "The assessment is grounded in the packet: both PRs make the same `musicgen_melody` cache initialization change, and #45737 only adds a regression test on top of that fix. The soft-edge accept is conservative enough because the diffs are mergeable into one PR for the same bugfix." + }, + "evaluator_used": true, + "retried": false, + "accepted_nontrivial_soft_edge": true, + "error_kind": null, + "error_message": null + } + ] +} diff --git a/snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/manifest.json b/snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/manifest.json new file mode 100644 index 0000000000000000000000000000000000000000..593940caf4ceefffb642a33dc0698cc512aa8099 --- /dev/null +++ b/snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/manifest.json @@ -0,0 +1,14 @@ +{ + "analysis_id": "hybrid-model-20260505t060044z", + "artifacts": { + "hybrid": "snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/analysis-report-hybrid.json", + "hybrid_reviews": "snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/analysis-report-hybrid.llm-reviews.json" + }, + "channel": "canonical", + "model": null, + "published_at": "2026-05-05T06:09:30Z", + "repo": "huggingface/transformers", + "schema_version": 1, + "snapshot_id": "20260505T060044Z", + "variant": "hybrid" +} diff --git a/snapshots/20260505T060044Z/manifest.json b/snapshots/20260505T060044Z/manifest.json index 477e848346af0325935101316b142440eb35d5d8..21e5659207fb9d8a8bc663d14fd38a863ee887f0 100644 --- a/snapshots/20260505T060044Z/manifest.json +++ b/snapshots/20260505T060044Z/manifest.json @@ -32,6 +32,24 @@ "timeline_events": 138 }, "extracted_at": "2026-05-05T06:00:44Z", + "published_analysis": { + "canonical_analysis_id": "hybrid-model-20260505t060044z", + "runs": { + "hybrid-model-20260505t060044z": { + "analysis_id": "hybrid-model-20260505t060044z", + "artifacts": { + "hybrid": "snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/analysis-report-hybrid.json", + "hybrid_reviews": "snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/analysis-report-hybrid.llm-reviews.json" + }, + "channel": "canonical", + "manifest_path": "snapshots/20260505T060044Z/analysis-runs/hybrid-model-20260505t060044z/manifest.json", + "model": null, + "published_at": "2026-05-05T06:09:30Z", + "variant": "hybrid" + } + }, + "schema_version": 1 + }, "repo": "huggingface/transformers", "snapshot_id": "20260505T060044Z", "watermark": {