diff --git "a/metadata_diff_model_4_19.json" "b/metadata_diff_model_4_19.json" new file mode 100644--- /dev/null +++ "b/metadata_diff_model_4_19.json" @@ -0,0 +1,17980 @@ +[ + { + "idx": 1660, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Casual_Reason/1.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1660/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1660/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1660/loser.png", + "save_id": 1227, + "prompt_en": "Remove the first book beneath the telescope.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1661, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/1.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1661/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1661/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1661/loser.png", + "save_id": 1227, + "prompt_en": "Remove the first book beneath the telescope.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1662, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Casual_Reason/1.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1662/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1662/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1662/loser.png", + "save_id": 1227, + "prompt_en": "Remove the first book beneath the telescope.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1663, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/1.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1663/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1663/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1663/loser.png", + "save_id": 1227, + "prompt_en": "Remove the first book beneath the telescope.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1664, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/3.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1664/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1664/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1664/loser.png", + "save_id": 1229, + "prompt_en": "Break the horizontally placed can of cola.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1665, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Casual_Reason/3.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1665/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1665/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1665/loser.png", + "save_id": 1229, + "prompt_en": "Break the horizontally placed can of cola.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1666, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/4.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1666/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1666/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1666/loser.png", + "save_id": 1230, + "prompt_en": "What will happen to this blood pressure monitor if the person has low blood pressure?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1667, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/4.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1667/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1667/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1667/loser.png", + "save_id": 1230, + "prompt_en": "What will happen to this blood pressure monitor if the person has low blood pressure?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1668, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "NextStep-1-HF" + ], + "source_image_ori": "Part2/Casual_Reason/4.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1668/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1668/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1668/loser.png", + "save_id": 1230, + "prompt_en": "What will happen to this blood pressure monitor if the person has low blood pressure?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1669, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/7.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1669/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1669/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1669/loser.png", + "save_id": 1233, + "prompt_en": "What will happen if you push the yellow button on the tape measure upward?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1670, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/7.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1670/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1670/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1670/loser.png", + "save_id": 1233, + "prompt_en": "What will happen if you push the yellow button on the tape measure upward?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1671, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/8.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1671/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1671/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1671/loser.png", + "save_id": 1234, + "prompt_en": "Close the lighter’s lid.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1672, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/8.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1672/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1672/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1672/loser.png", + "save_id": 1234, + "prompt_en": "Close the lighter’s lid.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1673, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "NextStep-1-HF" + ], + "source_image_ori": "Part2/Casual_Reason/8.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1673/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1673/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1673/loser.png", + "save_id": 1234, + "prompt_en": "Close the lighter’s lid.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1674, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/9.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1674/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1674/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1674/tie_2.png", + "save_id": 1235, + "prompt_en": "What will happen if you poke the balloon with a needle?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1675, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/9.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1675/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1675/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1675/loser.png", + "save_id": 1235, + "prompt_en": "What will happen if you poke the balloon with a needle?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1676, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Casual_Reason/9.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1676/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1676/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1676/loser.png", + "save_id": 1235, + "prompt_en": "What will happen if you poke the balloon with a needle?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1677, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/10.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1677/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1677/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1677/loser.png", + "save_id": 1236, + "prompt_en": "What will happen if you cut the kite’s tether line?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1678, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/10.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1678/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1678/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1678/loser.png", + "save_id": 1236, + "prompt_en": "What will happen if you cut the kite’s tether line?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1679, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/12.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1679/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1679/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1679/loser.png", + "save_id": 1238, + "prompt_en": "What will happen if you strike the nail forcefully with a hammer?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1680, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Casual_Reason/13.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1680/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1680/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1680/loser.png", + "save_id": 1239, + "prompt_en": "What will happen if the hamburger is bitten into?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1681, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/14.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1681/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1681/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1681/loser.png", + "save_id": 1240, + "prompt_en": "What happens if you keep pouring coffee into the cup?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1682, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/18.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1682/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1682/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1682/loser.png", + "save_id": 1243, + "prompt_en": "What will happen if you click ‘关机’ on the monitor, shown on the display?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1683, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Casual_Reason/18.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1683/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1683/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1683/loser.png", + "save_id": 1243, + "prompt_en": "What will happen if you click ‘关机’ on the monitor, shown on the display?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1684, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/20.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1684/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1684/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1684/loser.png", + "save_id": 1245, + "prompt_en": "What will happen if you remove the skin of a pineapple?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1685, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Casual_Reason/20.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1685/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1685/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1685/loser.png", + "save_id": 1245, + "prompt_en": "What will happen if you remove the skin of a pineapple?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1686, + "Sample_Model": [ + "EMU3.5", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/22.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1686/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1686/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1686/tie_2.png", + "save_id": 1247, + "prompt_en": "What will happen if you put the orange into the juicer and turn it on?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1687, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/22.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1687/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1687/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1687/loser.png", + "save_id": 1247, + "prompt_en": "What will happen if you put the orange into the juicer and turn it on?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1688, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/23.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1688/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1688/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1688/tie_2.png", + "save_id": 1248, + "prompt_en": "What will happen to this gauge when the driver steps on the accelerator?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1689, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/23.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1689/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1689/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1689/loser.png", + "save_id": 1248, + "prompt_en": "What will happen to this gauge when the driver steps on the accelerator?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1690, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/23.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1690/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1690/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1690/loser.png", + "save_id": 1248, + "prompt_en": "What will happen to this gauge when the driver steps on the accelerator?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1691, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Casual_Reason/23.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1691/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1691/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1691/tie_2.png", + "save_id": 1248, + "prompt_en": "What will happen to this gauge when the driver steps on the accelerator?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1692, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/25.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1692/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1692/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1692/loser.png", + "save_id": 1250, + "prompt_en": "What will happen if the boy in the picture presses the air conditioner’s power-off button?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1693, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/25.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1693/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1693/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1693/loser.png", + "save_id": 1250, + "prompt_en": "What will happen if the boy in the picture presses the air conditioner’s power-off button?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1694, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "NextStep-1-HF" + ], + "source_image_ori": "Part2/Casual_Reason/25.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1694/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1694/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1694/loser.png", + "save_id": 1250, + "prompt_en": "What will happen if the boy in the picture presses the air conditioner’s power-off button?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1695, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/27.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1695/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1695/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1695/tie_2.png", + "save_id": 1252, + "prompt_en": "What will happen to the blood pressure monitor if this person has hypertension?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1696, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/27.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1696/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1696/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1696/loser.png", + "save_id": 1252, + "prompt_en": "What will happen to the blood pressure monitor if this person has hypertension?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1697, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/27.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1697/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1697/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1697/loser.png", + "save_id": 1252, + "prompt_en": "What will happen to the blood pressure monitor if this person has hypertension?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1698, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/29.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1698/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1698/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1698/loser.png", + "save_id": 1254, + "prompt_en": "What will happen if you let go of the coin?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1699, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "NextStep-1-HF" + ], + "source_image_ori": "Part2/Casual_Reason/29.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1699/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1699/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1699/tie_2.png", + "save_id": 1254, + "prompt_en": "What will happen if you let go of the coin?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1700, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Casual_Reason/30.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1700/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1700/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1700/tie_2.png", + "save_id": 1255, + "prompt_en": "What will happen if you turn off the bathroom’s water supply?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1701, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/30.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1701/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1701/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1701/loser.png", + "save_id": 1255, + "prompt_en": "What will happen if you turn off the bathroom’s water supply?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1702, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/30.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1702/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1702/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1702/loser.png", + "save_id": 1255, + "prompt_en": "What will happen if you turn off the bathroom’s water supply?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1703, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/31.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1703/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1703/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1703/loser.png", + "save_id": 1256, + "prompt_en": "What will happen if the girl accidentally fails to hold onto the laptop?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1704, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/31.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1704/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1704/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1704/loser.png", + "save_id": 1256, + "prompt_en": "What will happen if the girl accidentally fails to hold onto the laptop?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1705, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Casual_Reason/31.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1705/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1705/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1705/tie_2.png", + "save_id": 1256, + "prompt_en": "What will happen if the girl accidentally fails to hold onto the laptop?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1706, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/32.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1706/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1706/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1706/loser.png", + "save_id": 1257, + "prompt_en": "What will happen if this person steps off the digital scale?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1707, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/34.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1707/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1707/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1707/tie_2.png", + "save_id": 1259, + "prompt_en": "What will happen if you twist the towel forcefully with your hands?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1708, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/34.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1708/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1708/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1708/loser.png", + "save_id": 1259, + "prompt_en": "What will happen if you twist the towel forcefully with your hands?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1709, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Casual_Reason/36.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1709/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1709/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1709/tie_2.png", + "save_id": 1261, + "prompt_en": "What will happen when the fire is turned off?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1710, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Casual_Reason/36.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1710/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1710/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1710/loser.png", + "save_id": 1261, + "prompt_en": "What will happen when the fire is turned off?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1711, + "Sample_Model": [ + "Qwen-Image-Edit-2511", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/37.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1711/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1711/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1711/loser.png", + "save_id": 1262, + "prompt_en": "What will happen when the rice is fully cooked?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1712, + "Sample_Model": [ + "Qwen-Image-Edit-2511", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/37.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1712/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1712/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1712/loser.png", + "save_id": 1262, + "prompt_en": "What will happen when the rice is fully cooked?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1713, + "Sample_Model": [ + "EMU3.5", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/37.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1713/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1713/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1713/tie_2.png", + "save_id": 1262, + "prompt_en": "What will happen when the rice is fully cooked?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1714, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/38.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1714/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1714/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1714/tie_2.png", + "save_id": 1263, + "prompt_en": "What will happen if the girl accidentally slips her hand?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1715, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/38.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1715/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1715/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1715/loser.png", + "save_id": 1263, + "prompt_en": "What will happen if the girl accidentally slips her hand?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1716, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/38.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1716/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1716/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1716/loser.png", + "save_id": 1263, + "prompt_en": "What will happen if the girl accidentally slips her hand?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1717, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/39.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1717/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1717/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1717/loser.png", + "save_id": 1264, + "prompt_en": "What will happen if you pinch the clip on the blue sock?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1718, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/39.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1718/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1718/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1718/loser.png", + "save_id": 1264, + "prompt_en": "What will happen if you pinch the clip on the blue sock?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1719, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/41.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1719/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1719/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1719/loser.png", + "save_id": 1266, + "prompt_en": "What will happen if the candle goes out?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1720, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Casual_Reason/41.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1720/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1720/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1720/loser.png", + "save_id": 1266, + "prompt_en": "What will happen if the candle goes out?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1721, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/44.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1721/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1721/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1721/loser.png", + "save_id": 1268, + "prompt_en": "What will happen if you drop a Mentos mint into the bottle?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1722, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Casual_Reason/44.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1722/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1722/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1722/loser.png", + "save_id": 1268, + "prompt_en": "What will happen if you drop a Mentos mint into the bottle?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1723, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Casual_Reason/44.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1723/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1723/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1723/tie_2.png", + "save_id": 1268, + "prompt_en": "What will happen if you drop a Mentos mint into the bottle?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1724, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/48.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1724/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1724/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1724/loser.png", + "save_id": 1271, + "prompt_en": "What will happen if this painting is replaced with a mirror?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1725, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/49.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1725/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1725/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1725/tie_2.png", + "save_id": 1272, + "prompt_en": "What will happen if you pull the nail out of the wall?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1726, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/49.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1726/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1726/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1726/loser.png", + "save_id": 1272, + "prompt_en": "What will happen if you pull the nail out of the wall?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1727, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/49.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1727/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1727/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1727/loser.png", + "save_id": 1272, + "prompt_en": "What will happen if you pull the nail out of the wall?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1728, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/50.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1728/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1728/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1728/tie_2.png", + "save_id": 1273, + "prompt_en": "What will happen if this person has high blood sugar?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1729, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/50.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1729/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1729/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1729/loser.png", + "save_id": 1273, + "prompt_en": "What will happen if this person has high blood sugar?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1730, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Casual_Reason/50.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1730/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1730/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1730/loser.png", + "save_id": 1273, + "prompt_en": "What will happen if this person has high blood sugar?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1731, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/51.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1731/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1731/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1731/loser.png", + "save_id": 1274, + "prompt_en": "If this is an automatic umbrella, what will happen when you press the button to open it?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1732, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/51.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1732/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1732/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1732/loser.png", + "save_id": 1274, + "prompt_en": "If this is an automatic umbrella, what will happen when you press the button to open it?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1733, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Casual_Reason/51.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1733/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1733/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1733/loser.png", + "save_id": 1274, + "prompt_en": "If this is an automatic umbrella, what will happen when you press the button to open it?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1734, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/52.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1734/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1734/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1734/loser.png", + "save_id": 1275, + "prompt_en": "What will happen if you unplug the power cord from the socket?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1735, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Casual_Reason/52.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1735/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1735/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1735/loser.png", + "save_id": 1275, + "prompt_en": "What will happen if you unplug the power cord from the socket?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1736, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/60.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1736/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1736/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1736/loser.png", + "save_id": 1282, + "prompt_en": "What will happen if you dump a bucket of water on the fire?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1737, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Casual_Reason/60.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1737/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1737/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1737/loser.png", + "save_id": 1282, + "prompt_en": "What will happen if you dump a bucket of water on the fire?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1738, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/64.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1738/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1738/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1738/tie_2.png", + "save_id": 1285, + "prompt_en": "What will happen if you drop a drop of red ink into the water?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1739, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/64.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1739/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1739/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1739/tie_2.png", + "save_id": 1285, + "prompt_en": "What will happen if you drop a drop of red ink into the water?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1740, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/65.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1740/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1740/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1740/tie_2.png", + "save_id": 1286, + "prompt_en": "Show the sneakers after being worn for a muddy hike.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1741, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/65.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1741/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1741/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1741/loser.png", + "save_id": 1286, + "prompt_en": "Show the sneakers after being worn for a muddy hike.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1742, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Casual_Reason/65.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1742/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1742/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1742/loser.png", + "save_id": 1286, + "prompt_en": "Show the sneakers after being worn for a muddy hike.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1743, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/69.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1743/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1743/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1743/loser.png", + "save_id": 1290, + "prompt_en": "What will happen if these bowls and chopsticks are washed? Please keep their positions unchanged.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1744, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/70.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1744/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1744/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1744/loser.png", + "save_id": 1291, + "prompt_en": "What will happen if this piece of butter is heated in a microwave?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1745, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Casual_Reason/73.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1745/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1745/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1745/loser.png", + "save_id": 1294, + "prompt_en": "Show the window after a baseball hits it.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1746, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/74.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1746/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1746/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1746/loser.png", + "save_id": 1295, + "prompt_en": "What happens when water is poured onto the paper shown in the image, making it wet?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1747, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/74.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1747/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1747/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1747/loser.png", + "save_id": 1295, + "prompt_en": "What happens when water is poured onto the paper shown in the image, making it wet?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1748, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Casual_Reason/74.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1748/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1748/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1748/loser.png", + "save_id": 1295, + "prompt_en": "What happens when water is poured onto the paper shown in the image, making it wet?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1749, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/77.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1749/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1749/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1749/loser.png", + "save_id": 1296, + "prompt_en": "Show the skin shortly after a mosquito bite.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1750, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Casual_Reason/77.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1750/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1750/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1750/loser.png", + "save_id": 1296, + "prompt_en": "Show the skin shortly after a mosquito bite.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1751, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Casual_Reason/77.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1751/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1751/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1751/loser.png", + "save_id": 1296, + "prompt_en": "Show the skin shortly after a mosquito bite.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1752, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/78.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1752/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1752/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1752/loser.png", + "save_id": 1297, + "prompt_en": "Show the tea bag after being used and pulled out of a cup.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1753, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/78.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1753/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1753/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1753/loser.png", + "save_id": 1297, + "prompt_en": "Show the tea bag after being used and pulled out of a cup.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1754, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Casual_Reason/78.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1754/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1754/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1754/loser.png", + "save_id": 1297, + "prompt_en": "Show the tea bag after being used and pulled out of a cup.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1755, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/79.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1755/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1755/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1755/tie_2.png", + "save_id": 1298, + "prompt_en": "Show the shirt after being ironed.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1756, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Casual_Reason/79.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1756/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1756/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1756/loser.png", + "save_id": 1298, + "prompt_en": "Show the shirt after being ironed.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1757, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/80.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1757/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1757/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1757/loser.png", + "save_id": 1299, + "prompt_en": "What will happen to the girl's hair after a balloon is repeatedly rubbed against it in the image?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1758, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Casual_Reason/80.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1758/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1758/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1758/loser.png", + "save_id": 1299, + "prompt_en": "What will happen to the girl's hair after a balloon is repeatedly rubbed against it in the image?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1759, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/81.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1759/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1759/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1759/loser.png", + "save_id": 1300, + "prompt_en": "What will the spinach in the image look like after it is cooked?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1760, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Casual_Reason/81.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1760/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1760/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1760/loser.png", + "save_id": 1300, + "prompt_en": "What will the spinach in the image look like after it is cooked?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1761, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "NextStep-1-HF" + ], + "source_image_ori": "Part2/Casual_Reason/81.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1761/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1761/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1761/loser.png", + "save_id": 1300, + "prompt_en": "What will the spinach in the image look like after it is cooked?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1762, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/82.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1762/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1762/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1762/loser.png", + "save_id": 1301, + "prompt_en": "Show the phone after being dropped on concrete.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1763, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/87.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1763/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1763/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1763/loser.png", + "save_id": 1304, + "prompt_en": "Show the holder when the paper runs out.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1764, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Casual_Reason/87.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1764/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1764/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1764/loser.png", + "save_id": 1304, + "prompt_en": "Show the holder when the paper runs out.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1765, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Casual_Reason/87.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1765/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1765/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1765/loser.png", + "save_id": 1304, + "prompt_en": "Show the holder when the paper runs out.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1766, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/91.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1766/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1766/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1766/loser.png", + "save_id": 1308, + "prompt_en": "Show the pizza after a party where most of it was eaten.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1767, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Casual_Reason/92.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1767/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1767/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1767/tie_2.png", + "save_id": 2, + "prompt_en": "Make the boy in the image pick up the teapot and pour water into the teacup.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1768, + "Sample_Model": [ + "Qwen-Image-Edit-2511", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Casual_Reason/92.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1768/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1768/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1768/loser.png", + "save_id": 2, + "prompt_en": "Make the boy in the image pick up the teapot and pour water into the teacup.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1769, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_001.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1769/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1769/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1769/loser.png", + "save_id": 1309, + "prompt_en": "What happens to a banana after one month?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1770, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_001.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1770/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1770/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1770/loser.png", + "save_id": 1309, + "prompt_en": "What happens to a banana after one month?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1771, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_001.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1771/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1771/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1771/loser.png", + "save_id": 1309, + "prompt_en": "What happens to a banana after one month?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1772, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "OmniGen2" + ], + "source_image_ori": "Part2/Time_Reason/tr_002.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1772/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1772/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1772/loser.png", + "save_id": 1310, + "prompt_en": "What will happen to this tire after it has been driven for 50,000 miles?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1773, + "Sample_Model": [ + "FLUX.2-dev", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_002.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1773/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1773/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1773/loser.png", + "save_id": 1310, + "prompt_en": "What will happen to this tire after it has been driven for 50,000 miles?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1774, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_004.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1774/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1774/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1774/loser.png", + "save_id": 1312, + "prompt_en": "Show the bread after being left in a damp room for a week.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1775, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_004.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1775/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1775/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1775/loser.png", + "save_id": 1312, + "prompt_en": "Show the bread after being left in a damp room for a week.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1776, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_004.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1776/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1776/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1776/loser.png", + "save_id": 1312, + "prompt_en": "Show the bread after being left in a damp room for a week.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1777, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_004.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1777/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1777/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1777/tie_2.png", + "save_id": 1312, + "prompt_en": "Show the bread after being left in a damp room for a week.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1778, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_006.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1778/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1778/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1778/tie_2.png", + "save_id": 1313, + "prompt_en": "Show the flowers after being neglected for two months..", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1779, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "OmniGen2" + ], + "source_image_ori": "Part2/Time_Reason/tr_006.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1779/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1779/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1779/loser.png", + "save_id": 1313, + "prompt_en": "Show the flowers after being neglected for two months..", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1780, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Time_Reason/tr_007.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1780/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1780/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1780/loser.png", + "save_id": 1314, + "prompt_en": "Show the appearance of this idle room after being neglected for many years.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1781, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "OmniGen2" + ], + "source_image_ori": "Part2/Time_Reason/tr_007.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1781/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1781/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1781/loser.png", + "save_id": 1314, + "prompt_en": "Show the appearance of this idle room after being neglected for many years.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1782, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_009.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1782/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1782/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1782/tie_2.png", + "save_id": 1316, + "prompt_en": "Show what he will look like thirty years from now.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1783, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_009.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1783/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1783/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1783/loser.png", + "save_id": 1316, + "prompt_en": "Show what he will look like thirty years from now.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1784, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_009.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1784/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1784/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1784/loser.png", + "save_id": 1316, + "prompt_en": "Show what he will look like thirty years from now.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1785, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_010.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1785/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1785/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1785/loser.png", + "save_id": 1317, + "prompt_en": "Show the man after he stopped shaving for two months.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1786, + "Sample_Model": [ + "Qwen-Image-Edit-2511", + "OmniGen2" + ], + "source_image_ori": "Part2/Time_Reason/tr_010.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1786/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1786/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1786/loser.png", + "save_id": 1317, + "prompt_en": "Show the man after he stopped shaving for two months.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1787, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_011.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1787/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1787/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1787/loser.png", + "save_id": 1318, + "prompt_en": "Show how her hair would look after not being cut for one year.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1788, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_011.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1788/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1788/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1788/loser.png", + "save_id": 1318, + "prompt_en": "Show how her hair would look after not being cut for one year.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1789, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "OmniGen2" + ], + "source_image_ori": "Part2/Time_Reason/tr_011.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1789/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1789/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1789/loser.png", + "save_id": 1318, + "prompt_en": "Show how her hair would look after not being cut for one year.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1790, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Time_Reason/tr_013.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1790/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1790/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1790/loser.png", + "save_id": 1320, + "prompt_en": "Show the skin after three weeks of healing.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1791, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_013.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1791/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1791/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1791/loser.png", + "save_id": 1320, + "prompt_en": "Show the skin after three weeks of healing.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1792, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_013.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1792/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1792/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1792/tie_2.png", + "save_id": 1320, + "prompt_en": "Show the skin after three weeks of healing.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1793, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_014.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1793/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1793/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1793/loser.png", + "save_id": 1321, + "prompt_en": "How will this person’s hands change after practicing weightlifting for many years?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1794, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "OmniGen2" + ], + "source_image_ori": "Part2/Time_Reason/tr_014.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1794/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1794/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1794/loser.png", + "save_id": 1321, + "prompt_en": "How will this person’s hands change after practicing weightlifting for many years?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1795, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_014.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1795/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1795/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1795/loser.png", + "save_id": 1321, + "prompt_en": "How will this person’s hands change after practicing weightlifting for many years?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1796, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_015.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1796/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1796/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1796/loser.png", + "save_id": 1322, + "prompt_en": "What happens when this instant noodle is cooked for 20 minutes?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1797, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "NextStep-1-HF" + ], + "source_image_ori": "Part2/Time_Reason/tr_015.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1797/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1797/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1797/loser.png", + "save_id": 1322, + "prompt_en": "What happens when this instant noodle is cooked for 20 minutes?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1798, + "Sample_Model": [ + "Qwen-Image-Edit-2511", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Time_Reason/tr_015.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1798/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1798/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1798/loser.png", + "save_id": 1322, + "prompt_en": "What happens when this instant noodle is cooked for 20 minutes?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1799, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_016.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1799/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1799/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1799/loser.png", + "save_id": 1323, + "prompt_en": "Show how this bicycle would look after being abandoned outdoors for ten years.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1800, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_016.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1800/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1800/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1800/tie_2.png", + "save_id": 1323, + "prompt_en": "Show how this bicycle would look after being abandoned outdoors for ten years.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1801, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_016.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1801/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1801/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1801/loser.png", + "save_id": 1323, + "prompt_en": "Show how this bicycle would look after being abandoned outdoors for ten years.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1802, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_018.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1802/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1802/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1802/tie_2.png", + "save_id": 1325, + "prompt_en": "Show how the scenery in the image looks in spring.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1803, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_018.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1803/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1803/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1803/loser.png", + "save_id": 1325, + "prompt_en": "Show how the scenery in the image looks in spring.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1804, + "Sample_Model": [ + "Qwen-Image-Edit-2511", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Time_Reason/tr_018.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1804/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1804/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1804/loser.png", + "save_id": 1325, + "prompt_en": "Show how the scenery in the image looks in spring.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1805, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_019.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1805/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1805/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1805/loser.png", + "save_id": 1326, + "prompt_en": "Show the newspaper after years in the sun.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1806, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_019.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1806/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1806/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1806/tie_2.png", + "save_id": 1326, + "prompt_en": "Show the newspaper after years in the sun.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1807, + "Sample_Model": [ + "EMU3.5", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_019.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1807/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1807/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1807/loser.png", + "save_id": 1326, + "prompt_en": "Show the newspaper after years in the sun.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1808, + "Sample_Model": [ + "EMU3.5", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_019.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1808/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1808/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1808/loser.png", + "save_id": 1326, + "prompt_en": "Show the newspaper after years in the sun.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1809, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_021.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1809/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1809/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1809/tie_2.png", + "save_id": 1328, + "prompt_en": "Show how this teddy bear looks after being played with by a child for many years.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1810, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_021.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1810/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1810/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1810/loser.png", + "save_id": 1328, + "prompt_en": "Show how this teddy bear looks after being played with by a child for many years.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1811, + "Sample_Model": [ + "EMU3.5", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_021.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1811/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1811/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1811/loser.png", + "save_id": 1328, + "prompt_en": "Show how this teddy bear looks after being played with by a child for many years.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1812, + "Sample_Model": [ + "FLUX.2-dev", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Time_Reason/tr_022.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1812/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1812/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1812/loser.png", + "save_id": 1329, + "prompt_en": "Depict the appearance of this soap after thirty shower uses.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1813, + "Sample_Model": [ + "FLUX.2-dev", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_022.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1813/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1813/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1813/loser.png", + "save_id": 1329, + "prompt_en": "Depict the appearance of this soap after thirty shower uses.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1814, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_023.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1814/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1814/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1814/loser.png", + "save_id": 1330, + "prompt_en": "Show how the toothpaste tube looks after being used to brush teeth 100 times.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1815, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_023.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1815/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1815/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1815/loser.png", + "save_id": 1330, + "prompt_en": "Show how the toothpaste tube looks after being used to brush teeth 100 times.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1816, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Time_Reason/tr_023.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1816/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1816/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1816/loser.png", + "save_id": 1330, + "prompt_en": "Show how the toothpaste tube looks after being used to brush teeth 100 times.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1817, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_024.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1817/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1817/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1817/loser.png", + "save_id": 1331, + "prompt_en": "Show the board after a year of cooking.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1818, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_024.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1818/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1818/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1818/loser.png", + "save_id": 1331, + "prompt_en": "Show the board after a year of cooking.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1819, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_024.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1819/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1819/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1819/loser.png", + "save_id": 1331, + "prompt_en": "Show the board after a year of cooking.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1820, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_026.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1820/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1820/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1820/loser.png", + "save_id": 1333, + "prompt_en": "What happens to her hair after six months?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1821, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_026.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1821/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1821/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1821/loser.png", + "save_id": 1333, + "prompt_en": "What happens to her hair after six months?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1822, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_026.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1822/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1822/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1822/loser.png", + "save_id": 1333, + "prompt_en": "What happens to her hair after six months?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1823, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_026.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1823/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1823/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1823/loser.png", + "save_id": 1333, + "prompt_en": "What happens to her hair after six months?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1824, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_026.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1824/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1824/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1824/loser.png", + "save_id": 1333, + "prompt_en": "What happens to her hair after six months?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1825, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_028.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1825/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1825/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1825/loser.png", + "save_id": 1335, + "prompt_en": "What happens to beer fifteen minutes later?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1826, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_028.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1826/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1826/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1826/loser.png", + "save_id": 1335, + "prompt_en": "What happens to beer fifteen minutes later?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1827, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_028.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1827/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1827/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1827/loser.png", + "save_id": 1335, + "prompt_en": "What happens to beer fifteen minutes later?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1828, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_028.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1828/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1828/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1828/loser.png", + "save_id": 1335, + "prompt_en": "What happens to beer fifteen minutes later?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1829, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_029.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1829/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1829/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1829/loser.png", + "save_id": 1336, + "prompt_en": "Show the ice cube after sitting in a warm room for 20 minutes.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1830, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_029.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1830/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1830/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1830/loser.png", + "save_id": 1336, + "prompt_en": "Show the ice cube after sitting in a warm room for 20 minutes.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1831, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_029.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1831/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1831/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1831/loser.png", + "save_id": 1336, + "prompt_en": "Show the ice cube after sitting in a warm room for 20 minutes.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1832, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Time_Reason/tr_030.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1832/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1832/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1832/loser.png", + "save_id": 1337, + "prompt_en": "How will the fireworks look three minutes from now?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1833, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_030.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1833/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1833/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1833/loser.png", + "save_id": 1337, + "prompt_en": "How will the fireworks look three minutes from now?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1834, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_030.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1834/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1834/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1834/loser.png", + "save_id": 1337, + "prompt_en": "How will the fireworks look three minutes from now?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1835, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_031.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1835/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1835/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1835/loser.png", + "save_id": 1338, + "prompt_en": "Show the appearance of the apple after three days.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1836, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_031.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1836/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1836/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1836/loser.png", + "save_id": 1338, + "prompt_en": "Show the appearance of the apple after three days.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1837, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_031.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1837/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1837/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1837/loser.png", + "save_id": 1338, + "prompt_en": "Show the appearance of the apple after three days.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1838, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_032.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1838/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1838/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1838/tie_2.png", + "save_id": 1339, + "prompt_en": "Show how the milk will look after ten minutes at room temperature.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1839, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_032.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1839/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1839/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1839/loser.png", + "save_id": 1339, + "prompt_en": "Show how the milk will look after ten minutes at room temperature.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1840, + "Sample_Model": [ + "Qwen-Image-Edit-2511", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_032.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1840/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1840/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1840/loser.png", + "save_id": 1339, + "prompt_en": "Show how the milk will look after ten minutes at room temperature.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1841, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_033.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1841/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1841/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1841/loser.png", + "save_id": 1340, + "prompt_en": "Show the egg after cooking for 6 minutes.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1842, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_033.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1842/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1842/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1842/loser.png", + "save_id": 1340, + "prompt_en": "Show the egg after cooking for 6 minutes.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1843, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_033.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1843/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1843/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1843/loser.png", + "save_id": 1340, + "prompt_en": "Show the egg after cooking for 6 minutes.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1844, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_034.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1844/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1844/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1844/loser.png", + "save_id": 1341, + "prompt_en": "Show the can after sitting on a table for 15 minutes.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1845, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_034.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1845/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1845/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1845/loser.png", + "save_id": 1341, + "prompt_en": "Show the can after sitting on a table for 15 minutes.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1846, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_034.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1846/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1846/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1846/loser.png", + "save_id": 1341, + "prompt_en": "Show the can after sitting on a table for 15 minutes.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1847, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_038.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1847/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1847/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1847/loser.png", + "save_id": 1345, + "prompt_en": "Show the screen after 8 hours of heavy gaming.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1848, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_038.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1848/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1848/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1848/loser.png", + "save_id": 1345, + "prompt_en": "Show the screen after 8 hours of heavy gaming.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1849, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_038.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1849/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1849/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1849/loser.png", + "save_id": 1345, + "prompt_en": "Show the screen after 8 hours of heavy gaming.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1850, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Time_Reason/tr_040.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1850/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1850/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1850/tie_2.png", + "save_id": 1347, + "prompt_en": "What will the road surface look like after a few days of vehicle traffic?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1851, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_040.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1851/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1851/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1851/loser.png", + "save_id": 1347, + "prompt_en": "What will the road surface look like after a few days of vehicle traffic?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1852, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Time_Reason/tr_040.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1852/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1852/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1852/loser.png", + "save_id": 1347, + "prompt_en": "What will the road surface look like after a few days of vehicle traffic?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1853, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_042.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1853/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1853/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1853/loser.png", + "save_id": 1349, + "prompt_en": "Show the pancake after 6 minutes of cooking.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1854, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_042.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1854/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1854/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1854/loser.png", + "save_id": 1349, + "prompt_en": "Show the pancake after 6 minutes of cooking.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1855, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_042.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1855/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1855/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1855/loser.png", + "save_id": 1349, + "prompt_en": "Show the pancake after 6 minutes of cooking.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1856, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_047.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1856/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1856/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1856/loser.png", + "save_id": 1352, + "prompt_en": "How will this hand sanitizer change after being used once a day for a month?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1857, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_047.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1857/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1857/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1857/loser.png", + "save_id": 1352, + "prompt_en": "How will this hand sanitizer change after being used once a day for a month?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1858, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_047.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1858/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1858/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1858/loser.png", + "save_id": 1352, + "prompt_en": "How will this hand sanitizer change after being used once a day for a month?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1859, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Time_Reason/tr_047.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1859/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1859/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1859/loser.png", + "save_id": 1352, + "prompt_en": "How will this hand sanitizer change after being used once a day for a month?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1860, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_048.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1860/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1860/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1860/loser.png", + "save_id": 1353, + "prompt_en": "Show the pencil after weeks of writing.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1861, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_048.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1861/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1861/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1861/loser.png", + "save_id": 1353, + "prompt_en": "Show the pencil after weeks of writing.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1862, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_048.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1862/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1862/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1862/loser.png", + "save_id": 1353, + "prompt_en": "Show the pencil after weeks of writing.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1863, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_049.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1863/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1863/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1863/loser.png", + "save_id": 1354, + "prompt_en": "Show the fire pit the next morning.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1864, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_049.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1864/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1864/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1864/loser.png", + "save_id": 1354, + "prompt_en": "Show the fire pit the next morning.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1865, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_049.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1865/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1865/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1865/loser.png", + "save_id": 1354, + "prompt_en": "Show the fire pit the next morning.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1866, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_050.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1866/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1866/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1866/loser.png", + "save_id": 1355, + "prompt_en": "Show the road after several hours of sunlight exposure.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1867, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_052.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1867/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1867/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1867/loser.png", + "save_id": 1356, + "prompt_en": "What will happen to the window after five hours?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1868, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_052.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1868/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1868/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1868/loser.png", + "save_id": 1356, + "prompt_en": "What will happen to the window after five hours?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1869, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_052.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1869/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1869/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1869/loser.png", + "save_id": 1356, + "prompt_en": "What will happen to the window after five hours?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1870, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_053.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1870/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1870/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1870/loser.png", + "save_id": 1357, + "prompt_en": "What will this window look like in half an hour?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1871, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_053.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1871/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1871/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1871/loser.png", + "save_id": 1357, + "prompt_en": "What will this window look like in half an hour?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1872, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_053.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1872/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1872/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1872/loser.png", + "save_id": 1357, + "prompt_en": "What will this window look like in half an hour?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1873, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_054.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1873/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1873/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1873/loser.png", + "save_id": 1358, + "prompt_en": "How will the cup look after ten minutes?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1874, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_054.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1874/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1874/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1874/loser.png", + "save_id": 1358, + "prompt_en": "How will the cup look after ten minutes?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1875, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_055.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1875/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1875/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1875/loser.png", + "save_id": 1359, + "prompt_en": "Show the sticker after years of being carried in a backpack.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1876, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_055.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1876/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1876/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1876/loser.png", + "save_id": 1359, + "prompt_en": "Show the sticker after years of being carried in a backpack.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1877, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_056.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1877/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1877/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1877/loser.png", + "save_id": 1360, + "prompt_en": "Show the strawberries after being left out for 8 days.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1878, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Time_Reason/tr_056.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1878/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1878/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1878/tie_2.png", + "save_id": 1360, + "prompt_en": "Show the strawberries after being left out for 8 days.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1879, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_056.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1879/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1879/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1879/tie_2.png", + "save_id": 1360, + "prompt_en": "Show the strawberries after being left out for 8 days.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1880, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_056.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1880/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1880/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1880/loser.png", + "save_id": 1360, + "prompt_en": "Show the strawberries after being left out for 8 days.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1881, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_059.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1881/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1881/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1881/loser.png", + "save_id": 1362, + "prompt_en": "What will this sweater look like after being worn for two years?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1882, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_059.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1882/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1882/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1882/loser.png", + "save_id": 1362, + "prompt_en": "What will this sweater look like after being worn for two years?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1883, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_059.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1883/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1883/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1883/loser.png", + "save_id": 1362, + "prompt_en": "What will this sweater look like after being worn for two years?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1884, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_060.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1884/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1884/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1884/loser.png", + "save_id": 1363, + "prompt_en": "What will the stopwatch look like after being started and running for 30 seconds?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1885, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_060.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1885/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1885/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1885/loser.png", + "save_id": 1363, + "prompt_en": "What will the stopwatch look like after being started and running for 30 seconds?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1886, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_060.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1886/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1886/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1886/loser.png", + "save_id": 1363, + "prompt_en": "What will the stopwatch look like after being started and running for 30 seconds?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1887, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_061.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1887/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1887/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1887/loser.png", + "save_id": 1364, + "prompt_en": "What did the pizza look like an hour ago before it was eaten?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1888, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_061.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1888/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1888/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1888/loser.png", + "save_id": 1364, + "prompt_en": "What did the pizza look like an hour ago before it was eaten?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1889, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_061.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1889/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1889/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1889/loser.png", + "save_id": 1364, + "prompt_en": "What did the pizza look like an hour ago before it was eaten?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1890, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_062.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1890/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1890/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1890/loser.png", + "save_id": 1365, + "prompt_en": "What did the snowman look like an hour ago when it had just been built?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1891, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_062.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1891/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1891/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1891/loser.png", + "save_id": 1365, + "prompt_en": "What did the snowman look like an hour ago when it had just been built?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1892, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_062.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1892/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1892/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1892/loser.png", + "save_id": 1365, + "prompt_en": "What did the snowman look like an hour ago when it had just been built?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1893, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Time_Reason/tr_065.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1893/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1893/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1893/tie_2.png", + "save_id": 1367, + "prompt_en": "Show the state of the puzzle pieces scattered around one hour ago.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1894, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_065.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1894/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1894/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1894/loser.png", + "save_id": 1367, + "prompt_en": "Show the state of the puzzle pieces scattered around one hour ago.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1895, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_065.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1895/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1895/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1895/loser.png", + "save_id": 1367, + "prompt_en": "Show the state of the puzzle pieces scattered around one hour ago.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1896, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_065.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1896/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1896/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1896/loser.png", + "save_id": 1367, + "prompt_en": "Show the state of the puzzle pieces scattered around one hour ago.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1897, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_065.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1897/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1897/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1897/loser.png", + "save_id": 1367, + "prompt_en": "Show the state of the puzzle pieces scattered around one hour ago.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1898, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_066.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1898/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1898/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1898/tie_2.png", + "save_id": 1368, + "prompt_en": "Show the state of this slice of bread before it was put into the toaster.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1899, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_066.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1899/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1899/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1899/loser.png", + "save_id": 1368, + "prompt_en": "Show the state of this slice of bread before it was put into the toaster.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1900, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_066.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1900/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1900/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1900/loser.png", + "save_id": 1368, + "prompt_en": "Show the state of this slice of bread before it was put into the toaster.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1901, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_066.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1901/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1901/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1901/loser.png", + "save_id": 1368, + "prompt_en": "Show the state of this slice of bread before it was put into the toaster.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1902, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_066.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1902/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1902/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1902/loser.png", + "save_id": 1368, + "prompt_en": "Show the state of this slice of bread before it was put into the toaster.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1903, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_069.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1903/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1903/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1903/loser.png", + "save_id": 1370, + "prompt_en": "Show the tower when it is half-built.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1904, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_069.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1904/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1904/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1904/loser.png", + "save_id": 1370, + "prompt_en": "Show the tower when it is half-built.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1905, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_069.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1905/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1905/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1905/loser.png", + "save_id": 1370, + "prompt_en": "Show the tower when it is half-built.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1906, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_069.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1906/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1906/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1906/loser.png", + "save_id": 1370, + "prompt_en": "Show the tower when it is half-built.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1907, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Time_Reason/tr_070.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1907/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1907/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1907/loser.png", + "save_id": 1371, + "prompt_en": "What did the box look like before it was opened?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1908, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_070.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1908/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1908/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1908/loser.png", + "save_id": 1371, + "prompt_en": "What did the box look like before it was opened?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1909, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_070.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1909/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1909/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1909/loser.png", + "save_id": 1371, + "prompt_en": "What did the box look like before it was opened?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1910, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_072.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1910/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1910/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1910/loser.png", + "save_id": 1373, + "prompt_en": "Show the state of the road before it rained.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1911, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_072.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1911/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1911/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1911/loser.png", + "save_id": 1373, + "prompt_en": "Show the state of the road before it rained.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1912, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_072.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1912/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1912/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1912/loser.png", + "save_id": 1373, + "prompt_en": "Show the state of the road before it rained.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1913, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_073.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1913/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1913/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1913/loser.png", + "save_id": 1374, + "prompt_en": "Show what the ashes looked like one hour ago.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1914, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_073.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1914/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1914/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1914/tie_2.png", + "save_id": 1374, + "prompt_en": "Show what the ashes looked like one hour ago.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1915, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_073.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1915/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1915/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1915/loser.png", + "save_id": 1374, + "prompt_en": "Show what the ashes looked like one hour ago.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1916, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_075.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1916/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1916/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1916/loser.png", + "save_id": 1375, + "prompt_en": "Show the mirror before the shower.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1917, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_075.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1917/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1917/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1917/loser.png", + "save_id": 1375, + "prompt_en": "Show the mirror before the shower.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1918, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_076.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1918/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1918/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1918/tie_2.png", + "save_id": 1376, + "prompt_en": "Show the unopened bag of chips from ten minutes ago.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1919, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_076.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1919/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1919/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1919/loser.png", + "save_id": 1376, + "prompt_en": "Show the unopened bag of chips from ten minutes ago.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1920, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_076.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1920/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1920/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1920/loser.png", + "save_id": 1376, + "prompt_en": "Show the unopened bag of chips from ten minutes ago.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1921, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_077.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1921/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1921/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1921/loser.png", + "save_id": 1377, + "prompt_en": "Show the drink before it is opened.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1922, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_077.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1922/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1922/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1922/loser.png", + "save_id": 1377, + "prompt_en": "Show the drink before it is opened.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1923, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_077.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1923/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1923/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1923/tie_2.png", + "save_id": 1377, + "prompt_en": "Show the drink before it is opened.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1924, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Time_Reason/tr_078.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1924/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1924/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1924/tie_2.png", + "save_id": 1378, + "prompt_en": "Show a house of cards when it is halfway through construction.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1925, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_078.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1925/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1925/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1925/tie_2.png", + "save_id": 1378, + "prompt_en": "Show a house of cards when it is halfway through construction.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1926, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_078.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1926/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1926/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1926/tie_2.png", + "save_id": 1378, + "prompt_en": "Show a house of cards when it is halfway through construction.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1927, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_079.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1927/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1927/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1927/loser.png", + "save_id": 1379, + "prompt_en": "Show the paper before it was crumpled, with a smooth surface.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1928, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_079.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1928/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1928/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1928/loser.png", + "save_id": 1379, + "prompt_en": "Show the paper before it was crumpled, with a smooth surface.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1929, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_079.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1929/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1929/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1929/loser.png", + "save_id": 1379, + "prompt_en": "Show the paper before it was crumpled, with a smooth surface.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1930, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_080.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1930/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1930/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1930/loser.png", + "save_id": 1380, + "prompt_en": "What about the chicken drumstick I didn’t eat ten minutes ago?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1931, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_080.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1931/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1931/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1931/loser.png", + "save_id": 1380, + "prompt_en": "What about the chicken drumstick I didn’t eat ten minutes ago?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1932, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_080.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1932/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1932/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1932/loser.png", + "save_id": 1380, + "prompt_en": "What about the chicken drumstick I didn’t eat ten minutes ago?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1933, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_081.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1933/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1933/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1933/loser.png", + "save_id": 1381, + "prompt_en": "What was the paper like before folding a thousand cranes?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1934, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_081.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1934/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1934/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1934/loser.png", + "save_id": 1381, + "prompt_en": "What was the paper like before folding a thousand cranes?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1935, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_081.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1935/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1935/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1935/loser.png", + "save_id": 1381, + "prompt_en": "What was the paper like before folding a thousand cranes?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1936, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_083.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1936/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1936/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1936/tie_2.png", + "save_id": 1383, + "prompt_en": "What does the fish I didn’t eat an hour ago look like?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1937, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_083.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1937/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1937/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1937/loser.png", + "save_id": 1383, + "prompt_en": "What does the fish I didn’t eat an hour ago look like?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1938, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_083.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1938/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1938/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1938/loser.png", + "save_id": 1383, + "prompt_en": "What does the fish I didn’t eat an hour ago look like?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1939, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_086.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1939/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1939/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1939/loser.png", + "save_id": 1384, + "prompt_en": "What does an unopened gift box look like?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1940, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_087.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1940/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1940/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1940/loser.png", + "save_id": 1385, + "prompt_en": "Here’s what this cup of milk looks like before it’s spilled.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1941, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_087.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1941/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1941/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1941/loser.png", + "save_id": 1385, + "prompt_en": "Here’s what this cup of milk looks like before it’s spilled.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1942, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Time_Reason/tr_088.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1942/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1942/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1942/tie_2.png", + "save_id": 1386, + "prompt_en": "Show the onion as it looked uncut five minutes ago.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1943, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_088.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1943/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1943/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1943/loser.png", + "save_id": 1386, + "prompt_en": "Show the onion as it looked uncut five minutes ago.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1944, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_088.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1944/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1944/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1944/loser.png", + "save_id": 1386, + "prompt_en": "Show the onion as it looked uncut five minutes ago.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1945, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_088.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1945/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1945/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1945/loser.png", + "save_id": 1386, + "prompt_en": "Show the onion as it looked uncut five minutes ago.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1946, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_090.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1946/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1946/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1946/tie_2.png", + "save_id": 1388, + "prompt_en": "What did this slice of watermelon look like ten minutes ago before it was eaten?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1947, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_090.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1947/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1947/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1947/loser.png", + "save_id": 1388, + "prompt_en": "What did this slice of watermelon look like ten minutes ago before it was eaten?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1948, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_090.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1948/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1948/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1948/loser.png", + "save_id": 1388, + "prompt_en": "What did this slice of watermelon look like ten minutes ago before it was eaten?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1949, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_091.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1949/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1949/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1949/loser.png", + "save_id": 1389, + "prompt_en": "What did this lemon look like two minutes ago before it was squeezed?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1950, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Time_Reason/tr_091.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1950/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1950/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1950/loser.png", + "save_id": 1389, + "prompt_en": "What did this lemon look like two minutes ago before it was squeezed?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1951, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Time_Reason/tr_092.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1951/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1951/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1951/loser.png", + "save_id": 1390, + "prompt_en": "What did the water in the pot look like fifteen minutes ago when it had just started boiling?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1952, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_092.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1952/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1952/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1952/loser.png", + "save_id": 1390, + "prompt_en": "What did the water in the pot look like fifteen minutes ago when it had just started boiling?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1953, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Time_Reason/tr_093.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1953/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1953/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1953/loser.png", + "save_id": 1391, + "prompt_en": "What does the medicine look like after it's finished?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1954, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_093.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1954/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1954/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1954/tie_2.png", + "save_id": 1391, + "prompt_en": "What does the medicine look like after it's finished?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1955, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Time_Reason/tr_094.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1955/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1955/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1955/tie_2.png", + "save_id": 1392, + "prompt_en": "What does the medicine look like before it's taken?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1956, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Time_Reason/tr_094.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1956/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1956/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1956/tie_2.png", + "save_id": 1392, + "prompt_en": "What does the medicine look like before it's taken?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1957, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Time_Reason/tr_094.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1957/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1957/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1957/tie_2.png", + "save_id": 1392, + "prompt_en": "What does the medicine look like before it's taken?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1958, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Chemical_Reason/1.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1958/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1958/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1958/tie_2.png", + "save_id": 1393, + "prompt_en": "Based on the blackboard hints, complete the equation for the reaction between phenol and metallic sodium.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1959, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Chemical_Reason/1.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1959/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1959/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1959/loser.png", + "save_id": 1393, + "prompt_en": "Based on the blackboard hints, complete the equation for the reaction between phenol and metallic sodium.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1960, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/3.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1960/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1960/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1960/loser.png", + "save_id": 1395, + "prompt_en": "Complete the chemical equation depicting the catalytic oxidation of isopropanol, following the blackboard notes.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1961, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/4.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1961/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1961/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1961/loser.png", + "save_id": 1396, + "prompt_en": "Deduce the missing products from the reactants and complete the equation on the blackboard.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1962, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Chemical_Reason/4.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1962/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1962/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1962/tie_2.png", + "save_id": 1396, + "prompt_en": "Deduce the missing products from the reactants and complete the equation on the blackboard.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1963, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Chemical_Reason/5.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1963/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1963/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1963/tie_2.png", + "save_id": 1397, + "prompt_en": "As suggested by the blackboard prompt, provide the completed equation for the hydrogenation of toluene.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1964, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/5.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1964/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1964/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1964/loser.png", + "save_id": 1397, + "prompt_en": "As suggested by the blackboard prompt, provide the completed equation for the hydrogenation of toluene.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1965, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Chemical_Reason/6.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1965/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1965/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1965/loser.png", + "save_id": 1398, + "prompt_en": "Fill in the blanks to complete the reaction equation between acetylene and water, using the board's hints.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1966, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Chemical_Reason/7.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1966/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1966/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1966/tie_2.png", + "save_id": 1399, + "prompt_en": "Finish the equation for the reduction reaction of nitrobenzene as indicated by the blackboard.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1967, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/7.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1967/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1967/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1967/loser.png", + "save_id": 1399, + "prompt_en": "Finish the equation for the reduction reaction of nitrobenzene as indicated by the blackboard.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1968, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Chemical_Reason/8.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1968/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1968/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1968/loser.png", + "save_id": 1400, + "prompt_en": "Refer to the blackboard hints to complete the given chemical equation.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1969, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Chemical_Reason/10.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1969/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1969/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1969/loser.png", + "save_id": 1402, + "prompt_en": "Follow the guidance on the blackboard to complete the chemical equation.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1970, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/10.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1970/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1970/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1970/loser.png", + "save_id": 1402, + "prompt_en": "Follow the guidance on the blackboard to complete the chemical equation.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1971, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Chemical_Reason/10.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1971/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1971/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1971/tie_2.png", + "save_id": 1402, + "prompt_en": "Follow the guidance on the blackboard to complete the chemical equation.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1972, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Chemical_Reason/11.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1972/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1972/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1972/tie_2.png", + "save_id": 1403, + "prompt_en": "Based on the clues provided on the board, fill in the missing parts of the chemical equation.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1973, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/11.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1973/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1973/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1973/tie_2.png", + "save_id": 1403, + "prompt_en": "Based on the clues provided on the board, fill in the missing parts of the chemical equation.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1974, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/11.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1974/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1974/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1974/tie_2.png", + "save_id": 1403, + "prompt_en": "Based on the clues provided on the board, fill in the missing parts of the chemical equation.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1975, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Chemical_Reason/12.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1975/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1975/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1975/tie_2.png", + "save_id": 1404, + "prompt_en": "Complete the specific reaction equation shown on the blackboard using the context hints.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1976, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Chemical_Reason/12.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1976/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1976/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1976/loser.png", + "save_id": 1404, + "prompt_en": "Complete the specific reaction equation shown on the blackboard using the context hints.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1977, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/12.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1977/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1977/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1977/loser.png", + "save_id": 1404, + "prompt_en": "Complete the specific reaction equation shown on the blackboard using the context hints.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1978, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/13.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1978/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1978/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1978/loser.png", + "save_id": 1405, + "prompt_en": "Finish the incomplete chemical equation as prompted by the notes on the board.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1979, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Chemical_Reason/14.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1979/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1979/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1979/loser.png", + "save_id": 1406, + "prompt_en": "Use the blackboard's prompt to finalize the chemical reaction equation.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1980, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Chemical_Reason/14.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1980/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1980/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1980/loser.png", + "save_id": 1406, + "prompt_en": "Use the blackboard's prompt to finalize the chemical reaction equation.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1981, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Chemical_Reason/15.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1981/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1981/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1981/loser.png", + "save_id": 1407, + "prompt_en": "Complete the chemical equation in accordance with the hints visible on the blackboard.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1982, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Chemical_Reason/15.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1982/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1982/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1982/loser.png", + "save_id": 1407, + "prompt_en": "Complete the chemical equation in accordance with the hints visible on the blackboard.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1983, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Chemical_Reason/16.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1983/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1983/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1983/tie_2.png", + "save_id": 1408, + "prompt_en": "What will happen if you pour concentrated sulfuric acid onto the sugar?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1984, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/16.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1984/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1984/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1984/loser.png", + "save_id": 1408, + "prompt_en": "What will happen if you pour concentrated sulfuric acid onto the sugar?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1985, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Chemical_Reason/18.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1985/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1985/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1985/tie_2.png", + "save_id": 1410, + "prompt_en": "What will happen if you apply bleach directly to the stain?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1986, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Chemical_Reason/18.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1986/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1986/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1986/loser.png", + "save_id": 1410, + "prompt_en": "What will happen if you apply bleach directly to the stain?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1987, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/19.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1987/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1987/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1987/loser.png", + "save_id": 1411, + "prompt_en": "Add the correct stoichiometric coefficients to balance this ammonia synthesis equation.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1988, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/20.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1988/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1988/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1988/loser.png", + "save_id": 1412, + "prompt_en": "Add the correct stoichiometric coefficients to balance this chemical equation.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1989, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Chemical_Reason/22.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1989/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1989/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1989/loser.png", + "save_id": 1414, + "prompt_en": "Balance the chemical equation.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1990, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Chemical_Reason/23.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1990/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1990/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1990/tie_2.png", + "save_id": 1415, + "prompt_en": "Complete the chemical equation.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1991, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/23.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1991/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1991/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1991/loser.png", + "save_id": 1415, + "prompt_en": "Complete the chemical equation.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1992, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/24.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1992/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1992/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1992/loser.png", + "save_id": 1416, + "prompt_en": "Complete the chemical equation by identifying the missing reactant.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1993, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/29.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1993/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1993/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1993/loser.png", + "save_id": 1420, + "prompt_en": "Add the correct stoichiometric coefficients to balance this laboratory chlorine preparation equation.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1994, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Chemical_Reason/38.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1994/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1994/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1994/loser.png", + "save_id": 1429, + "prompt_en": "What will happen if lemon juice is squeezed onto the blue litmus paper next to it?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1995, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/38.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1995/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1995/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1995/loser.png", + "save_id": 1429, + "prompt_en": "What will happen if lemon juice is squeezed onto the blue litmus paper next to it?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1996, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "NextStep-1-HF" + ], + "source_image_ori": "Part2/Chemical_Reason/38.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1996/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1996/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1996/loser.png", + "save_id": 1429, + "prompt_en": "What will happen if lemon juice is squeezed onto the blue litmus paper next to it?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1997, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Chemical_Reason/38.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1997/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1997/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1997/tie_2.png", + "save_id": 1429, + "prompt_en": "What will happen if lemon juice is squeezed onto the blue litmus paper next to it?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1998, + "Sample_Model": [ + "EMU3.5", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Chemical_Reason/40.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1998/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1998/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1998/loser.png", + "save_id": 1430, + "prompt_en": "What will happen if the iodine solution in the image is dropped onto the potato next to it?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 1999, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Chemical_Reason/40.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1999/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1999/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/1999/tie_2.png", + "save_id": 1430, + "prompt_en": "What will happen if the iodine solution in the image is dropped onto the potato next to it?", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2000, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/42.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2000/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2000/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2000/loser.png", + "save_id": 1432, + "prompt_en": "What will happen if the nail in the image is placed into the nearby copper sulfate solution and left there for one day?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2001, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Chemical_Reason/43.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2001/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2001/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2001/loser.png", + "save_id": 1433, + "prompt_en": "What will happen if a piece of sodium is dropped into the beaker of water shown in the image?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2002, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Chemical_Reason/46.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2002/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2002/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2002/loser.png", + "save_id": 1435, + "prompt_en": "Change the structural formula in the image to ethylene.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2003, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Chemical_Reason/48.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2003/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2003/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2003/loser.png", + "save_id": 1437, + "prompt_en": "What happens when a small amount of white vinegar is added to the red cabbage juice in the image?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2004, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "OmniGen2" + ], + "source_image_ori": "Part2/Chemical_Reason/49.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2004/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2004/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2004/loser.png", + "save_id": 1438, + "prompt_en": "What would happen if boiling water were poured into the bowl in the image?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2005, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Chemical_Reason/50.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2005/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2005/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2005/loser.png", + "save_id": 1439, + "prompt_en": "Change the flame color to correspond to the metal element found in table salt.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2006, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Chemical_Reason/52.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2006/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2006/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2006/loser.png", + "save_id": 1441, + "prompt_en": "What will happen if alum is added to this glass of water and left to stand for a period of time?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2007, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Chemical_Reason/53.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2007/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2007/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2007/loser.png", + "save_id": 1442, + "prompt_en": "What would happen if lemon juice were added to the cup of milk in the image?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2008, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Chemical_Reason/54.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2008/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2008/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2008/loser.png", + "save_id": 1443, + "prompt_en": "A beaker contains blue CuSO₄·5H₂O crystals. What will occur upon continuous heating?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2009, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/54.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2009/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2009/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2009/loser.png", + "save_id": 1443, + "prompt_en": "A beaker contains blue CuSO₄·5H₂O crystals. What will occur upon continuous heating?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2010, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Chemical_Reason/55.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2010/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2010/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2010/loser.png", + "save_id": 1444, + "prompt_en": "The graduated cylinder contains a hydrogen peroxide solution mixed with dish soap and a purple dye. What will happen when potassium iodide is added to the cylinder as a catalyst?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2011, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "OmniGen2" + ], + "source_image_ori": "Part2/Chemical_Reason/55.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2011/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2011/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2011/loser.png", + "save_id": 1444, + "prompt_en": "The graduated cylinder contains a hydrogen peroxide solution mixed with dish soap and a purple dye. What will happen when potassium iodide is added to the cylinder as a catalyst?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2012, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Chemical_Reason/55.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2012/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2012/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2012/loser.png", + "save_id": 1444, + "prompt_en": "The graduated cylinder contains a hydrogen peroxide solution mixed with dish soap and a purple dye. What will happen when potassium iodide is added to the cylinder as a catalyst?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2013, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Chemical_Reason/55.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2013/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2013/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2013/loser.png", + "save_id": 1444, + "prompt_en": "The graduated cylinder contains a hydrogen peroxide solution mixed with dish soap and a purple dye. What will happen when potassium iodide is added to the cylinder as a catalyst?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2014, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Chemical_Reason/56.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2014/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2014/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2014/loser.png", + "save_id": 1445, + "prompt_en": "When a beaker containing a small amount of purple-black solid iodine is gently heated at the bottom, what will occur?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2015, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Chemical_Reason/56.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2015/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2015/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2015/loser.png", + "save_id": 1445, + "prompt_en": "When a beaker containing a small amount of purple-black solid iodine is gently heated at the bottom, what will occur?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2016, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Chemical_Reason/57.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2016/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2016/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2016/loser.png", + "save_id": 1446, + "prompt_en": "What happens when you bend this glow stick and shake it vigorously?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2017, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Chemical_Reason/59.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2017/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2017/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2017/loser.png", + "save_id": 1448, + "prompt_en": "What will happen if activated carbon is added to this glass of water and left to stand for a period of time?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2018, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/1.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2018/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2018/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2018/loser.png", + "save_id": 1526, + "prompt_en": "Change the minimum number of cards to make it a “Three of a Kind” in Zha Jin Hua.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2019, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Game/1.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2019/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2019/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2019/loser.png", + "save_id": 1526, + "prompt_en": "Change the minimum number of cards to make it a “Three of a Kind” in Zha Jin Hua.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2020, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "NextStep-1-HF" + ], + "source_image_ori": "Part2/Game/3.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2020/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2020/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2020/loser.png", + "save_id": 1528, + "prompt_en": "Change the minimum number of cards while keeping the ranks unchanged to make it a “Flush” in Zha Jin Hua.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2021, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/3.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2021/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2021/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2021/loser.png", + "save_id": 1528, + "prompt_en": "Change the minimum number of cards while keeping the ranks unchanged to make it a “Flush” in Zha Jin Hua.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2022, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/4.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2022/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2022/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2022/loser.png", + "save_id": 1529, + "prompt_en": "Change the cards in the image to make it a “Rocket” (pair of jokers) in Dou Di Zhu.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2023, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Game/4.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2023/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2023/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2023/loser.png", + "save_id": 1529, + "prompt_en": "Change the cards in the image to make it a “Rocket” (pair of jokers) in Dou Di Zhu.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2024, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Game/5.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2024/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2024/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2024/loser.png", + "save_id": 1530, + "prompt_en": "Add a third card to make the total score exactly 21.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2025, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part2/Game/6.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2025/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2025/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2025/tie_2.png", + "save_id": 1531, + "prompt_en": "Replace the 2 of Diamonds with a card that makes the hand bust.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2026, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Game/6.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2026/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2026/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2026/loser.png", + "save_id": 1531, + "prompt_en": "Replace the 2 of Diamonds with a card that makes the hand bust.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2027, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Game/8.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2027/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2027/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2027/loser.png", + "save_id": 1533, + "prompt_en": "Change the minimum number of cards to make it a “Flush” in Texas Hold’em.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2028, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Game/9.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2028/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2028/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2028/loser.png", + "save_id": 1534, + "prompt_en": "Add a fifth card to this set to form the highest possible straight in Texas Hold’em.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2029, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "UniWorld-V1" + ], + "source_image_ori": "Part2/Game/12.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2029/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2029/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2029/tie_2.png", + "save_id": 1537, + "prompt_en": "Change the dice faces so they sum up to 7.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2030, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/12.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2030/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2030/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2030/loser.png", + "save_id": 1537, + "prompt_en": "Change the dice faces so they sum up to 7.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2031, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Game/14.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2031/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2031/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2031/loser.png", + "save_id": 1539, + "prompt_en": "Change the minimum number of cards to make it a “Bomb” in Dou Di Zhu.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2032, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Game/14.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2032/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2032/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2032/loser.png", + "save_id": 1539, + "prompt_en": "Change the minimum number of cards to make it a “Bomb” in Dou Di Zhu.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2033, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Game/19.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2033/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2033/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2033/loser.png", + "save_id": 1543, + "prompt_en": "Change the 7 to a King to make Four of a Kind.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2034, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Game/19.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2034/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2034/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2034/loser.png", + "save_id": 1543, + "prompt_en": "Change the 7 to a King to make Four of a Kind.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2035, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/21.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2035/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2035/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2035/loser.png", + "save_id": 1544, + "prompt_en": "Change the dice in the image to show the highest number.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2036, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Game/21.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2036/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2036/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2036/loser.png", + "save_id": 1544, + "prompt_en": "Change the dice in the image to show the highest number.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2037, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Game/21.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2037/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2037/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2037/loser.png", + "save_id": 1544, + "prompt_en": "Change the dice in the image to show the highest number.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2038, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/22.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2038/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2038/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2038/loser.png", + "save_id": 1545, + "prompt_en": "Change the dice in the image to show the lowest number.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2039, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Game/22.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2039/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2039/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2039/loser.png", + "save_id": 1545, + "prompt_en": "Change the dice in the image to show the lowest number.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2040, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Game/22.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2040/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2040/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2040/loser.png", + "save_id": 1545, + "prompt_en": "Change the dice in the image to show the lowest number.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2041, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Game/23.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2041/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2041/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2041/loser.png", + "save_id": 1546, + "prompt_en": "Add a fourth Mahjong tile to form a kong.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2042, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/23.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2042/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2042/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2042/tie_2.png", + "save_id": 1546, + "prompt_en": "Add a fourth Mahjong tile to form a kong.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2043, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Game/23.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2043/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2043/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2043/loser.png", + "save_id": 1546, + "prompt_en": "Add a fourth Mahjong tile to form a kong.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2044, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Game/24.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2044/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2044/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2044/loser.png", + "save_id": 1547, + "prompt_en": "With the red heart going first, where should the next move be placed to win, shown on the image?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2045, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/24.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2045/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2045/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2045/loser.png", + "save_id": 1547, + "prompt_en": "With the red heart going first, where should the next move be placed to win, shown on the image?", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2046, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/51.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2046/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2046/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2046/loser.png", + "save_id": 1573, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x2 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 4.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2047, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Game/51.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2047/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2047/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2047/loser.png", + "save_id": 1573, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x2 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 4.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2048, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/52.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2048/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2048/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2048/loser.png", + "save_id": 1574, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x2 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 4.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2049, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Game/53.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2049/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2049/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2049/loser.png", + "save_id": 1575, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x2 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 4.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2050, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Game/54.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2050/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2050/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2050/loser.png", + "save_id": 1576, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x2 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 4.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2051, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Game/54.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2051/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2051/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2051/loser.png", + "save_id": 1576, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x2 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 4.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2052, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part2/Game/55.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2052/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2052/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2052/loser.png", + "save_id": 1577, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x2 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 4.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2053, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/56.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2053/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2053/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2053/loser.png", + "save_id": 1578, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x2 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 4.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2054, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/57.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2054/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2054/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2054/loser.png", + "save_id": 1579, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2055, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Game/57.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2055/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2055/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2055/loser.png", + "save_id": 1579, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2056, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Game/57.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2056/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2056/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2056/loser.png", + "save_id": 1579, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2057, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/58.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2057/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2057/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2057/loser.png", + "save_id": 1580, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2058, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Game/58.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2058/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2058/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2058/loser.png", + "save_id": 1580, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2059, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Game/58.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2059/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2059/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2059/loser.png", + "save_id": 1580, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2060, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/59.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2060/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2060/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2060/loser.png", + "save_id": 1581, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2061, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Game/59.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2061/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2061/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2061/loser.png", + "save_id": 1581, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2062, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Game/59.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2062/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2062/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2062/loser.png", + "save_id": 1581, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2063, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/60.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2063/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2063/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2063/loser.png", + "save_id": 1582, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2064, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Game/61.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2064/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2064/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2064/loser.png", + "save_id": 1583, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2065, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/62.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2065/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2065/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2065/loser.png", + "save_id": 1584, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2066, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Game/62.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2066/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2066/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2066/loser.png", + "save_id": 1584, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2067, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Game/62.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2067/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2067/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2067/loser.png", + "save_id": 1584, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2068, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Game/62.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2068/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2068/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2068/loser.png", + "save_id": 1584, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 6.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2069, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Game/63.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2069/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2069/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2069/loser.png", + "save_id": 1585, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x4 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 8.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2070, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Game/63.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2070/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2070/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2070/loser.png", + "save_id": 1585, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x4 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 8.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2071, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/64.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2071/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2071/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2071/loser.png", + "save_id": 1586, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x4 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 8.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2072, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Game/64.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2072/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2072/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2072/loser.png", + "save_id": 1586, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x4 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 8.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2073, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Game/64.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2073/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2073/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2073/loser.png", + "save_id": 1586, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x4 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 8.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2074, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Game/65.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2074/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2074/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2074/loser.png", + "save_id": 1587, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x4 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 8.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2075, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/65.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2075/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2075/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2075/loser.png", + "save_id": 1587, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x4 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 8.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2076, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Game/65.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2076/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2076/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2076/loser.png", + "save_id": 1587, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x4 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 8.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2077, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Game/66.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2077/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2077/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2077/loser.png", + "save_id": 1588, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x4 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 8.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2078, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Game/66.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2078/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2078/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2078/loser.png", + "save_id": 1588, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 2x4 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 8.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2079, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/69.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2079/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2079/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2079/loser.png", + "save_id": 1591, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 3x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 9.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2080, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Game/69.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2080/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2080/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2080/loser.png", + "save_id": 1591, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 3x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 9.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2081, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part2/Game/69.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2081/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2081/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2081/loser.png", + "save_id": 1591, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 3x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 9.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2082, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Game/72.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2082/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2082/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2082/loser.png", + "save_id": 1594, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 3x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 9.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2083, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part2/Game/72.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2083/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2083/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2083/loser.png", + "save_id": 1594, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 3x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 9.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2084, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part2/Game/72.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2084/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2084/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2084/loser.png", + "save_id": 1594, + "prompt_en": "Draw a solution ensuring that every row, every column, and each 3x3 sub-grid delimited by thick solid lines is a unique permutation of the numbers 1 to 9.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2085, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part4/Bag/knap_000_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2085/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2085/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2085/loser.png", + "save_id": 1864, + "prompt_en": "Given a capacity of 10, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2086, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_000_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2086/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2086/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2086/loser.png", + "save_id": 1864, + "prompt_en": "Given a capacity of 10, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2087, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_000_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2087/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2087/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2087/loser.png", + "save_id": 1864, + "prompt_en": "Given a capacity of 10, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2088, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_002_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2088/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2088/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2088/loser.png", + "save_id": 1866, + "prompt_en": "Given a capacity of 11, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2089, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_002_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2089/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2089/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2089/loser.png", + "save_id": 1866, + "prompt_en": "Given a capacity of 11, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2090, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_004_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2090/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2090/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2090/tie_2.png", + "save_id": 1868, + "prompt_en": "Given a capacity of 17, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2091, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_004_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2091/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2091/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2091/loser.png", + "save_id": 1868, + "prompt_en": "Given a capacity of 17, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2092, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_004_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2092/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2092/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2092/loser.png", + "save_id": 1868, + "prompt_en": "Given a capacity of 17, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2093, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part4/Bag/knap_004_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2093/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2093/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2093/loser.png", + "save_id": 1868, + "prompt_en": "Given a capacity of 17, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2094, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_006_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2094/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2094/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2094/loser.png", + "save_id": 1870, + "prompt_en": "Given a capacity of 20, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2095, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_006_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2095/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2095/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2095/loser.png", + "save_id": 1870, + "prompt_en": "Given a capacity of 20, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2096, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_006_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2096/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2096/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2096/loser.png", + "save_id": 1870, + "prompt_en": "Given a capacity of 20, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2097, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part4/Bag/knap_006_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2097/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2097/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2097/loser.png", + "save_id": 1870, + "prompt_en": "Given a capacity of 20, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2098, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_008_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2098/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2098/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2098/tie_2.png", + "save_id": 1872, + "prompt_en": "Given a capacity of 12, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2099, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part4/Bag/knap_008_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2099/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2099/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2099/loser.png", + "save_id": 1872, + "prompt_en": "Given a capacity of 12, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2100, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_008_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2100/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2100/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2100/loser.png", + "save_id": 1872, + "prompt_en": "Given a capacity of 12, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2101, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_008_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2101/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2101/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2101/loser.png", + "save_id": 1872, + "prompt_en": "Given a capacity of 12, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2102, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_010_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2102/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2102/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2102/loser.png", + "save_id": 1874, + "prompt_en": "Given a capacity of 15, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2103, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_010_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2103/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2103/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2103/loser.png", + "save_id": 1874, + "prompt_en": "Given a capacity of 15, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2104, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_010_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2104/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2104/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2104/loser.png", + "save_id": 1874, + "prompt_en": "Given a capacity of 15, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2105, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/Bag/knap_010_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2105/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2105/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2105/loser.png", + "save_id": 1874, + "prompt_en": "Given a capacity of 15, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2106, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_012_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2106/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2106/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2106/loser.png", + "save_id": 1876, + "prompt_en": "Given a capacity of 16, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2107, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_012_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2107/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2107/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2107/loser.png", + "save_id": 1876, + "prompt_en": "Given a capacity of 16, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2108, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_012_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2108/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2108/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2108/loser.png", + "save_id": 1876, + "prompt_en": "Given a capacity of 16, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2109, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_014_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2109/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2109/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2109/loser.png", + "save_id": 1878, + "prompt_en": "Given a capacity of 16, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2110, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_014_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2110/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2110/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2110/loser.png", + "save_id": 1878, + "prompt_en": "Given a capacity of 16, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2111, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_014_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2111/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2111/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2111/loser.png", + "save_id": 1878, + "prompt_en": "Given a capacity of 16, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2112, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/Bag/knap_014_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2112/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2112/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2112/loser.png", + "save_id": 1878, + "prompt_en": "Given a capacity of 16, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2113, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_016_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2113/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2113/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2113/loser.png", + "save_id": 1880, + "prompt_en": "Given a capacity of 15, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2114, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/Bag/knap_016_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2114/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2114/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2114/loser.png", + "save_id": 1880, + "prompt_en": "Given a capacity of 15, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2115, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_016_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2115/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2115/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2115/loser.png", + "save_id": 1880, + "prompt_en": "Given a capacity of 15, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2116, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_016_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2116/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2116/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2116/loser.png", + "save_id": 1880, + "prompt_en": "Given a capacity of 15, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2117, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_020_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2117/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2117/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2117/tie_2.png", + "save_id": 1884, + "prompt_en": "Given a capacity of 17, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2118, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_020_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2118/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2118/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2118/loser.png", + "save_id": 1884, + "prompt_en": "Given a capacity of 17, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2119, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_020_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2119/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2119/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2119/loser.png", + "save_id": 1884, + "prompt_en": "Given a capacity of 17, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2120, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_022_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2120/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2120/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2120/loser.png", + "save_id": 1886, + "prompt_en": "Given a capacity of 14, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2121, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_022_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2121/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2121/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2121/loser.png", + "save_id": 1886, + "prompt_en": "Given a capacity of 14, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2122, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_024_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2122/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2122/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2122/loser.png", + "save_id": 1888, + "prompt_en": "Given a capacity of 18, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2123, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_024_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2123/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2123/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2123/loser.png", + "save_id": 1888, + "prompt_en": "Given a capacity of 18, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2124, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_024_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2124/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2124/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2124/loser.png", + "save_id": 1888, + "prompt_en": "Given a capacity of 18, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2125, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_026_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2125/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2125/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2125/loser.png", + "save_id": 1890, + "prompt_en": "Given a capacity of 17, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2126, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_026_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2126/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2126/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2126/loser.png", + "save_id": 1890, + "prompt_en": "Given a capacity of 17, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2127, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_030_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2127/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2127/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2127/tie_2.png", + "save_id": 1894, + "prompt_en": "Given a capacity of 12, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2128, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_030_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2128/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2128/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2128/loser.png", + "save_id": 1894, + "prompt_en": "Given a capacity of 12, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2129, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/Bag/knap_030_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2129/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2129/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2129/loser.png", + "save_id": 1894, + "prompt_en": "Given a capacity of 12, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2130, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_032_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2130/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2130/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2130/loser.png", + "save_id": 1896, + "prompt_en": "Given a capacity of 11, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2131, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_034_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2131/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2131/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2131/loser.png", + "save_id": 1898, + "prompt_en": "Given a capacity of 14, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2132, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_036_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2132/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2132/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2132/loser.png", + "save_id": 1900, + "prompt_en": "Given a capacity of 11, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2133, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "OmniGen2" + ], + "source_image_ori": "Part4/Bag/knap_036_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2133/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2133/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2133/loser.png", + "save_id": 1900, + "prompt_en": "Given a capacity of 11, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2134, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_036_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2134/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2134/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2134/loser.png", + "save_id": 1900, + "prompt_en": "Given a capacity of 11, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2135, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_038_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2135/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2135/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2135/loser.png", + "save_id": 1902, + "prompt_en": "Given a capacity of 20, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2136, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_038_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2136/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2136/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2136/loser.png", + "save_id": 1902, + "prompt_en": "Given a capacity of 20, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2137, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_040_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2137/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2137/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2137/loser.png", + "save_id": 1904, + "prompt_en": "Given a capacity of 15, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2138, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_040_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2138/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2138/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2138/loser.png", + "save_id": 1904, + "prompt_en": "Given a capacity of 15, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2139, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_041_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2139/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2139/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2139/tie_2.png", + "save_id": 1905, + "prompt_en": "Given a capacity of 17, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2140, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_041_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2140/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2140/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2140/loser.png", + "save_id": 1905, + "prompt_en": "Given a capacity of 17, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2141, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_042_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2141/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2141/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2141/loser.png", + "save_id": 1906, + "prompt_en": "Given a capacity of 19, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2142, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part4/Bag/knap_042_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2142/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2142/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2142/loser.png", + "save_id": 1906, + "prompt_en": "Given a capacity of 19, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2143, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_043_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2143/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2143/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2143/loser.png", + "save_id": 1907, + "prompt_en": "Given a capacity of 12, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2144, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_044_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2144/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2144/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2144/loser.png", + "save_id": 1908, + "prompt_en": "Given a capacity of 10, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2145, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_044_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2145/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2145/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2145/loser.png", + "save_id": 1908, + "prompt_en": "Given a capacity of 10, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2146, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_045_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2146/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2146/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2146/tie_2.png", + "save_id": 1909, + "prompt_en": "Given a capacity of 10, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2147, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/Bag/knap_045_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2147/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2147/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2147/loser.png", + "save_id": 1909, + "prompt_en": "Given a capacity of 10, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2148, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_046_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2148/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2148/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2148/loser.png", + "save_id": 1910, + "prompt_en": "Given a capacity of 17, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2149, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/Bag/knap_047_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2149/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2149/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2149/loser.png", + "save_id": 1911, + "prompt_en": "Given a capacity of 20, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2150, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_048_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2150/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2150/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2150/loser.png", + "save_id": 1912, + "prompt_en": "Given a capacity of 10, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2151, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_048_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2151/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2151/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2151/loser.png", + "save_id": 1912, + "prompt_en": "Given a capacity of 10, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2152, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_049_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2152/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2152/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2152/loser.png", + "save_id": 1913, + "prompt_en": "Given a capacity of 19, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2153, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/Bag/knap_050_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2153/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2153/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2153/loser.png", + "save_id": 1914, + "prompt_en": "Given a capacity of 17, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2154, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_050_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2154/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2154/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2154/loser.png", + "save_id": 1914, + "prompt_en": "Given a capacity of 17, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2155, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_051_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2155/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2155/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2155/loser.png", + "save_id": 1915, + "prompt_en": "Given a capacity of 15, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2156, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_052_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2156/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2156/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2156/loser.png", + "save_id": 1916, + "prompt_en": "Given a capacity of 16, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2157, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_053_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2157/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2157/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2157/loser.png", + "save_id": 1917, + "prompt_en": "Given a capacity of 14, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2158, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/Bag/knap_054_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2158/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2158/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2158/loser.png", + "save_id": 1918, + "prompt_en": "Given a capacity of 19, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2159, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "OmniGen2" + ], + "source_image_ori": "Part4/Bag/knap_055_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2159/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2159/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2159/loser.png", + "save_id": 1919, + "prompt_en": "Given a capacity of 12, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2160, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/Bag/knap_055_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2160/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2160/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2160/loser.png", + "save_id": 1919, + "prompt_en": "Given a capacity of 12, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2161, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/Bag/knap_056_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2161/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2161/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2161/loser.png", + "save_id": 1920, + "prompt_en": "Given a capacity of 16, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2162, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/Bag/knap_057_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2162/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2162/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2162/loser.png", + "save_id": 1921, + "prompt_en": "Given a capacity of 10, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2163, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_058_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2163/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2163/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2163/loser.png", + "save_id": 1922, + "prompt_en": "Given a capacity of 13, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2164, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Bag/knap_058_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2164/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2164/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2164/loser.png", + "save_id": 1922, + "prompt_en": "Given a capacity of 13, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2165, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Bag/knap_059_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2165/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2165/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2165/loser.png", + "save_id": 1923, + "prompt_en": "Given a capacity of 15, highlight the items in green that maximize total value without exceeding the weight limit.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2166, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_000_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2166/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2166/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2166/loser.png", + "save_id": 1964, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2167, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_000_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2167/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2167/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2167/loser.png", + "save_id": 1964, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2168, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_000_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2168/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2168/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2168/loser.png", + "save_id": 1964, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2169, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_002_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2169/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2169/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2169/loser.png", + "save_id": 1966, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2170, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_006_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2170/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2170/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2170/loser.png", + "save_id": 1970, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2171, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_008_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2171/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2171/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2171/loser.png", + "save_id": 1972, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2172, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_012_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2172/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2172/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2172/loser.png", + "save_id": 1976, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2173, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_014_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2173/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2173/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2173/loser.png", + "save_id": 1978, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2174, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_016_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2174/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2174/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2174/loser.png", + "save_id": 1980, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2175, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_018_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2175/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2175/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2175/loser.png", + "save_id": 1982, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2176, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_022_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2176/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2176/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2176/loser.png", + "save_id": 1986, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2177, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_028_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2177/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2177/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2177/loser.png", + "save_id": 1992, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2178, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_036_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2178/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2178/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2178/loser.png", + "save_id": 2000, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2179, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_038_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2179/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2179/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2179/loser.png", + "save_id": 2002, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2180, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_040_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2180/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2180/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2180/loser.png", + "save_id": 2004, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2181, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_041_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2181/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2181/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2181/loser.png", + "save_id": 2005, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2182, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_048_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2182/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2182/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2182/loser.png", + "save_id": 2012, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2183, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_053_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2183/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2183/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2183/loser.png", + "save_id": 2017, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2184, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/dataset_convex_hull/hull_058_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2184/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2184/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2184/loser.png", + "save_id": 2022, + "prompt_en": "Connect the outermost points to form the minimum convex polygon that encloses all points in the image.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2185, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_000_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2185/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2185/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2185/loser.png", + "save_id": 2064, + "prompt_en": "Find the 2x4 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2186, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_002_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2186/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2186/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2186/loser.png", + "save_id": 2066, + "prompt_en": "Find the 4x3 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2187, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_002_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2187/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2187/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2187/loser.png", + "save_id": 2066, + "prompt_en": "Find the 4x3 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2188, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_004_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2188/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2188/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2188/loser.png", + "save_id": 2068, + "prompt_en": "Find the 4x2 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2189, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_006_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2189/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2189/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2189/loser.png", + "save_id": 2070, + "prompt_en": "Find the 4x4 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2190, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_008_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2190/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2190/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2190/loser.png", + "save_id": 2072, + "prompt_en": "Find the 2x2 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2191, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_010_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2191/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2191/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2191/loser.png", + "save_id": 2074, + "prompt_en": "Find the 4x3 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2192, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_010_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2192/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2192/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2192/loser.png", + "save_id": 2074, + "prompt_en": "Find the 4x3 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2193, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_012_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2193/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2193/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2193/loser.png", + "save_id": 2076, + "prompt_en": "Find the 3x2 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2194, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_012_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2194/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2194/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2194/loser.png", + "save_id": 2076, + "prompt_en": "Find the 3x2 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2195, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_014_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2195/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2195/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2195/loser.png", + "save_id": 2078, + "prompt_en": "Find the 3x4 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2196, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_014_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2196/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2196/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2196/loser.png", + "save_id": 2078, + "prompt_en": "Find the 3x4 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2197, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_016_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2197/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2197/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2197/loser.png", + "save_id": 2080, + "prompt_en": "Find the 2x3 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2198, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_018_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2198/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2198/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2198/loser.png", + "save_id": 2082, + "prompt_en": "Find the 3x4 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2199, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_022_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2199/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2199/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2199/loser.png", + "save_id": 2086, + "prompt_en": "Find the 3x3 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2200, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_026_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2200/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2200/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2200/loser.png", + "save_id": 2090, + "prompt_en": "Find the 2x4 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2201, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_028_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2201/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2201/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2201/loser.png", + "save_id": 2092, + "prompt_en": "Find the 3x2 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2202, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_030_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2202/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2202/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2202/loser.png", + "save_id": 2094, + "prompt_en": "Find the 4x4 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2203, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_032_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2203/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2203/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2203/loser.png", + "save_id": 2096, + "prompt_en": "Find the 2x3 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2204, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_036_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2204/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2204/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2204/loser.png", + "save_id": 2100, + "prompt_en": "Find the 4x4 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2205, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_038_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2205/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2205/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2205/loser.png", + "save_id": 2102, + "prompt_en": "Find the 3x3 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2206, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_040_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2206/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2206/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2206/loser.png", + "save_id": 2104, + "prompt_en": "Find the 2x4 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2207, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_041_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2207/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2207/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2207/loser.png", + "save_id": 2105, + "prompt_en": "Find the 2x4 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2208, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_042_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2208/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2208/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2208/loser.png", + "save_id": 2106, + "prompt_en": "Find the 2x4 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2209, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_043_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2209/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2209/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2209/loser.png", + "save_id": 2107, + "prompt_en": "Find the 4x4 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2210, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_044_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2210/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2210/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2210/loser.png", + "save_id": 2108, + "prompt_en": "Find the 3x3 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2211, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_045_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2211/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2211/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2211/loser.png", + "save_id": 2109, + "prompt_en": "Find the 4x2 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2212, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_046_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2212/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2212/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2212/loser.png", + "save_id": 2110, + "prompt_en": "Find the 4x2 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2213, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_048_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2213/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2213/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2213/loser.png", + "save_id": 2112, + "prompt_en": "Find the 3x3 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2214, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/dataset_max_submatrix/conv_049_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2214/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2214/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2214/loser.png", + "save_id": 2113, + "prompt_en": "Find the 3x3 rectangular region in the image with the highest sum of values and highlight this region in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2215, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_no_start/global_002_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2215/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2215/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2215/loser.png", + "save_id": 2166, + "prompt_en": "Help me find the longest word in this image. It can only go down or right. Please highlight the path in green.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2216, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_no_start/global_006_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2216/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2216/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2216/loser.png", + "save_id": 2170, + "prompt_en": "Help me find the longest word in this image. It can only go down or right. Please highlight the path in green.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2217, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_no_start/global_044_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2217/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2217/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2217/loser.png", + "save_id": 2208, + "prompt_en": "Help me find the longest word in this image. It can only go down or right. Please highlight the path in green.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2218, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_no_start/global_056_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2218/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2218/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2218/loser.png", + "save_id": 2220, + "prompt_en": "Help me find the longest word in this image. It can only go down or right. Please highlight the path in green.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2219, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_no_start/global_060_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2219/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2219/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2219/loser.png", + "save_id": 2224, + "prompt_en": "Help me find the longest word in this image. It can only go down or right. Please highlight the path in green.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2220, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_no_start/global_066_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2220/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2220/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2220/loser.png", + "save_id": 2230, + "prompt_en": "Help me find the longest word in this image. It can only go down or right. Please highlight the path in green.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2221, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_no_start/global_068_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2221/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2221/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2221/loser.png", + "save_id": 2232, + "prompt_en": "Help me find the longest word in this image. It can only go down or right. Please highlight the path in green.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2222, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Longest_Word_with_start/task_start_001_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2222/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2222/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2222/loser.png", + "save_id": 2265, + "prompt_en": "Starting from the yellow cell, find the longest valid English word by moving only Down or Right, and highlight its path in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2223, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_with_start/task_start_005_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2223/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2223/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2223/loser.png", + "save_id": 2269, + "prompt_en": "Starting from the yellow cell, find the longest valid English word by moving only Down or Right, and highlight its path in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2224, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_with_start/task_start_011_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2224/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2224/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2224/loser.png", + "save_id": 2275, + "prompt_en": "Starting from the yellow cell, find the longest valid English word by moving only Down or Right, and highlight its path in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2225, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_with_start/task_start_021_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2225/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2225/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2225/loser.png", + "save_id": 2285, + "prompt_en": "Starting from the yellow cell, find the longest valid English word by moving only Down or Right, and highlight its path in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2226, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_with_start/task_start_023_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2226/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2226/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2226/loser.png", + "save_id": 2287, + "prompt_en": "Starting from the yellow cell, find the longest valid English word by moving only Down or Right, and highlight its path in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2227, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_with_start/task_start_025_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2227/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2227/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2227/loser.png", + "save_id": 2289, + "prompt_en": "Starting from the yellow cell, find the longest valid English word by moving only Down or Right, and highlight its path in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2228, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_with_start/task_start_035_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2228/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2228/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2228/loser.png", + "save_id": 2299, + "prompt_en": "Starting from the yellow cell, find the longest valid English word by moving only Down or Right, and highlight its path in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2229, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_with_start/task_start_039_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2229/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2229/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2229/loser.png", + "save_id": 2303, + "prompt_en": "Starting from the yellow cell, find the longest valid English word by moving only Down or Right, and highlight its path in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2230, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Longest_Word_with_start/task_start_041_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2230/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2230/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2230/loser.png", + "save_id": 2305, + "prompt_en": "Starting from the yellow cell, find the longest valid English word by moving only Down or Right, and highlight its path in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2231, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Longest_Word_with_start/task_start_043_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2231/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2231/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2231/loser.png", + "save_id": 2307, + "prompt_en": "Starting from the yellow cell, find the longest valid English word by moving only Down or Right, and highlight its path in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2232, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/Longest_Word_with_start/task_start_059_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2232/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2232/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2232/loser.png", + "save_id": 2323, + "prompt_en": "Starting from the yellow cell, find the longest valid English word by moving only Down or Right, and highlight its path in yellow.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2233, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_000_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2233/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2233/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2233/loser.png", + "save_id": 2364, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2234, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_004_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2234/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2234/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2234/loser.png", + "save_id": 2368, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2235, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_008_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2235/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2235/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2235/loser.png", + "save_id": 2372, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2236, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_012_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2236/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2236/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2236/loser.png", + "save_id": 2376, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2237, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/max_bonus/flex_024_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2237/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2237/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2237/loser.png", + "save_id": 2388, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2238, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/max_bonus/flex_028_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2238/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2238/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2238/loser.png", + "save_id": 2392, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2239, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_036_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2239/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2239/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2239/loser.png", + "save_id": 2400, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2240, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_040_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2240/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2240/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2240/loser.png", + "save_id": 2404, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2241, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_048_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2241/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2241/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2241/loser.png", + "save_id": 2412, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2242, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_052_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2242/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2242/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2242/loser.png", + "save_id": 2416, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2243, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_058_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2243/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2243/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2243/loser.png", + "save_id": 2422, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2244, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_060_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2244/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2244/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2244/loser.png", + "save_id": 2424, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2245, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_061_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2245/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2245/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2245/loser.png", + "save_id": 2425, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2246, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_062_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2246/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2246/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2246/loser.png", + "save_id": 2426, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2247, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_063_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2247/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2247/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2247/loser.png", + "save_id": 2427, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2248, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_064_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2248/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2248/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2248/loser.png", + "save_id": 2428, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2249, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_065_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2249/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2249/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2249/loser.png", + "save_id": 2429, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2250, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_066_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2250/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2250/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2250/loser.png", + "save_id": 2430, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2251, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_067_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2251/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2251/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2251/loser.png", + "save_id": 2431, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2252, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/max_bonus/flex_072_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2252/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2252/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2252/loser.png", + "save_id": 2436, + "prompt_en": "Color a path from the blue 'S' to the red 'E' in yellow. You can only move down or right. Passing through each square acquires the number inside it. The goal is to maximize the sum of all numbers collected along the path.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2253, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_000_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2253/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2253/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2253/loser.png", + "save_id": 2484, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2254, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_004_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2254/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2254/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2254/loser.png", + "save_id": 2488, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2255, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_012_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2255/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2255/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2255/loser.png", + "save_id": 2496, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2256, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_016_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2256/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2256/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2256/loser.png", + "save_id": 2500, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2257, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_018_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2257/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2257/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2257/loser.png", + "save_id": 2502, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2258, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_020_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2258/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2258/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2258/loser.png", + "save_id": 2504, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2259, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_028_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2259/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2259/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2259/loser.png", + "save_id": 2512, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2260, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_030_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2260/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2260/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2260/loser.png", + "save_id": 2514, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2261, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_034_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2261/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2261/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2261/loser.png", + "save_id": 2518, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2262, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_038_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2262/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2262/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2262/loser.png", + "save_id": 2522, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2263, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_047_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2263/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2263/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2263/loser.png", + "save_id": 2531, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2264, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_048_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2264/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2264/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2264/loser.png", + "save_id": 2532, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2265, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_051_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2265/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2265/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2265/loser.png", + "save_id": 2535, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2266, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_052_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2266/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2266/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2266/loser.png", + "save_id": 2536, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2267, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_055_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2267/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2267/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2267/tie_2.png", + "save_id": 2539, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2268, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_057_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2268/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2268/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2268/loser.png", + "save_id": 2541, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2269, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/numberlink/flow_visual_060_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2269/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2269/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2269/tie_2.png", + "save_id": 2544, + "prompt_en": "Connect circles of the same color. Paths may move only horizontally or vertically, and each line must match the color of its circles. Paths of different colors must not intersect.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2270, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Shortest_path2/rpg_000_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2270/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2270/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2270/loser.png", + "save_id": 4222, + "prompt_en": "Draw a lowest-cost path from S to E using red arrows. Movement is restricted to horizontal or vertical directions only, and the path may pass only through gray, green, and blue cells. The cost is 1 for gray cells, 3 for green cells, and 8 for blue cells.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2271, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Shortest_path2/rpg_003_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2271/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2271/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2271/loser.png", + "save_id": 4225, + "prompt_en": "Draw a lowest-cost path from S to E using red arrows. Movement is restricted to horizontal or vertical directions only, and the path may pass only through gray, green, and blue cells. The cost is 1 for gray cells, 3 for green cells, and 8 for blue cells.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2272, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Shortest_path2/rpg_004_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2272/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2272/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2272/loser.png", + "save_id": 4226, + "prompt_en": "Draw a lowest-cost path from S to E using red arrows. Movement is restricted to horizontal or vertical directions only, and the path may pass only through gray, green, and blue cells. The cost is 1 for gray cells, 3 for green cells, and 8 for blue cells.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2273, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Shortest_path2/rpg_006_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2273/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2273/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2273/loser.png", + "save_id": 4228, + "prompt_en": "Draw a lowest-cost path from S to E using red arrows. Movement is restricted to horizontal or vertical directions only, and the path may pass only through gray, green, and blue cells. The cost is 1 for gray cells, 3 for green cells, and 8 for blue cells.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2274, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Shortest_path2/rpg_007_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2274/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2274/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2274/loser.png", + "save_id": 4229, + "prompt_en": "Draw a lowest-cost path from S to E using red arrows. Movement is restricted to horizontal or vertical directions only, and the path may pass only through gray, green, and blue cells. The cost is 1 for gray cells, 3 for green cells, and 8 for blue cells.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2275, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Shortest_path2/rpg_008_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2275/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2275/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2275/loser.png", + "save_id": 4230, + "prompt_en": "Draw a lowest-cost path from S to E using red arrows. Movement is restricted to horizontal or vertical directions only, and the path may pass only through gray, green, and blue cells. The cost is 1 for gray cells, 3 for green cells, and 8 for blue cells.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2276, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Shortest_path2/rpg_010_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2276/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2276/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2276/loser.png", + "save_id": 4232, + "prompt_en": "Draw a lowest-cost path from S to E using red arrows. Movement is restricted to horizontal or vertical directions only, and the path may pass only through gray, green, and blue cells. The cost is 1 for gray cells, 3 for green cells, and 8 for blue cells.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2277, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Shortest_path2/rpg_017_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2277/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2277/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2277/loser.png", + "save_id": 4239, + "prompt_en": "Draw a lowest-cost path from S to E using red arrows. Movement is restricted to horizontal or vertical directions only, and the path may pass only through gray, green, and blue cells. The cost is 1 for gray cells, 3 for green cells, and 8 for blue cells.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2278, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Shortest_path2/rpg_021_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2278/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2278/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2278/loser.png", + "save_id": 4243, + "prompt_en": "Draw a lowest-cost path from S to E using red arrows. Movement is restricted to horizontal or vertical directions only, and the path may pass only through gray, green, and blue cells. The cost is 1 for gray cells, 3 for green cells, and 8 for blue cells.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2279, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Shortest_path2/rpg_025_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2279/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2279/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2279/loser.png", + "save_id": 4247, + "prompt_en": "Draw a lowest-cost path from S to E using red arrows. Movement is restricted to horizontal or vertical directions only, and the path may pass only through gray, green, and blue cells. The cost is 1 for gray cells, 3 for green cells, and 8 for blue cells.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2280, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Shortest_path2/rpg_026_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2280/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2280/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2280/loser.png", + "save_id": 4248, + "prompt_en": "Draw a lowest-cost path from S to E using red arrows. Movement is restricted to horizontal or vertical directions only, and the path may pass only through gray, green, and blue cells. The cost is 1 for gray cells, 3 for green cells, and 8 for blue cells.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2281, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Shortest_path2/rpg_027_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2281/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2281/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2281/loser.png", + "save_id": 4249, + "prompt_en": "Draw a lowest-cost path from S to E using red arrows. Movement is restricted to horizontal or vertical directions only, and the path may pass only through gray, green, and blue cells. The cost is 1 for gray cells, 3 for green cells, and 8 for blue cells.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2282, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/Shortest_path2/rpg_032_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2282/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2282/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2282/loser.png", + "save_id": 4254, + "prompt_en": "Draw a lowest-cost path from S to E using red arrows. Movement is restricted to horizontal or vertical directions only, and the path may pass only through gray, green, and blue cells. The cost is 1 for gray cells, 3 for green cells, and 8 for blue cells.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2283, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_no_start/task_001_DOG_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2283/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2283/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2283/loser.png", + "save_id": 2685, + "prompt_en": "Find the word 'DOG'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2284, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part4/word_no_start/task_003_MAP_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2284/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2284/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2284/loser.png", + "save_id": 2687, + "prompt_en": "Find the word 'MAP'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2285, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_no_start/task_005_BOX_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2285/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2285/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2285/loser.png", + "save_id": 2689, + "prompt_en": "Find the word 'BOX'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2286, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/word_no_start/task_007_EYE_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2286/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2286/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2286/loser.png", + "save_id": 2691, + "prompt_en": "Find the word 'EYE'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2287, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part4/word_no_start/task_009_TEA_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2287/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2287/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2287/loser.png", + "save_id": 2693, + "prompt_en": "Find the word 'TEA'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2288, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_no_start/task_011_FISH_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2288/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2288/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2288/loser.png", + "save_id": 2695, + "prompt_en": "Find the word 'FISH'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2289, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/word_no_start/task_013_MILK_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2289/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2289/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2289/loser.png", + "save_id": 2697, + "prompt_en": "Find the word 'MILK'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2290, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_no_start/task_015_DUCK_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2290/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2290/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2290/loser.png", + "save_id": 2699, + "prompt_en": "Find the word 'DUCK'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2291, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part4/word_no_start/task_017_GOAT_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2291/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2291/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2291/loser.png", + "save_id": 2701, + "prompt_en": "Find the word 'GOAT'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2292, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/word_no_start/task_019_CODE_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2292/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2292/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2292/loser.png", + "save_id": 2703, + "prompt_en": "Find the word 'CODE'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2293, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_no_start/task_041_MONKEY_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2293/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2293/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2293/loser.png", + "save_id": 2725, + "prompt_en": "Find the word 'MONKEY'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2294, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part4/word_no_start/task_043_ORANGE_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2294/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2294/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2294/loser.png", + "save_id": 2727, + "prompt_en": "Find the word 'ORANGE'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2295, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/word_no_start/task_045_DOCTOR_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2295/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2295/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2295/loser.png", + "save_id": 2729, + "prompt_en": "Find the word 'DOCTOR'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2296, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_no_start/task_047_FRIEND_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2296/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2296/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2296/loser.png", + "save_id": 2731, + "prompt_en": "Find the word 'FRIEND'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2297, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_no_start/task_049_PYTHON_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2297/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2297/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2297/loser.png", + "save_id": 2733, + "prompt_en": "Find the word 'PYTHON'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2298, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/word_no_start/task_051_GARDEN_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2298/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2298/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2298/loser.png", + "save_id": 2735, + "prompt_en": "Find the word 'GARDEN'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2299, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part4/word_no_start/task_053_APPLE_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2299/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2299/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2299/loser.png", + "save_id": 2737, + "prompt_en": "Find the word 'APPLE'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2300, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_no_start/task_055_LEMON_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2300/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2300/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2300/loser.png", + "save_id": 2739, + "prompt_en": "Find the word 'LEMON'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2301, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part4/word_no_start/task_057_RIVER_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2301/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2301/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2301/loser.png", + "save_id": 2741, + "prompt_en": "Find the word 'RIVER'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2302, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/word_no_start/task_059_CLOUD_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2302/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2302/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2302/loser.png", + "save_id": 2743, + "prompt_en": "Find the word 'CLOUD'. Highlight its characters in red. Note: The word flows only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2303, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part4/word_with_start/task_001_DOG_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2303/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2303/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2303/loser.png", + "save_id": 2785, + "prompt_en": "Complete the red path for the word 'DOG' starting from the red cell. Move only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2304, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Bagel" + ], + "source_image_ori": "Part4/word_with_start/task_003_MAP_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2304/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2304/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2304/loser.png", + "save_id": 2787, + "prompt_en": "Complete the red path for the word 'MAP' starting from the red cell. Move only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2305, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "FLUX.2-dev" + ], + "source_image_ori": "Part4/word_with_start/task_005_BOX_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2305/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2305/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2305/loser.png", + "save_id": 2789, + "prompt_en": "Complete the red path for the word 'BOX' starting from the red cell. Move only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2306, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_with_start/task_007_EYE_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2306/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2306/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2306/loser.png", + "save_id": 2791, + "prompt_en": "Complete the red path for the word 'EYE' starting from the red cell. Move only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2307, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part4/word_with_start/task_009_TEA_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2307/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2307/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2307/loser.png", + "save_id": 2793, + "prompt_en": "Complete the red path for the word 'TEA' starting from the red cell. Move only Down or Right.", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2308, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "EMU3.5" + ], + "source_image_ori": "Part4/word_with_start/task_011_FISH_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2308/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2308/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2308/tie_2.png", + "save_id": 2795, + "prompt_en": "Complete the red path for the word 'FISH' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2309, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_with_start/task_041_MONKEY_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2309/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2309/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2309/tie_2.png", + "save_id": 2825, + "prompt_en": "Complete the red path for the word 'MONKEY' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2310, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/word_with_start/task_043_ORANGE_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2310/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2310/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2310/tie_2.png", + "save_id": 2827, + "prompt_en": "Complete the red path for the word 'ORANGE' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2311, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_with_start/task_047_FRIEND_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2311/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2311/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2311/tie_2.png", + "save_id": 2831, + "prompt_en": "Complete the red path for the word 'FRIEND' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2312, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Bagel" + ], + "source_image_ori": "Part4/word_with_start/task_049_PYTHON_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2312/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2312/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2312/tie_2.png", + "save_id": 2833, + "prompt_en": "Complete the red path for the word 'PYTHON' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2313, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part4/word_with_start/task_051_GARDEN_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2313/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2313/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2313/tie_2.png", + "save_id": 2835, + "prompt_en": "Complete the red path for the word 'GARDEN' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2314, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/word_with_start/task_053_APPLE_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2314/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2314/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2314/tie_2.png", + "save_id": 2837, + "prompt_en": "Complete the red path for the word 'APPLE' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2315, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_with_start/task_071_KANGAROO_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2315/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2315/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2315/tie_2.png", + "save_id": 2855, + "prompt_en": "Complete the red path for the word 'KANGAROO' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2316, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_with_start/task_073_SANDWICH_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2316/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2316/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2316/tie_2.png", + "save_id": 2857, + "prompt_en": "Complete the red path for the word 'SANDWICH' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2317, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/word_with_start/task_081_TELEPHONE_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2317/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2317/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2317/tie_2.png", + "save_id": 2865, + "prompt_en": "Complete the red path for the word 'TELEPHONE' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2318, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part4/word_with_start/task_085_VACATION_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2318/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2318/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2318/tie_2.png", + "save_id": 2869, + "prompt_en": "Complete the red path for the word 'VACATION' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2319, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_with_start/task_089_PINEAPPLE_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2319/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2319/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2319/tie_2.png", + "save_id": 2873, + "prompt_en": "Complete the red path for the word 'PINEAPPLE' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2320, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part4/word_with_start/task_091_CHEMISTRY_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2320/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2320/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2320/tie_2.png", + "save_id": 2875, + "prompt_en": "Complete the red path for the word 'CHEMISTRY' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2321, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_with_start/task_093_LANGUAGE_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2321/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2321/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2321/tie_2.png", + "save_id": 2877, + "prompt_en": "Complete the red path for the word 'LANGUAGE' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2322, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "gemini-3-pro-image-preview" + ], + "source_image_ori": "Part4/word_with_start/task_097_PHYSICS_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2322/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2322/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2322/tie_2.png", + "save_id": 2881, + "prompt_en": "Complete the red path for the word 'PHYSICS' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2323, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "gemini-3.1-flash-image-preview" + ], + "source_image_ori": "Part4/word_with_start/task_099_BREAKFAST_input.jpg", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2323/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2323/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2323/tie_2.png", + "save_id": 2883, + "prompt_en": "Complete the red path for the word 'BREAKFAST' starting from the red cell. Move only Down or Right.", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "WA": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "WA": "WA_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2324, + "Sample_Model": [ + "gemini-3-pro-image-preview", + "Qwen-Image-Edit-2511" + ], + "source_image_ori": "Part2/Complex_paint/CN/1.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2324/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2324/tie_1.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2324/tie_2.png", + "save_id": 428, + "prompt_en": "", + "label": "tie", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + }, + { + "idx": 2325, + "Sample_Model": [ + "gemini-3.1-flash-image-preview", + "LongCat-Image-Edit" + ], + "source_image_ori": "Part2/Complex_paint/CN/2.png", + "source_image": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2325/source.png", + "winner_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2325/winner.png", + "loser_src": "/root/data/bxh/OmniBench_EVAL_RM/RM_Bench/IF/2325/loser.png", + "save_id": 429, + "prompt_en": "", + "label": "preference", + "dimension": "IF", + "system_prompt_name": { + "IF": "Single_image_prompt", + "VC": "Single_image_prompt", + "VQ": "Single_image_prompt" + }, + "parser_name": { + "IF": "IF_Parser", + "VC": "VC_Parser", + "VQ": "VQ_Parser" + } + } +] \ No newline at end of file