Model Open Ended VQA: % Human Rating Multiple Choice VQA: % Accuracy Hints-Multiple Choice VQA: % Accuracy Attributions-Multiple Choice VQA: % Accuracy Refernce Based-Automatic Evaluation: Accuracy of Judge Prediction Compared to Human Ratings Refernce Free-Automatic Evaluation: Accuracy of Judge Prediction Compared to Human Ratings Automatic Evaluation: % Auto-Rater Ratings Hints-Automatic Evaluation: % Auto-Rater Ratings Attributions-Automatic Evaluation: % Auto-Rater Ratings | |
Humans 82 * * * * * 78 * * | |
Gemini Pro 1.5 40 38 66 72 87 52 53 62 29 | |
Gemini Pro Vision 30 41 62 * 75 38 34 47 | |
GPT4 34 45 69 82 86 51 38 61 25 | |
LlaVA-1.6-34B 15 24 30 * 76 43 21 16 * | |
LlaVA-1.5-7B 13 17 29 * 70 35 19 30 * | |
InstructBlip 13 * * * * * 20 28 * | |
Gemini Pro 1.5 Caption _ Gemini Pro 1.5 23 * * * * * * * * | |
Human (Oracle) Caption _ Gemini Pro 1.5 50 * * * * * * * * | |
Claude 3.5 Sonnet * 46 45 * * * 39 * * | |
GPT4o * 55 83 * * * 50 * * | |
Qwen-VL-Max * 35 53 * * * 26 * * | |
Molmo-7B * 34 42 * * * 36 * * |