HallusionBench / gpt4v_benchmark.py
rayguan's picture
Upload 6 files
8f587e4
raw
history blame contribute delete
No virus
12.7 kB
import csv
import json
from tqdm import tqdm
import numpy as np
from prettytable import PrettyTable
import os
from utils import *
import openai
### to evaluate your method, implement and run generate_answer function!
root_dir = "."
llava = False
# llava = True
# load_json = False
load_json = True
input_file_name = "HallusionBench.tsv"
save_json_path_vd = "./hallusion_output_vd.json"
save_json_path_vs = "./hallusion_output_vs.json"
model_output_entry = "gpt4v_output"
model_correctness_entry = "gpt4v_output_gpt_check"
model_correctness_entry_human = "gpt4v_output_human_check"
if llava:
save_json_path_vd = "./hallusion_output_vd_llava.json"
save_json_path_vs = "./hallusion_output_vs_llava.json"
model_output_entry = "llava_1_5_output"
model_correctness_entry = "llava_1_5_output_gpt_check"
model_correctness_entry_human = "llava_1_5_output_human_check"
col_idx = {
'category':0,
'subcategory':1,
'visual_input':2,
'set_id':3,
'figure_id':4,
'sample_note':5,
'question_id':6,
'question':7,
'gt_answer_details':8,
'gt_answer':9,
'gpt4v_output':10,
'gpt4v_output_human_check': 11,
'llava_1_5_output':12,
'llava_1_5_output_human_check': 13,
}
def generate_answer(data, model_output_entry):
## TODO
## implement this section with yout model!
## your_function(img_filename, question) -> "0" (No), "1" (Yes), "2" (Uncertain)
# for r in data:
# r[model_output_entry] = your_function(r["filename"], r["question"])
return data
if __name__ == "__main__":
data_vd = []
data_vs = []
with open(input_file_name) as file:
tsv_file = csv.reader(file, delimiter="\t")
flag = 0
for line in tsv_file:
if line[0] not in ["VD", "VS"]:
continue
data_dict = {}
for k, v in col_idx.items():
data_dict[k] = line[v]
data_dict["filename"] = get_image_file_location(root_dir, data_dict)
if line[0] == "VD":
data_vd.append(data_dict)
else:
data_vs.append(data_dict)
## TODO
data_vd = generate_answer(data_vd, model_output_entry)
data_vs = generate_answer(data_vs, model_output_entry)
## END
data_vd = evaluate_by_chatgpt(data_vd, model_output_entry, model_correctness_entry, load_json=load_json, save_json_path=save_json_path_vd)
data_vd = check_same_by_chatgpt(data_vd, model_output_entry, load_json=load_json, save_json_path=save_json_path_vd)
data_vs = evaluate_by_chatgpt(data_vs, model_output_entry, model_correctness_entry, load_json=load_json, save_json_path=save_json_path_vs)
data_vs = check_same_by_chatgpt(data_vs, model_output_entry, load_json=load_json, save_json_path=save_json_path_vs)
data_vd = assign_correctness(data_vd, correctness_entry=model_correctness_entry_human)
data_vs = assign_correctness(data_vs, correctness_entry=model_correctness_entry_human)
data = data_vd + data_vs
all_data = get_eval_all(data, model_correctness_entry_human)
all_vd = get_eval_all(data_vd, model_correctness_entry_human)
all_vs = get_eval_all(data_vs, model_correctness_entry_human)
human_check_correctness = [i["correct"] for i in data]
print("##### Human Evaluate #####")
# question level
table1 = [["per question", "Total"],
["VD", round(100 * all_vd["correct"]/all_vd["total"], 4)],
["VS", round(100 * all_vs["correct"]/all_vs["total"], 4)],
["Overall", round(100 * all_data["correct"]/all_data["total"], 4)]]
tab1 = PrettyTable(table1[0])
tab1.add_rows(table1[1:])
print(tab1)
q_acc_human = round(100 * all_data["correct"]/all_data["total"], 4)
all_data = get_eval_pair_all(data, model_correctness_entry_human)
easy = get_eval_pair_easy(data)
hard = get_eval_pair_hard(data)
all_vd = get_eval_pair_all(data_vd, model_correctness_entry_human)
easy_vd = get_eval_pair_easy(data_vd)
hard_vd = get_eval_pair_hard(data_vd)
all_vs = get_eval_pair_all(data_vs, model_correctness_entry_human)
easy_vs = get_eval_pair_easy(data_vs)
hard_vs = get_eval_pair_hard(data_vs)
# question pair level
table3 = [["per question pair", "Easy", "Hard", "Total"],
["VD", round(100 * easy_vd["correct"]/easy_vd["total"], 4), round(100 * hard_vd["correct"]/hard_vd["total"], 4), round(100 * all_vd["correct"]/all_vd["total"], 4)],
["VS", round(100 * easy_vs["correct"]/easy_vs["total"], 4), round(100 * hard_vs["correct"]/hard_vs["total"], 4), round(100 * all_vs["correct"]/all_vs["total"], 4)],
["Overall", round(100 * easy["correct"]/easy["total"], 4), round(100 * hard["correct"]/hard["total"], 4), round(100 * all_data["correct"]/all_data["total"], 4)]]
tab3 = PrettyTable(table3[0])
tab3.add_rows(table3[1:])
print(tab3)
fig_all = get_eval_fig(data)
fig_vd = get_eval_fig(data_vd)
fig_vs = get_eval_fig(data_vs)
fig_all_human = fig_all
all_data_human = all_data
# image level
table2 = [["per figure", "Correct", "Inconsistant", "Wrong", "Score"],
["VD", round(100 * fig_vd["correct"]/fig_vd["total"], 4), round(100 * fig_vd["inconsistent"]/fig_vd["total"], 4), round(100 * fig_vd["wrong"]/fig_vd["total"], 4), round(100 * fig_vd["score"], 4)],
["VS", round(100 * fig_vs["correct"]/fig_vs["total"], 4), round(100 * fig_vs["inconsistent"]/fig_vs["total"], 4), round(100 * fig_vs["wrong"]/fig_vs["total"], 4), round(100 * fig_vs["score"], 4)],
["Overall", round(100 * fig_all["correct"]/fig_all["total"], 4), round(100 * fig_all["inconsistent"]/fig_all["total"], 4), round(100 * fig_all["wrong"]/fig_all["total"], 4), round(100 * fig_all["score"], 4)]]
tab2 = PrettyTable(table2[0])
tab2.add_rows(table2[1:])
print(tab2)
pair_acc_human = round(100 * all_data["correct"]/all_data["total"], 4)
figure_acc_human = round(100 * fig_all["correct"]/fig_all["total"], 4)
easy_acc_human = round(100 * easy["correct"]/easy["total"], 4)
hard_acc_human = round(100 * hard["correct"]/hard["total"], 4)
stats_human = yes_ratio_stats(data)
# from IPython import embed;embed()
############################################
print("##### GPT Evaluate #####")
data_vd = assign_correctness(data_vd, correctness_entry=model_correctness_entry)
data_vs = assign_correctness(data_vs, correctness_entry=model_correctness_entry)
data = data_vd + data_vs
all_data = get_eval_all(data, model_correctness_entry)
all_vd = get_eval_all(data_vd, model_correctness_entry)
all_vs = get_eval_all(data_vs, model_correctness_entry)
gpt_check_correctness = [i["correct"] for i in data]
# question level
table1 = [["per question", "Total"],
["VD", round(100 * all_vd["correct"]/all_vd["total"], 4)],
["VS", round(100 * all_vs["correct"]/all_vs["total"], 4)],
["Overall", round(100 * all_data["correct"]/all_data["total"], 4)]]
tab1 = PrettyTable(table1[0])
tab1.add_rows(table1[1:])
print(tab1)
q_acc_gpt = round(100 * all_data["correct"]/all_data["total"], 4)
all_data = get_eval_pair_all(data, model_correctness_entry)
easy = get_eval_pair_easy(data)
hard = get_eval_pair_hard(data)
all_vd = get_eval_pair_all(data_vd, model_correctness_entry)
easy_vd = get_eval_pair_easy(data_vd)
hard_vd = get_eval_pair_hard(data_vd)
all_vs = get_eval_pair_all(data_vs, model_correctness_entry)
easy_vs = get_eval_pair_easy(data_vs)
hard_vs = get_eval_pair_hard(data_vs)
# question pair level
table3 = [["per question pair", "Easy", "Hard", "Total"],
["VD", round(100 * easy_vd["correct"]/easy_vd["total"], 4), round(100 * hard_vd["correct"]/hard_vd["total"], 4), round(100 * all_vd["correct"]/all_vd["total"], 4)],
["VS", round(100 * easy_vs["correct"]/easy_vs["total"], 4), round(100 * hard_vs["correct"]/hard_vs["total"], 4), round(100 * all_vs["correct"]/all_vs["total"], 4)],
["Overall", round(100 * easy["correct"]/easy["total"], 4), round(100 * hard["correct"]/hard["total"], 4), round(100 * all_data["correct"]/all_data["total"], 4)]]
tab3 = PrettyTable(table3[0])
tab3.add_rows(table3[1:])
print(tab3)
fig_all = get_eval_fig(data)
fig_vd = get_eval_fig(data_vd)
fig_vs = get_eval_fig(data_vs)
# image level
table2 = [["per figure", "Correct", "Wrong", "Score"],
["VD", round(100 * fig_vd["correct"]/fig_vd["total"], 4), round(100 * fig_vd["inconsistent"]/fig_vd["total"], 4) + round(100 * fig_vd["wrong"]/fig_vd["total"], 4), round(fig_vd["score"], 4)],
["VS", round(100 * fig_vs["correct"]/fig_vs["total"], 4), round(100 * fig_vs["inconsistent"]/fig_vs["total"], 4) + round(100 * fig_vs["wrong"]/fig_vs["total"], 4), round(fig_vs["score"], 4)],
["Overall", round(100 * fig_all["correct"]/fig_all["total"], 4), round(100 * fig_all["inconsistent"]/fig_all["total"], 4) + round(100 * fig_all["wrong"]/fig_all["total"], 4), round(fig_all["score"], 4)]]
tab2 = PrettyTable(table2[0])
tab2.add_rows(table2[1:])
print(tab2)
pair_acc_gpt = round(100 * all_data["correct"]/all_data["total"], 4)
figure_acc_gpt = round(100 * fig_all["correct"]/fig_all["total"], 4)
easy_acc_gpt = round(100 * easy["correct"]/easy["total"], 4)
hard_acc_gpt = round(100 * hard["correct"]/hard["total"], 4)
##############################
print("##### Question Stats #####")
print("Easy Questions: " + str(easy_vd["total_q"]) + "(Visual Dependent) + " + str(easy_vs["total_q"]) + "(Visual Supplement)")
print("Hard Questions: " + str(hard_vd["total_q"]) + "(Visual Dependent) + " + str(hard_vs["total_q"]) + "(Visual Supplement)")
print("Total Questions: " + str(all_data["total_q"]))
print("##### Figure Stats #####")
print("Visual Dependent Figures: " + str(fig_vd["total"]))
print("Visual Supplement Figures: " + str(fig_vs["total"]))
print("Total Figures: " + str(fig_all["total"]))
print("##### Leaderboard Stats #####")
table = [["", "Acc per question pair (qAcc)", "Acc per figure (fAcc)", "Acc per easy question (easy aAcc)", "Acc per hard question (hard aAcc)", "Acc per question (aAcc)"],
["Human Eval", pair_acc_human, figure_acc_human, easy_acc_human, hard_acc_human, q_acc_human],
["GPT Eval", pair_acc_gpt, figure_acc_gpt, easy_acc_gpt, hard_acc_gpt, q_acc_gpt]]
leaderboard = PrettyTable(table[0])
leaderboard.add_rows(table[1:])
print(leaderboard)
print(all_data["total"], all_data["wrong"], all_data["LH"], all_data["VI"], all_data["Mix"])
print(all_data["total_q"], all_data["LH_cg"], all_data["VI_cg"], all_data["Mix_cg"])
print(len(gpt_check_correctness))
print(len(human_check_correctness))
print(sum(np.array(human_check_correctness) == np.array(gpt_check_correctness)))
print(sum(np.array(human_check_correctness) == np.array(gpt_check_correctness)) / len(gpt_check_correctness))
yes = [int(i["gt_answer"]) for i in data]
print(sum(yes))
print(len(yes))
print(sum(yes)/len(yes))
stats_gpt = yes_ratio_stats(data)
table = [["", "Yes/No Bias (Pct Diff)", "Yes/No Bias (FP Ratio)", "Consistency Test (correct)", "Consistency Test (inconsistent)", "Consistency Test (wrong)", "LH", "VI", "Mixed"],
["Human Eval", stats_human["diff"], stats_human["fp"], round(100 * fig_all_human["correct"]/fig_all_human["total"], 4), round(100 * fig_all_human["inconsistent"]/fig_all_human["total"], 4), round(100 * fig_all_human["wrong"]/fig_all_human["total"], 4), round(100 * all_data_human["LH_cg"]/(all_data_human["LH_cg"] + all_data_human["VI_cg"] + all_data_human["Mix_cg"]), 4), round(100 * all_data_human["VI_cg"]/(all_data_human["LH_cg"] + all_data_human["VI_cg"] + all_data_human["Mix_cg"]), 4), round(100 * all_data_human["Mix_cg"]/(all_data_human["LH_cg"] + all_data_human["VI_cg"] + all_data_human["Mix_cg"]), 4)],
["GPT Eval", stats_gpt["diff"], stats_gpt["fp"], round(100 * fig_all["correct"]/fig_all["total"], 4), round(100 * fig_all["inconsistent"]/fig_all["total"], 4), round(100 * fig_all["wrong"]/fig_all["total"], 4), round(100 * all_data["LH_cg"]/(all_data["LH_cg"] + all_data["VI_cg"] + all_data["Mix_cg"]), 4), round(100 * all_data["VI_cg"]/(all_data["LH_cg"] + all_data["VI_cg"] + all_data["Mix_cg"]), 4), round(100 * all_data["Mix_cg"]/(all_data["LH_cg"] + all_data["VI_cg"] + all_data["Mix_cg"]), 4)]]
test = PrettyTable(table[0])
test.add_rows(table[1:])
print(test)