name,Faithfulness,Answer_Relevancy,Answer_Correctness,Answer_Similarity Yi-6B-Chat,0.7727272727272728,0.39624895698624696,0.7542070590484817,0.8510823274180981 Internlm2-Chat-20B,0.4974937343358396,0.23934623278394898,0.6204328631578558,0.8259980417246064 Yi-6B,0.14285714285714285,0.5190011151439459,0.613405924657078,0.7862767598528019 Qwen1.5-0.5B-Chat,0.42857142857142855,0.488269773660657,0.5753016779392393,0.8382536957664072 Baichuan2-13B-Chat,0.2857142857142857,0.7405192357284925,0.5708584068359933,0.8691557081776934 Internlm2-Chat-7B,0.5,0.22600361749502174,0.570841424928178,0.835409602287096 Gemma-2B,0.011904761904761904,0.2803251363458862,0.557515859867883,0.8041128629884543 Qwen1.5-4B-Chat,0.42857142857142855,0.7503365683315548,0.5549539275574193,0.8483871388011062 Vicuna-13B-V1.5,0.7223381250590647,0.5398926752022207,0.5186514512565293,0.6634719608098028 Yi-9B,0.25,0.31590509156704094,0.4925135259219739,0.7803641425669777 Baichuan2-7B-Chat,0.42857142857142855,0.5020120070047182,0.4880968089375992,0.8529672546477044 Vicuna-7B-V1.5,0.14285714285714285,0.5019194353900974,0.4722927782884535,0.8320120749350508 Gpt-3.5-Turbo,0.6428571428571429,0.7368772770427375,0.46705283670036873,0.8518465563087741 Qwen1.5-1.8B-Chat,0.40816326530612246,0.6187149363151768,0.46468776283363905,0.8536228462063511 Qwen1.5-14B-Chat,0.5,0.62189263577569,0.44546489401960593,0.8573697801600568 Qwen1.5-7B-Chat,0.5714285714285714,0.6304668631060011,0.4380453591507842,0.8549025250385108 Gemma-7B,0.25,0.09449475661345018,0.4189708246631338,0.7939624827076257 Mistral-7B,0.14285714285714285,0.10603847794715389,0.3972735247665132,0.7962369562089103