javier-ab-bsc commited on
Commit
a3bd347
1 Parent(s): 881cbeb

Add eval section

Browse files
Files changed (1) hide show
  1. README.md +332 -1
README.md CHANGED
@@ -614,7 +614,338 @@ This instruction-tuned variant has been trained with a mixture of 276k English,
614
 
615
  ## Evaluation
616
 
617
- <span style="color:red">TODO</span>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
618
 
619
  ## Ethical Considerations and Limitations
620
 
 
614
 
615
  ## Evaluation
616
 
617
+ ### Gold-standard benchmarks
618
+
619
+ Evaluation is done using the Language Model Evaluation Harness (Gao et al., 2024). We evaluate on a set of tasks taken from [SpanishBench](https://github.com/EleutherAI/lm-evaluation-harness/pull/2157), [CatalanBench](https://github.com/EleutherAI/lm-evaluation-harness/pull/2154), [BasqueBench](https://github.com/EleutherAI/lm-evaluation-harness/pull/2153) and [GalicianBench](https://github.com/EleutherAI/lm-evaluation-harness/pull/2155). These benchmarks include both new and existing tasks and datasets. Given that this is an instructed model, we add LM Evaluation Harness's native feature of `chat-template` to the setup. In the tables below, we include the results in a selection of evaluation datasets that represent model's performance across a variety of tasks within these benchmarks.
620
+
621
+ We only use tasks that are either human generated, human translated, or with a strong human-in-the-loop (i.e., machine translation followed by professional revision or machine generation followed by human revision and annotation). This is the reason behind the variety in number of tasks reported across languages. As more tasks that fulfill these requirements are published, we will update the presented results. We also intend to expand the evaluation to other languages, as long as the datasets meet our quality standards.
622
+
623
+ During the implementation of the evaluation we observed a series of issues worth considering when replicating and interpreting the results presented. These issues include ≈1.5% variances in performance in some tasks depending on the version of the `transformers` library used, and depending on the use (or lack of use) of tensor parallelism when loading a model. When implementing existing tasks, we carry out a comprehensive quality evaluation of the dataset, the Harness task itself, and what kind of input models see during evaluation. Our implementation (see links above) addresses multiple existing problems such as errors in datasets and prompts, and lack of pre-processing. All this means that results will vary if using other Harness implementations, and may slightly vary depending on the replication setup.
624
+
625
+ It should be noted that these results are subject to all the drawbacks of every current gold-standard evaluation, and that the figures do not fully represent the models capabilities and potential. We thus advise caution when reading and interpreting the results.
626
+
627
+ A full list of results compared to other baselines, a discussion of the model's performance across tasks and its implications, and details regarding problem-solving with task implementation will soon be available in the technical report.
628
+
629
+ All results reported below are on a 0-shot setting.
630
+
631
+ #### Spanish
632
+
633
+ <table><thead>
634
+ <tr>
635
+ <th>Category</th>
636
+ <th>Task</th>
637
+ <th>Metric</th>
638
+ <th>Result</th>
639
+ </tr></thead>
640
+ <tbody>
641
+ <tr>
642
+ <td>Commonsense Reasoning</td>
643
+ <td>xstorycloze_es</td>
644
+ <td>acc</td>
645
+ <td>00</td>
646
+ </tr>
647
+ <tr>
648
+ <td>Math</td>
649
+ <td>mgsm_direct_es</td>
650
+ <td>em</td>
651
+ <td>00</td>
652
+ </tr>
653
+ <tr>
654
+ <td rowspan="2">NLI</td>
655
+ <td>wnli_es</td>
656
+ <td>acc</td>
657
+ <td>45.07</td>
658
+ </tr>
659
+ <tr>
660
+ <td>xnli_es</td>
661
+ <td>acc</td>
662
+ <td>51.49</td>
663
+ </tr>
664
+ <tr>
665
+ <td>Paraphrasing</td>
666
+ <td>paws_es</td>
667
+ <td>acc</td>
668
+ <td>59.4</td>
669
+ </tr>
670
+ <tr>
671
+ <td>QA</td>
672
+ <td>xquad_es</td>
673
+ <td>acc</td>
674
+ <td>43.82</td>
675
+ </tr>
676
+ <tr>
677
+ <td>Translation</td>
678
+ <td>flores_es</td>
679
+ <td>bleu</td>
680
+ <td>22.98</td>
681
+ </tr>
682
+ </tbody>
683
+ </table>
684
+
685
+ #### Catalan
686
+
687
+ <table><thead>
688
+ <tr>
689
+ <th>Category</th>
690
+ <th>Task</th>
691
+ <th>Metric</th>
692
+ <th>Result</th>
693
+ </tr></thead>
694
+ <tbody>
695
+ <tr>
696
+ <td rowspan="2">Commonsense Reasoning</td>
697
+ <td>copa_ca</td>
698
+ <td>acc</td>
699
+ <td>81.2</td>
700
+ </tr>
701
+ <tr>
702
+ <td>xstorycloze_ca</td>
703
+ <td>acc</td>
704
+ <td>70.68</td>
705
+ </tr>
706
+ <tr>
707
+ <td>Math</td>
708
+ <td>mgsm_direct_ca</td>
709
+ <td>em</td>
710
+ <td>0</td>
711
+ </tr>
712
+ <tr>
713
+ <td rowspan="2">NLI</td>
714
+ <td>wnli_ca</td>
715
+ <td>acc</td>
716
+ <td>50.7</td>
717
+ </tr>
718
+ <tr>
719
+ <td>xnli_ca</td>
720
+ <td>acc</td>
721
+ <td>55.14</td>
722
+ </tr>
723
+ <tr>
724
+ <td rowspan="2">Paraphrasing</td>
725
+ <td>parafraseja</td>
726
+ <td>acc</td>
727
+ <td>65.18</td>
728
+ </tr>
729
+ <tr>
730
+ <td> paws_ca</td>
731
+ <td>acc</td>
732
+ <td>62.95</td>
733
+ </tr>
734
+ <tr>
735
+ <td rowspan="5">QA</td>
736
+ <td>arc_ca_easy</td>
737
+ <td>acc</td>
738
+ <td>64.98</td>
739
+ </tr>
740
+ <tr>
741
+ <td> arc_ca_challenge</td>
742
+ <td>acc</td>
743
+ <td>41.89</td>
744
+ </tr>
745
+ <tr>
746
+ <td> openbookqa_ca</td>
747
+ <td>acc</td>
748
+ <td>35.2</td>
749
+ </tr>
750
+ <tr>
751
+ <td> piqa_ca</td>
752
+ <td>acc</td>
753
+ <td>69.53</td>
754
+ </tr>
755
+ <tr>
756
+ <td> siqa_ca</td>
757
+ <td>acc</td>
758
+ <td>48.62</td>
759
+ </tr>
760
+ <tr>
761
+ <td>Translation</td>
762
+ <td>flores_ca</td>
763
+ <td>bleu</td>
764
+ <td>28.65</td>
765
+ </tr>
766
+ </tbody></table>
767
+
768
+ #### Basque
769
+
770
+ <table><thead>
771
+ <tr>
772
+ <th>Category</th>
773
+ <th>Task</th>
774
+ <th>Metric</th>
775
+ <th>Result</th>
776
+ </tr></thead>
777
+ <tbody>
778
+ <tr>
779
+ <td rowspan="2">Commonsense Reasoning</td>
780
+ <td>xcopa_eu</td>
781
+ <td>acc</td>
782
+ <td>61.6</td>
783
+ </tr>
784
+ <tr>
785
+ <td>xstorycloze_eu</td>
786
+ <td>acc</td>
787
+ <td>61.15</td>
788
+ </tr>
789
+ <tr>
790
+ <td>Math</td>
791
+ <td>mgsm_direct_eu</td>
792
+ <td>em</td>
793
+ <td>1</td>
794
+ </tr>
795
+ <tr>
796
+ <td rowspan="2">NLI</td>
797
+ <td>wnli_eu</td>
798
+ <td>acc</td>
799
+ <td>45.07</td>
800
+ </tr>
801
+ <tr>
802
+ <td>xnli_eu</td>
803
+ <td>acc</td>
804
+ <td>46.81</td>
805
+ </tr>
806
+ <tr>
807
+ <td rowspan="3">QA</td>
808
+ <td>eus_exams</td>
809
+ <td>acc</td>
810
+ <td>39.09</td>
811
+ </tr>
812
+ <tr>
813
+ <td>eus_proficiency</td>
814
+ <td>acc</td>
815
+ <td>36.93</td>
816
+ </tr>
817
+ <tr>
818
+ <td>eus_trivia</td>
819
+ <td>acc</td>
820
+ <td>46.94</td>
821
+ </tr>
822
+ <tr>
823
+ <td>Reading Comprehension</td>
824
+ <td>eus_reading</td>
825
+ <td>acc</td>
826
+ <td>00</td>
827
+ </tr>
828
+ <tr>
829
+ <td>Translation</td>
830
+ <td>flores_eu</td>
831
+ <td>bleu</td>
832
+ <td>14.89</td>
833
+ </tr>
834
+ </tbody></table>
835
+
836
+ #### Galician
837
+
838
+ <table><thead>
839
+ <tr>
840
+ <th>Category</th>
841
+ <th>Task</th>
842
+ <th>Metric</th>
843
+ <th>Result</th>
844
+ </tr></thead>
845
+ <tbody>
846
+ <tr>
847
+ <td>Math</td>
848
+ <td>mgsm_direct_gl</td>
849
+ <td>em</td>
850
+ <td>0</td>
851
+ </tr>
852
+ <tr>
853
+ <td rowspan="2">Paraphrasing</td>
854
+ <td>parafrases_gl</td>
855
+ <td>acc</td>
856
+ <td>55.44</td>
857
+ </tr>
858
+ <tr>
859
+ <td>paws_gl</td>
860
+ <td>acc</td>
861
+ <td>56.55</td>
862
+ </tr>
863
+ <tr>
864
+ <td>QA</td>
865
+ <td>openbookqa_gl</td>
866
+ <td>acc</td>
867
+ <td>38.4</td>
868
+ </tr>
869
+ <tr>
870
+ <td>Translation</td>
871
+ <td>flores_gl</td>
872
+ <td>bleu</td>
873
+ <td>27.03</td>
874
+ </tr>
875
+ </tbody>
876
+ </table>
877
+
878
+ ### LLM-as-a-judge
879
+
880
+ We use [Prometheus-2 8x7B](https://huggingface.co/prometheus-eval/prometheus-8x7b-v2.0) as a judge to evaluate the responses of the model. Tasks are created from existing multilingual evaluation datasets covering the same categories as the ones measured in our gold-standard benchmarks. We randomly select a subset of 250 instances per language from the `test` set of each source dataset. To evaluate the responses of our model, we use task-specific criteria developed in-house for the _LLM-judge_ to use. Each criterion is measured either as a 5-point Likert scale or as a binary task depending on the idiosyncrasy of the task and criterion.
881
+
882
+ Prompts for each task are created in various ways to score the model's robustness in addition to these criteria. This is done by presenting the same source instance within three different prompts. We then calculate the variance between the scores assigned by the _LLM-judge_ to our model's responses to the three prompt styles and average it across all instances. Prompts are human translated to all languages measured. We do not provide the _LLM-judge_ with a reference answer.
883
+
884
+ The _judge_ prompt we use during evaluation is the same used to fine tune the Prometheus-2 family. We keep the _judge_ prompt and criteria used to present the _LLM-judge_ with the task prompts and model responses in English for evaluation across languages. The _judge_ prompt used is:
885
+
886
+ ```python
887
+ "You are a fair judge assistant tasked with providing clear, objective feedback based on specific criteria, ensuring each assessment reflects the absolute standards set for performance.
888
+
889
+ ###Task Description:
890
+ An instruction (might include an Input inside it), a response to evaluate, and a score rubric representing a evaluation criteria are given.
891
+ 1. Write a detailed feedback that assess the quality of the response strictly based on the given score rubric, not evaluating in general.
892
+ 2. After writing a feedback, write a score that is an integer between {a} and {b}. You should refer to the score rubric.
893
+ 3. The output format should look as follows: \"Feedback: (write a feedback for criteria) [RESULT] (an integer number between {a} and {b})\"
894
+ 4. Please do not generate any other opening, closing, and explanations.
895
+
896
+ ###The instruction to evaluate:
897
+ {input}
898
+
899
+ ###Response to evaluate:
900
+ {prediction}
901
+
902
+ ###Score Rubrics:
903
+ {criteria}
904
+
905
+ ###Feedback:"
906
+ ```
907
+
908
+
909
+ As an example, prompts for the Math task in English are based on instances from [MGSM](https://huggingface.co/datasets/juletxara/mgsm), and each instance is presented within these prompts:
910
+
911
+ ```python
912
+ "en": [
913
+ ("I need help with this math problem: \"", "\" Give me the answer step by step and also the final result separately."),
914
+ ("Can you please help me answer this? \"", "\" Explain the answer and give me the final result as well. Thanks."),
915
+ ("Help me with this problem: \"", "\" I need the answer explained and the final result separately.")
916
+ ]
917
+ ```
918
+
919
+
920
+ This task is then evaluated by the _LLM-judge_ using two criteria, reasoning capability (5-point Likert) and mathematical correctness (binary):
921
+
922
+ ```python
923
+ reasoning_capability_criteria = {
924
+ "reasoning_capability": """
925
+ [Does the model's answer demonstrate reasoning capability?]
926
+ Score 1: The answer demonstrates poor reasoning, with illogical arguments or conclusions that do not follow from the provided information.
927
+ Score 2: The answer shows weak reasoning, with some logical connections but also contains significant flaws or gaps in the argumentation.
928
+ Score 3: The answer demonstrates adequate reasoning, with generally logical arguments, but may have minor flaws or a lack of depth in the reasoning process.
929
+ Score 4: The answer shows strong reasoning, with well-structured arguments and conclusions that logically follow from the information provided.
930
+ Score 5: The answer demonstrates exceptional reasoning, with clear, coherent, and insightful arguments that are logically sound and well-supported by the information provided."""
931
+ }
932
+
933
+ mathematical_correctness_binary_criteria = {
934
+ "mathematical_correctness_binary": """
935
+ [Is the model's answer mathematically correct?]
936
+ Score 0: The answer contains mathematical errors that render the solution incorrect or unreliable.
937
+ Score 1: The answer is mathematically correct, with accurate calculations and appropriate use of mathematical concepts."""
938
+ }
939
+ ```
940
+
941
+
942
+ #### Multilingual results
943
+
944
+ Here, we present results for seven categories of tasks in Spanish, Catalan, Basque, Galician, and English. Results are presented for each task, criterion and language. Criteria with a `(B)` after their name are binary criteria (i.e., numbers go from 0 to 1, where 1 is best). The rest of the criteria are measured using a 5-point Likert scale, where 5 is best. The first number of the pair of numbers separated by `/` shows the average score for the criterion (and language). The second number of each pair is the robustness score, where numbers closer to 0 mean that the model generates similar responses when comparing the three prompt varieties for a single instance.
945
+
946
+ Further details on all tasks and criteria, a full list of results compared to other baselines, a discussion of the model's performance across tasks and its implications, and details regarding problem-solving with task implementation will soon be available in the technical report.
947
+
948
+ ![](./images/results_eval_7b_judge.png)
949
 
950
  ## Ethical Considerations and Limitations
951