eduagarcia commited on
Commit
2083a24
1 Parent(s): 9ff0e9d

Uploading raw results for bigscience/bloom-1b7

Browse files
Files changed (13) hide show
  1. .gitattributes +5 -0
  2. bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_rte.jsonl +3 -0
  3. bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_sts.jsonl +3 -0
  4. bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl +3 -0
  5. bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl +3 -0
  6. bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_faquad_nli.jsonl +0 -0
  7. bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl +3 -0
  8. bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_emotion-2021-cortiz-por.jsonl +0 -0
  9. bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_hate-2019-fortuna-por.jsonl +0 -0
  10. bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_sentiment-2016-mozetic-por.jsonl +0 -0
  11. bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_sentiment-2018-brum-por.jsonl +0 -0
  12. bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/results.json +1259 -0
  13. bigscience/bloom-1b7/results_2024-02-07T01-17-14.803861.json +346 -0
.gitattributes CHANGED
@@ -99,3 +99,8 @@ bigscience/bloom-560m/raw_2024-02-07T01-17-10.201449/pretrained__bigscience__blo
99
  bigscience/bloom-560m/raw_2024-02-07T01-17-10.201449/pretrained__bigscience__bloom-560m,dtype__float16,device__cuda:2,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl filter=lfs diff=lfs merge=lfs -text
100
  bigscience/bloom-560m/raw_2024-02-07T01-17-10.201449/pretrained__bigscience__bloom-560m,dtype__float16,device__cuda:2,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl filter=lfs diff=lfs merge=lfs -text
101
  bigscience/bloom-560m/raw_2024-02-07T01-17-10.201449/pretrained__bigscience__bloom-560m,dtype__float16,device__cuda:2,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
99
  bigscience/bloom-560m/raw_2024-02-07T01-17-10.201449/pretrained__bigscience__bloom-560m,dtype__float16,device__cuda:2,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl filter=lfs diff=lfs merge=lfs -text
100
  bigscience/bloom-560m/raw_2024-02-07T01-17-10.201449/pretrained__bigscience__bloom-560m,dtype__float16,device__cuda:2,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl filter=lfs diff=lfs merge=lfs -text
101
  bigscience/bloom-560m/raw_2024-02-07T01-17-10.201449/pretrained__bigscience__bloom-560m,dtype__float16,device__cuda:2,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl filter=lfs diff=lfs merge=lfs -text
102
+ bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_rte.jsonl filter=lfs diff=lfs merge=lfs -text
103
+ bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_sts.jsonl filter=lfs diff=lfs merge=lfs -text
104
+ bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl filter=lfs diff=lfs merge=lfs -text
105
+ bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl filter=lfs diff=lfs merge=lfs -text
106
+ bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl filter=lfs diff=lfs merge=lfs -text
bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_rte.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfafb6664452ba32ebe27565a605b01aa43b2509990a5d0e01f4ec7b06b24e78
3
+ size 29334761
bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_sts.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:785a446f7beef6f7841397a1afd489725efc38103afc168520edfd655106b079
3
+ size 30821901
bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:436c39635d0ac61d5444007db855f1d07bf5eb9243a742e276d1749720047a94
3
+ size 10572897
bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c699022f324f89a9ab3d0ae515ecfa6350e2637aa41d7406ec6e66fbfa9632b
3
+ size 23283518
bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_faquad_nli.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7ed24565932d245ffe9313073b20394c147090a3614c4a9f2d25fe33d1cf61f
3
+ size 32536881
bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_emotion-2021-cortiz-por.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_hate-2019-fortuna-por.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_sentiment-2016-mozetic-por.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/pretrained__bigscience__bloom-1b7,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_sentiment-2018-brum-por.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
bigscience/bloom-1b7/raw_2024-02-07T01-17-14.803861/results.json ADDED
@@ -0,0 +1,1259 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "assin2_rte": {
4
+ "f1_macro,all": 0.3378561857199312,
5
+ "acc,all": 0.5020424836601307,
6
+ "alias": "assin2_rte"
7
+ },
8
+ "assin2_sts": {
9
+ "pearson,all": 0.017927392336598496,
10
+ "mse,all": 2.7831372549019617,
11
+ "alias": "assin2_sts"
12
+ },
13
+ "bluex": {
14
+ "acc,all": 0.23087621696801114,
15
+ "acc,exam_id__USP_2019": 0.175,
16
+ "acc,exam_id__UNICAMP_2018": 0.24074074074074073,
17
+ "acc,exam_id__UNICAMP_2020": 0.23636363636363636,
18
+ "acc,exam_id__UNICAMP_2023": 0.23255813953488372,
19
+ "acc,exam_id__UNICAMP_2024": 0.2,
20
+ "acc,exam_id__UNICAMP_2019": 0.24,
21
+ "acc,exam_id__UNICAMP_2021_2": 0.19607843137254902,
22
+ "acc,exam_id__USP_2022": 0.30612244897959184,
23
+ "acc,exam_id__USP_2020": 0.30357142857142855,
24
+ "acc,exam_id__USP_2018": 0.2037037037037037,
25
+ "acc,exam_id__USP_2021": 0.19230769230769232,
26
+ "acc,exam_id__USP_2023": 0.25,
27
+ "acc,exam_id__USP_2024": 0.17073170731707318,
28
+ "acc,exam_id__UNICAMP_2021_1": 0.2608695652173913,
29
+ "acc,exam_id__UNICAMP_2022": 0.23076923076923078,
30
+ "alias": "bluex"
31
+ },
32
+ "enem_challenge": {
33
+ "alias": "enem",
34
+ "acc,all": 0.1966410076976907,
35
+ "acc,exam_id__2016_2": 0.17073170731707318,
36
+ "acc,exam_id__2016": 0.24793388429752067,
37
+ "acc,exam_id__2010": 0.17094017094017094,
38
+ "acc,exam_id__2014": 0.1743119266055046,
39
+ "acc,exam_id__2023": 0.2222222222222222,
40
+ "acc,exam_id__2017": 0.1724137931034483,
41
+ "acc,exam_id__2015": 0.21008403361344538,
42
+ "acc,exam_id__2012": 0.1206896551724138,
43
+ "acc,exam_id__2022": 0.21804511278195488,
44
+ "acc,exam_id__2013": 0.16666666666666666,
45
+ "acc,exam_id__2009": 0.19130434782608696,
46
+ "acc,exam_id__2011": 0.28205128205128205
47
+ },
48
+ "faquad_nli": {
49
+ "f1_macro,all": 0.4396551724137931,
50
+ "acc,all": 0.7846153846153846,
51
+ "alias": "faquad_nli"
52
+ },
53
+ "oab_exams": {
54
+ "acc,all": 0.2428246013667426,
55
+ "acc,exam_id__2014-13": 0.2875,
56
+ "acc,exam_id__2012-07": 0.175,
57
+ "acc,exam_id__2012-09": 0.24675324675324675,
58
+ "acc,exam_id__2011-03": 0.2222222222222222,
59
+ "acc,exam_id__2012-08": 0.275,
60
+ "acc,exam_id__2015-17": 0.20512820512820512,
61
+ "acc,exam_id__2014-15": 0.2564102564102564,
62
+ "acc,exam_id__2017-24": 0.175,
63
+ "acc,exam_id__2015-16": 0.2375,
64
+ "acc,exam_id__2017-23": 0.25,
65
+ "acc,exam_id__2011-04": 0.25,
66
+ "acc,exam_id__2010-02": 0.23,
67
+ "acc,exam_id__2016-19": 0.2564102564102564,
68
+ "acc,exam_id__2012-06": 0.25,
69
+ "acc,exam_id__2012-06a": 0.225,
70
+ "acc,exam_id__2013-12": 0.1875,
71
+ "acc,exam_id__2017-22": 0.3125,
72
+ "acc,exam_id__2010-01": 0.27058823529411763,
73
+ "acc,exam_id__2013-10": 0.225,
74
+ "acc,exam_id__2014-14": 0.25,
75
+ "acc,exam_id__2018-25": 0.2875,
76
+ "acc,exam_id__2016-20a": 0.2875,
77
+ "acc,exam_id__2015-18": 0.2875,
78
+ "acc,exam_id__2011-05": 0.2,
79
+ "acc,exam_id__2013-11": 0.175,
80
+ "acc,exam_id__2016-20": 0.275,
81
+ "acc,exam_id__2016-21": 0.2625,
82
+ "alias": "oab_exams"
83
+ },
84
+ "sparrow_emotion-2021-cortiz-por": {
85
+ "alias": "emotion-2021-cortiz-por",
86
+ "f1_macro,all": 0.048302336566881446,
87
+ "acc,all": 0.088
88
+ },
89
+ "sparrow_hate-2019-fortuna-por": {
90
+ "alias": "hate-2019-fortuna-por",
91
+ "f1_macro,all": 0.3932038834951456,
92
+ "acc,all": 0.648
93
+ },
94
+ "sparrow_sentiment-2016-mozetic-por": {
95
+ "alias": "sentiment-2016-mozetic-por",
96
+ "f1_macro,all": 0.09875428777757718,
97
+ "acc,all": 0.152
98
+ },
99
+ "sparrow_sentiment-2018-brum-por": {
100
+ "alias": "sentiment-2018-brum-por",
101
+ "f1_macro,all": 0.19385342789598106,
102
+ "acc,all": 0.41
103
+ }
104
+ },
105
+ "configs": {
106
+ "assin2_rte": {
107
+ "task": "assin2_rte",
108
+ "group": [
109
+ "pt_benchmark",
110
+ "assin2"
111
+ ],
112
+ "dataset_path": "assin2",
113
+ "test_split": "test",
114
+ "fewshot_split": "train",
115
+ "doc_to_text": "Premissa: {{premise}}\nHipótese: {{hypothesis}}\nPergunta: A hipótese pode ser inferida pela premissa?\nResposta:",
116
+ "doc_to_target": "{{['Não', 'Sim'][entailment_judgment]}}",
117
+ "description": "Abaixo contém pares de premissa e hipótese, para cada par você deve julgar se a hipótese pode ser inferida a partir da premissa, responda apenas com Sim ou Não.\n\n",
118
+ "target_delimiter": " ",
119
+ "fewshot_delimiter": "\n\n",
120
+ "fewshot_config": {
121
+ "sampler": "id_sampler",
122
+ "sampler_config": {
123
+ "id_list": [
124
+ 1,
125
+ 3251,
126
+ 2,
127
+ 3252,
128
+ 3,
129
+ 4,
130
+ 5,
131
+ 6,
132
+ 3253,
133
+ 7,
134
+ 3254,
135
+ 3255,
136
+ 3256,
137
+ 8,
138
+ 9,
139
+ 10,
140
+ 3257,
141
+ 11,
142
+ 3258,
143
+ 12,
144
+ 13,
145
+ 14,
146
+ 15,
147
+ 3259,
148
+ 3260,
149
+ 3261,
150
+ 3262,
151
+ 3263,
152
+ 16,
153
+ 17,
154
+ 3264,
155
+ 18,
156
+ 3265,
157
+ 3266,
158
+ 3267,
159
+ 19,
160
+ 20,
161
+ 3268,
162
+ 3269,
163
+ 21,
164
+ 3270,
165
+ 3271,
166
+ 22,
167
+ 3272,
168
+ 3273,
169
+ 23,
170
+ 3274,
171
+ 24,
172
+ 25,
173
+ 3275
174
+ ],
175
+ "id_column": "sentence_pair_id"
176
+ }
177
+ },
178
+ "num_fewshot": 15,
179
+ "metric_list": [
180
+ {
181
+ "metric": "f1_macro",
182
+ "aggregation": "f1_macro",
183
+ "higher_is_better": true
184
+ },
185
+ {
186
+ "metric": "acc",
187
+ "aggregation": "acc",
188
+ "higher_is_better": true
189
+ }
190
+ ],
191
+ "output_type": "generate_until",
192
+ "generation_kwargs": {
193
+ "max_gen_toks": 32,
194
+ "do_sample": false,
195
+ "temperature": 0.0,
196
+ "top_k": null,
197
+ "top_p": null,
198
+ "until": [
199
+ "\n\n"
200
+ ]
201
+ },
202
+ "repeats": 1,
203
+ "filter_list": [
204
+ {
205
+ "name": "all",
206
+ "filter": [
207
+ {
208
+ "function": "find_similar_label",
209
+ "labels": [
210
+ "Sim",
211
+ "Não"
212
+ ]
213
+ },
214
+ {
215
+ "function": "take_first"
216
+ }
217
+ ]
218
+ }
219
+ ],
220
+ "should_decontaminate": false,
221
+ "metadata": {
222
+ "version": 1.0
223
+ }
224
+ },
225
+ "assin2_sts": {
226
+ "task": "assin2_sts",
227
+ "group": [
228
+ "pt_benchmark",
229
+ "assin2"
230
+ ],
231
+ "dataset_path": "assin2",
232
+ "test_split": "test",
233
+ "fewshot_split": "train",
234
+ "doc_to_text": "Frase 1: {{premise}}\nFrase 2: {{hypothesis}}\nPergunta: Qual o grau de similaridade entre as duas frases de 1,0 a 5,0?\nResposta:",
235
+ "doc_to_target": "<function assin2_float_to_pt_str at 0x7f8bf0f71800>",
236
+ "description": "Abaixo contém pares de frases, para cada par você deve julgar o grau de similaridade de 1,0 a 5,0, responda apenas com o número.\n\n",
237
+ "target_delimiter": " ",
238
+ "fewshot_delimiter": "\n\n",
239
+ "fewshot_config": {
240
+ "sampler": "id_sampler",
241
+ "sampler_config": {
242
+ "id_list": [
243
+ 1,
244
+ 3251,
245
+ 2,
246
+ 3252,
247
+ 3,
248
+ 4,
249
+ 5,
250
+ 6,
251
+ 3253,
252
+ 7,
253
+ 3254,
254
+ 3255,
255
+ 3256,
256
+ 8,
257
+ 9,
258
+ 10,
259
+ 3257,
260
+ 11,
261
+ 3258,
262
+ 12,
263
+ 13,
264
+ 14,
265
+ 15,
266
+ 3259,
267
+ 3260,
268
+ 3261,
269
+ 3262,
270
+ 3263,
271
+ 16,
272
+ 17,
273
+ 3264,
274
+ 18,
275
+ 3265,
276
+ 3266,
277
+ 3267,
278
+ 19,
279
+ 20,
280
+ 3268,
281
+ 3269,
282
+ 21,
283
+ 3270,
284
+ 3271,
285
+ 22,
286
+ 3272,
287
+ 3273,
288
+ 23,
289
+ 3274,
290
+ 24,
291
+ 25,
292
+ 3275
293
+ ],
294
+ "id_column": "sentence_pair_id"
295
+ }
296
+ },
297
+ "num_fewshot": 15,
298
+ "metric_list": [
299
+ {
300
+ "metric": "pearson",
301
+ "aggregation": "pearsonr",
302
+ "higher_is_better": true
303
+ },
304
+ {
305
+ "metric": "mse",
306
+ "aggregation": "mean_squared_error",
307
+ "higher_is_better": false
308
+ }
309
+ ],
310
+ "output_type": "generate_until",
311
+ "generation_kwargs": {
312
+ "max_gen_toks": 32,
313
+ "do_sample": false,
314
+ "temperature": 0.0,
315
+ "top_k": null,
316
+ "top_p": null,
317
+ "until": [
318
+ "\n\n"
319
+ ]
320
+ },
321
+ "repeats": 1,
322
+ "filter_list": [
323
+ {
324
+ "name": "all",
325
+ "filter": [
326
+ {
327
+ "function": "number_filter",
328
+ "type": "float",
329
+ "range_min": 1.0,
330
+ "range_max": 5.0,
331
+ "on_outside_range": "clip",
332
+ "fallback": 5.0
333
+ },
334
+ {
335
+ "function": "take_first"
336
+ }
337
+ ]
338
+ }
339
+ ],
340
+ "should_decontaminate": false,
341
+ "metadata": {
342
+ "version": 1.0
343
+ }
344
+ },
345
+ "bluex": {
346
+ "task": "bluex",
347
+ "group": [
348
+ "pt_benchmark",
349
+ "vestibular"
350
+ ],
351
+ "dataset_path": "eduagarcia-temp/BLUEX_without_images",
352
+ "test_split": "train",
353
+ "fewshot_split": "train",
354
+ "doc_to_text": "<function enem_doc_to_text at 0x7f8bf0f711c0>",
355
+ "doc_to_target": "{{answerKey}}",
356
+ "description": "As perguntas a seguir são questões de multipla escolha de provas de vestibular de Universidades Brasileiras, reponda apenas com as letras A, B, C, D ou E.\n\n",
357
+ "target_delimiter": " ",
358
+ "fewshot_delimiter": "\n\n",
359
+ "fewshot_config": {
360
+ "sampler": "id_sampler",
361
+ "sampler_config": {
362
+ "id_list": [
363
+ "USP_2018_3",
364
+ "UNICAMP_2018_2",
365
+ "USP_2018_35",
366
+ "UNICAMP_2018_16",
367
+ "USP_2018_89"
368
+ ],
369
+ "id_column": "id",
370
+ "exclude_from_task": true
371
+ }
372
+ },
373
+ "num_fewshot": 3,
374
+ "metric_list": [
375
+ {
376
+ "metric": "acc",
377
+ "aggregation": "acc",
378
+ "higher_is_better": true
379
+ }
380
+ ],
381
+ "output_type": "generate_until",
382
+ "generation_kwargs": {
383
+ "max_gen_toks": 32,
384
+ "do_sample": false,
385
+ "temperature": 0.0,
386
+ "top_k": null,
387
+ "top_p": null,
388
+ "until": [
389
+ "\n\n"
390
+ ]
391
+ },
392
+ "repeats": 1,
393
+ "filter_list": [
394
+ {
395
+ "name": "all",
396
+ "filter": [
397
+ {
398
+ "function": "normalize_spaces"
399
+ },
400
+ {
401
+ "function": "remove_accents"
402
+ },
403
+ {
404
+ "function": "find_choices",
405
+ "choices": [
406
+ "A",
407
+ "B",
408
+ "C",
409
+ "D",
410
+ "E"
411
+ ],
412
+ "regex_patterns": [
413
+ "(?:[Ll]etra|[Aa]lternativa|[Rr]esposta|[Rr]esposta [Cc]orreta|[Rr]esposta[Cc]orreta e|[Oo]pcao):? ([ABCDE])\\b",
414
+ "\\b([ABCDE])\\.",
415
+ "\\b([ABCDE]) ?[.):-]",
416
+ "\\b([ABCDE])$",
417
+ "\\b([ABCDE])\\b"
418
+ ]
419
+ },
420
+ {
421
+ "function": "take_first"
422
+ }
423
+ ],
424
+ "group_by": {
425
+ "column": "exam_id"
426
+ }
427
+ }
428
+ ],
429
+ "should_decontaminate": true,
430
+ "doc_to_decontamination_query": "<function enem_doc_to_text at 0x7f8bf0f71440>",
431
+ "metadata": {
432
+ "version": 1.0
433
+ }
434
+ },
435
+ "enem_challenge": {
436
+ "task": "enem_challenge",
437
+ "task_alias": "enem",
438
+ "group": [
439
+ "pt_benchmark",
440
+ "vestibular"
441
+ ],
442
+ "dataset_path": "eduagarcia/enem_challenge",
443
+ "test_split": "train",
444
+ "fewshot_split": "train",
445
+ "doc_to_text": "<function enem_doc_to_text at 0x7f8bf0f719e0>",
446
+ "doc_to_target": "{{answerKey}}",
447
+ "description": "As perguntas a seguir são questões de multipla escolha do Exame Nacional do Ensino Médio (ENEM), reponda apenas com as letras A, B, C, D ou E.\n\n",
448
+ "target_delimiter": " ",
449
+ "fewshot_delimiter": "\n\n",
450
+ "fewshot_config": {
451
+ "sampler": "id_sampler",
452
+ "sampler_config": {
453
+ "id_list": [
454
+ "2022_21",
455
+ "2022_88",
456
+ "2022_143"
457
+ ],
458
+ "id_column": "id",
459
+ "exclude_from_task": true
460
+ }
461
+ },
462
+ "num_fewshot": 3,
463
+ "metric_list": [
464
+ {
465
+ "metric": "acc",
466
+ "aggregation": "acc",
467
+ "higher_is_better": true
468
+ }
469
+ ],
470
+ "output_type": "generate_until",
471
+ "generation_kwargs": {
472
+ "max_gen_toks": 32,
473
+ "do_sample": false,
474
+ "temperature": 0.0,
475
+ "top_k": null,
476
+ "top_p": null,
477
+ "until": [
478
+ "\n\n"
479
+ ]
480
+ },
481
+ "repeats": 1,
482
+ "filter_list": [
483
+ {
484
+ "name": "all",
485
+ "filter": [
486
+ {
487
+ "function": "normalize_spaces"
488
+ },
489
+ {
490
+ "function": "remove_accents"
491
+ },
492
+ {
493
+ "function": "find_choices",
494
+ "choices": [
495
+ "A",
496
+ "B",
497
+ "C",
498
+ "D",
499
+ "E"
500
+ ],
501
+ "regex_patterns": [
502
+ "(?:[Ll]etra|[Aa]lternativa|[Rr]esposta|[Rr]esposta [Cc]orreta|[Rr]esposta[Cc]orreta e|[Oo]pcao):? ([ABCDE])\\b",
503
+ "\\b([ABCDE])\\.",
504
+ "\\b([ABCDE]) ?[.):-]",
505
+ "\\b([ABCDE])$",
506
+ "\\b([ABCDE])\\b"
507
+ ]
508
+ },
509
+ {
510
+ "function": "take_first"
511
+ }
512
+ ],
513
+ "group_by": {
514
+ "column": "exam_id"
515
+ }
516
+ }
517
+ ],
518
+ "should_decontaminate": true,
519
+ "doc_to_decontamination_query": "<function enem_doc_to_text at 0x7f8bf0f71c60>",
520
+ "metadata": {
521
+ "version": 1.0
522
+ }
523
+ },
524
+ "faquad_nli": {
525
+ "task": "faquad_nli",
526
+ "group": [
527
+ "pt_benchmark"
528
+ ],
529
+ "dataset_path": "ruanchaves/faquad-nli",
530
+ "test_split": "test",
531
+ "fewshot_split": "train",
532
+ "doc_to_text": "Pergunta: {{question}}\nResposta: {{answer}}\nA resposta satisfaz a pergunta? Sim ou Não?",
533
+ "doc_to_target": "{{['Não', 'Sim'][label]}}",
534
+ "description": "Abaixo contém pares de pergunta e reposta, para cada par você deve julgar resposta responde a pergunta de maneira satisfatória e aparenta estar correta, escreva apenas Sim ou Não.\n\n",
535
+ "target_delimiter": " ",
536
+ "fewshot_delimiter": "\n\n",
537
+ "fewshot_config": {
538
+ "sampler": "first_n",
539
+ "sampler_config": {
540
+ "fewshot_indices": [
541
+ 1893,
542
+ 949,
543
+ 663,
544
+ 105,
545
+ 1169,
546
+ 2910,
547
+ 2227,
548
+ 2813,
549
+ 974,
550
+ 558,
551
+ 1503,
552
+ 1958,
553
+ 2918,
554
+ 601,
555
+ 1560,
556
+ 984,
557
+ 2388,
558
+ 995,
559
+ 2233,
560
+ 1982,
561
+ 165,
562
+ 2788,
563
+ 1312,
564
+ 2285,
565
+ 522,
566
+ 1113,
567
+ 1670,
568
+ 323,
569
+ 236,
570
+ 1263,
571
+ 1562,
572
+ 2519,
573
+ 1049,
574
+ 432,
575
+ 1167,
576
+ 1394,
577
+ 2022,
578
+ 2551,
579
+ 2194,
580
+ 2187,
581
+ 2282,
582
+ 2816,
583
+ 108,
584
+ 301,
585
+ 1185,
586
+ 1315,
587
+ 1420,
588
+ 2436,
589
+ 2322,
590
+ 766
591
+ ]
592
+ }
593
+ },
594
+ "num_fewshot": 15,
595
+ "metric_list": [
596
+ {
597
+ "metric": "f1_macro",
598
+ "aggregation": "f1_macro",
599
+ "higher_is_better": true
600
+ },
601
+ {
602
+ "metric": "acc",
603
+ "aggregation": "acc",
604
+ "higher_is_better": true
605
+ }
606
+ ],
607
+ "output_type": "generate_until",
608
+ "generation_kwargs": {
609
+ "max_gen_toks": 32,
610
+ "do_sample": false,
611
+ "temperature": 0.0,
612
+ "top_k": null,
613
+ "top_p": null,
614
+ "until": [
615
+ "\n\n"
616
+ ]
617
+ },
618
+ "repeats": 1,
619
+ "filter_list": [
620
+ {
621
+ "name": "all",
622
+ "filter": [
623
+ {
624
+ "function": "find_similar_label",
625
+ "labels": [
626
+ "Sim",
627
+ "Não"
628
+ ]
629
+ },
630
+ {
631
+ "function": "take_first"
632
+ }
633
+ ]
634
+ }
635
+ ],
636
+ "should_decontaminate": false,
637
+ "metadata": {
638
+ "version": 1.0
639
+ }
640
+ },
641
+ "oab_exams": {
642
+ "task": "oab_exams",
643
+ "group": [
644
+ "legal_benchmark",
645
+ "pt_benchmark"
646
+ ],
647
+ "dataset_path": "eduagarcia/oab_exams",
648
+ "test_split": "train",
649
+ "fewshot_split": "train",
650
+ "doc_to_text": "<function doc_to_text at 0x7f8bf0f70b80>",
651
+ "doc_to_target": "{{answerKey}}",
652
+ "description": "As perguntas a seguir são questões de multipla escolha do Exame de Ordem da Ordem dos Advogados do Brasil (OAB), reponda apenas com as letras A, B, C ou D.\n\n",
653
+ "target_delimiter": " ",
654
+ "fewshot_delimiter": "\n\n",
655
+ "fewshot_config": {
656
+ "sampler": "id_sampler",
657
+ "sampler_config": {
658
+ "id_list": [
659
+ "2010-01_1",
660
+ "2010-01_11",
661
+ "2010-01_13",
662
+ "2010-01_23",
663
+ "2010-01_26",
664
+ "2010-01_28",
665
+ "2010-01_38",
666
+ "2010-01_48",
667
+ "2010-01_58",
668
+ "2010-01_68",
669
+ "2010-01_76",
670
+ "2010-01_83",
671
+ "2010-01_85",
672
+ "2010-01_91",
673
+ "2010-01_99"
674
+ ],
675
+ "id_column": "id",
676
+ "exclude_from_task": true
677
+ }
678
+ },
679
+ "num_fewshot": 3,
680
+ "metric_list": [
681
+ {
682
+ "metric": "acc",
683
+ "aggregation": "acc",
684
+ "higher_is_better": true
685
+ }
686
+ ],
687
+ "output_type": "generate_until",
688
+ "generation_kwargs": {
689
+ "max_gen_toks": 32,
690
+ "do_sample": false,
691
+ "temperature": 0.0,
692
+ "top_k": null,
693
+ "top_p": null,
694
+ "until": [
695
+ "\n\n"
696
+ ]
697
+ },
698
+ "repeats": 1,
699
+ "filter_list": [
700
+ {
701
+ "name": "all",
702
+ "filter": [
703
+ {
704
+ "function": "normalize_spaces"
705
+ },
706
+ {
707
+ "function": "remove_accents"
708
+ },
709
+ {
710
+ "function": "find_choices",
711
+ "choices": [
712
+ "A",
713
+ "B",
714
+ "C",
715
+ "D"
716
+ ],
717
+ "regex_patterns": [
718
+ "(?:[Ll]etra|[Aa]lternativa|[Rr]esposta|[Rr]esposta [Cc]orreta|[Rr]esposta[Cc]orreta e|[Oo]pcao):? ([ABCD])\\b",
719
+ "\\b([ABCD])\\)",
720
+ "\\b([ABCD]) ?[.):-]",
721
+ "\\b([ABCD])$",
722
+ "\\b([ABCD])\\b"
723
+ ]
724
+ },
725
+ {
726
+ "function": "take_first"
727
+ }
728
+ ],
729
+ "group_by": {
730
+ "column": "exam_id"
731
+ }
732
+ }
733
+ ],
734
+ "should_decontaminate": true,
735
+ "doc_to_decontamination_query": "<function doc_to_text at 0x7f8bf0f70e00>",
736
+ "metadata": {
737
+ "version": 1.4
738
+ }
739
+ },
740
+ "sparrow_emotion-2021-cortiz-por": {
741
+ "task": "sparrow_emotion-2021-cortiz-por",
742
+ "task_alias": "emotion-2021-cortiz-por",
743
+ "group": [
744
+ "pt_benchmark",
745
+ "sparrow"
746
+ ],
747
+ "dataset_path": "UBC-NLP/sparrow",
748
+ "dataset_name": "emotion-2021-cortiz-por",
749
+ "test_split": "validation",
750
+ "fewshot_split": "train",
751
+ "doc_to_text": "Texto: {{content}}\nPergunta: Qual a principal emoção apresentada no texto?\nResposta:",
752
+ "doc_to_target": "<function sparrow_emotion_por_trans_label at 0x7f8bf0f71080>",
753
+ "description": "Abaixo contém o conteúdo de tweets de usuarios do Twitter em português, sua tarefa é extrair qual a principal emoção dos textos. Responda com apenas uma das seguintes opções:\n Admiração, Diversão, Raiva, Aborrecimento, Aprovação, Compaixão, Confusão, Curiosidade, Desejo, Decepção, Desaprovação, Nojo, Vergonha, Inveja, Entusiasmo, Medo, Gratidão, Luto, Alegria, Saudade, Amor, Nervosismo, Otimismo, Orgulho, Alívio, Remorso, Tristeza ou Surpresa.\n\n",
754
+ "target_delimiter": " ",
755
+ "fewshot_delimiter": "\n\n",
756
+ "fewshot_config": {
757
+ "sampler": "first_n"
758
+ },
759
+ "num_fewshot": 25,
760
+ "metric_list": [
761
+ {
762
+ "metric": "f1_macro",
763
+ "aggregation": "f1_macro",
764
+ "higher_is_better": true
765
+ },
766
+ {
767
+ "metric": "acc",
768
+ "aggregation": "acc",
769
+ "higher_is_better": true
770
+ }
771
+ ],
772
+ "output_type": "generate_until",
773
+ "generation_kwargs": {
774
+ "max_gen_toks": 32,
775
+ "do_sample": false,
776
+ "temperature": 0.0,
777
+ "top_k": null,
778
+ "top_p": null,
779
+ "until": [
780
+ "\n\n"
781
+ ]
782
+ },
783
+ "repeats": 1,
784
+ "filter_list": [
785
+ {
786
+ "name": "all",
787
+ "filter": [
788
+ {
789
+ "function": "find_similar_label",
790
+ "labels": [
791
+ "Admiração",
792
+ "Diversão",
793
+ "Raiva",
794
+ "Aborrecimento",
795
+ "Aprovação",
796
+ "Compaixão",
797
+ "Confusão",
798
+ "Curiosidade",
799
+ "Desejo",
800
+ "Decepção",
801
+ "Desaprovação",
802
+ "Nojo",
803
+ " Vergonha",
804
+ "Inveja",
805
+ "Entusiasmo",
806
+ "Medo",
807
+ "Gratidão",
808
+ "Luto",
809
+ "Alegria",
810
+ "Saudade",
811
+ "Amor",
812
+ "Nervosismo",
813
+ "Otimismo",
814
+ "Orgulho",
815
+ "Alívio",
816
+ "Remorso",
817
+ "Tristeza",
818
+ "Surpresa"
819
+ ]
820
+ },
821
+ {
822
+ "function": "take_first"
823
+ }
824
+ ]
825
+ }
826
+ ],
827
+ "should_decontaminate": false,
828
+ "limit": 500,
829
+ "metadata": {
830
+ "version": 1.0
831
+ }
832
+ },
833
+ "sparrow_hate-2019-fortuna-por": {
834
+ "task": "sparrow_hate-2019-fortuna-por",
835
+ "task_alias": "hate-2019-fortuna-por",
836
+ "group": [
837
+ "pt_benchmark",
838
+ "sparrow"
839
+ ],
840
+ "dataset_path": "UBC-NLP/sparrow",
841
+ "dataset_name": "hate-2019-fortuna-por",
842
+ "test_split": "validation",
843
+ "fewshot_split": "train",
844
+ "doc_to_text": "Texto: {{content}}\nPergunta: O texto contém discurso de ódio?\nResposta:",
845
+ "doc_to_target": "{{'Sim' if label == 'Hate' else 'Não'}}",
846
+ "description": "Abaixo contém o conteúdo de tweets de usuarios do Twitter em português, sua tarefa é classificar se o texto contem discurso de ódio our não. Responda apenas com Sim ou Não.\n\n",
847
+ "target_delimiter": " ",
848
+ "fewshot_delimiter": "\n\n",
849
+ "fewshot_config": {
850
+ "sampler": "first_n"
851
+ },
852
+ "num_fewshot": 25,
853
+ "metric_list": [
854
+ {
855
+ "metric": "f1_macro",
856
+ "aggregation": "f1_macro",
857
+ "higher_is_better": true
858
+ },
859
+ {
860
+ "metric": "acc",
861
+ "aggregation": "acc",
862
+ "higher_is_better": true
863
+ }
864
+ ],
865
+ "output_type": "generate_until",
866
+ "generation_kwargs": {
867
+ "max_gen_toks": 32,
868
+ "do_sample": false,
869
+ "temperature": 0.0,
870
+ "top_k": null,
871
+ "top_p": null,
872
+ "until": [
873
+ "\n\n"
874
+ ]
875
+ },
876
+ "repeats": 1,
877
+ "filter_list": [
878
+ {
879
+ "name": "all",
880
+ "filter": [
881
+ {
882
+ "function": "find_similar_label",
883
+ "labels": [
884
+ "Sim",
885
+ "Não"
886
+ ]
887
+ },
888
+ {
889
+ "function": "take_first"
890
+ }
891
+ ]
892
+ }
893
+ ],
894
+ "should_decontaminate": false,
895
+ "limit": 500,
896
+ "metadata": {
897
+ "version": 1.0
898
+ }
899
+ },
900
+ "sparrow_sentiment-2016-mozetic-por": {
901
+ "task": "sparrow_sentiment-2016-mozetic-por",
902
+ "task_alias": "sentiment-2016-mozetic-por",
903
+ "group": [
904
+ "pt_benchmark",
905
+ "sparrow"
906
+ ],
907
+ "dataset_path": "UBC-NLP/sparrow",
908
+ "dataset_name": "sentiment-2016-mozetic-por",
909
+ "test_split": "validation",
910
+ "fewshot_split": "train",
911
+ "doc_to_text": "Texto: {{content}}\nPergunta: O sentimento do texto é Positivo, Neutro ou Negativo?\nResposta:",
912
+ "doc_to_target": "{{'Positivo' if label == 'Positive' else ('Negativo' if label == 'Negative' else 'Neutro')}}",
913
+ "description": "Abaixo contém o conteúdo de tweets de usuarios do Twitter em português, sua tarefa é classificar se o sentimento do texto é Positivo, Neutro ou Negativo. Responda apenas com uma das opções.\n\n",
914
+ "target_delimiter": " ",
915
+ "fewshot_delimiter": "\n\n",
916
+ "fewshot_config": {
917
+ "sampler": "first_n"
918
+ },
919
+ "num_fewshot": 25,
920
+ "metric_list": [
921
+ {
922
+ "metric": "f1_macro",
923
+ "aggregation": "f1_macro",
924
+ "higher_is_better": true
925
+ },
926
+ {
927
+ "metric": "acc",
928
+ "aggregation": "acc",
929
+ "higher_is_better": true
930
+ }
931
+ ],
932
+ "output_type": "generate_until",
933
+ "generation_kwargs": {
934
+ "max_gen_toks": 32,
935
+ "do_sample": false,
936
+ "temperature": 0.0,
937
+ "top_k": null,
938
+ "top_p": null,
939
+ "until": [
940
+ "\n\n"
941
+ ]
942
+ },
943
+ "repeats": 1,
944
+ "filter_list": [
945
+ {
946
+ "name": "all",
947
+ "filter": [
948
+ {
949
+ "function": "find_similar_label",
950
+ "labels": [
951
+ "Positivo",
952
+ "Neutro",
953
+ "Negativo"
954
+ ]
955
+ },
956
+ {
957
+ "function": "take_first"
958
+ }
959
+ ]
960
+ }
961
+ ],
962
+ "should_decontaminate": false,
963
+ "limit": 500,
964
+ "metadata": {
965
+ "version": 1.0
966
+ }
967
+ },
968
+ "sparrow_sentiment-2018-brum-por": {
969
+ "task": "sparrow_sentiment-2018-brum-por",
970
+ "task_alias": "sentiment-2018-brum-por",
971
+ "group": [
972
+ "pt_benchmark",
973
+ "sparrow"
974
+ ],
975
+ "dataset_path": "UBC-NLP/sparrow",
976
+ "dataset_name": "sentiment-2018-brum-por",
977
+ "test_split": "validation",
978
+ "fewshot_split": "train",
979
+ "doc_to_text": "Texto: {{content}}\nPergunta: O sentimento do texto é Positivo, Neutro ou Negativo?\nResposta:",
980
+ "doc_to_target": "{{'Positivo' if label == 'Positive' else ('Negativo' if label == 'Negative' else 'Neutro')}}",
981
+ "description": "Abaixo contém o conteúdo de tweets de usuarios do Twitter em português, sua tarefa é classificar se o sentimento do texto é Positivo, Neutro ou Negativo. Responda apenas com uma das opções.\n\n",
982
+ "target_delimiter": " ",
983
+ "fewshot_delimiter": "\n\n",
984
+ "fewshot_config": {
985
+ "sampler": "first_n"
986
+ },
987
+ "num_fewshot": 25,
988
+ "metric_list": [
989
+ {
990
+ "metric": "f1_macro",
991
+ "aggregation": "f1_macro",
992
+ "higher_is_better": true
993
+ },
994
+ {
995
+ "metric": "acc",
996
+ "aggregation": "acc",
997
+ "higher_is_better": true
998
+ }
999
+ ],
1000
+ "output_type": "generate_until",
1001
+ "generation_kwargs": {
1002
+ "max_gen_toks": 32,
1003
+ "do_sample": false,
1004
+ "temperature": 0.0,
1005
+ "top_k": null,
1006
+ "top_p": null,
1007
+ "until": [
1008
+ "\n\n"
1009
+ ]
1010
+ },
1011
+ "repeats": 1,
1012
+ "filter_list": [
1013
+ {
1014
+ "name": "all",
1015
+ "filter": [
1016
+ {
1017
+ "function": "find_similar_label",
1018
+ "labels": [
1019
+ "Positivo",
1020
+ "Neutro",
1021
+ "Negativo"
1022
+ ]
1023
+ },
1024
+ {
1025
+ "function": "take_first"
1026
+ }
1027
+ ]
1028
+ }
1029
+ ],
1030
+ "should_decontaminate": false,
1031
+ "limit": 500,
1032
+ "metadata": {
1033
+ "version": 1.0
1034
+ }
1035
+ }
1036
+ },
1037
+ "versions": {
1038
+ "assin2_rte": 1.0,
1039
+ "assin2_sts": 1.0,
1040
+ "bluex": 1.0,
1041
+ "enem_challenge": 1.0,
1042
+ "faquad_nli": 1.0,
1043
+ "oab_exams": 1.4,
1044
+ "sparrow_emotion-2021-cortiz-por": 1.0,
1045
+ "sparrow_hate-2019-fortuna-por": 1.0,
1046
+ "sparrow_sentiment-2016-mozetic-por": 1.0,
1047
+ "sparrow_sentiment-2018-brum-por": 1.0
1048
+ },
1049
+ "n-shot": {
1050
+ "assin2_rte": 15,
1051
+ "assin2_sts": 15,
1052
+ "bluex": 3,
1053
+ "enem_challenge": 3,
1054
+ "faquad_nli": 15,
1055
+ "oab_exams": 3,
1056
+ "sparrow_emotion-2021-cortiz-por": 25,
1057
+ "sparrow_hate-2019-fortuna-por": 25,
1058
+ "sparrow_sentiment-2016-mozetic-por": 25,
1059
+ "sparrow_sentiment-2018-brum-por": 25
1060
+ },
1061
+ "model_meta": {
1062
+ "truncated": 1,
1063
+ "non_truncated": 11888,
1064
+ "padded": 0,
1065
+ "non_padded": 11889,
1066
+ "fewshots_truncated": 1,
1067
+ "has_chat_template": false,
1068
+ "chat_type": null,
1069
+ "n_gpus": 1,
1070
+ "accelerate_num_process": null,
1071
+ "model_sha": "cc72a88036c2fb937d65efeacc57a0c2ef5d6fe5",
1072
+ "model_dtype": "torch.float16",
1073
+ "model_memory_footprint": 3444817920,
1074
+ "model_num_parameters": 1722408960,
1075
+ "model_is_loaded_in_4bit": false,
1076
+ "model_is_loaded_in_8bit": false,
1077
+ "model_is_quantized": null,
1078
+ "model_device": "cuda:1",
1079
+ "batch_size": 16,
1080
+ "max_length": 2048,
1081
+ "max_ctx_length": 2016,
1082
+ "max_gen_toks": 32
1083
+ },
1084
+ "task_model_meta": {
1085
+ "assin2_rte": {
1086
+ "sample_size": 2448,
1087
+ "truncated": 0,
1088
+ "non_truncated": 2448,
1089
+ "padded": 0,
1090
+ "non_padded": 2448,
1091
+ "fewshots_truncated": 0,
1092
+ "mean_seq_length": 852.9178921568628,
1093
+ "min_seq_length": 838,
1094
+ "max_seq_length": 899,
1095
+ "max_ctx_length": 2016,
1096
+ "max_gen_toks": 32,
1097
+ "mean_original_fewshots_size": 15.0,
1098
+ "mean_effective_fewshot_size": 15.0
1099
+ },
1100
+ "assin2_sts": {
1101
+ "sample_size": 2448,
1102
+ "truncated": 0,
1103
+ "non_truncated": 2448,
1104
+ "padded": 0,
1105
+ "non_padded": 2448,
1106
+ "fewshots_truncated": 0,
1107
+ "mean_seq_length": 1023.9178921568628,
1108
+ "min_seq_length": 1009,
1109
+ "max_seq_length": 1070,
1110
+ "max_ctx_length": 2016,
1111
+ "max_gen_toks": 32,
1112
+ "mean_original_fewshots_size": 15.0,
1113
+ "mean_effective_fewshot_size": 15.0
1114
+ },
1115
+ "bluex": {
1116
+ "sample_size": 719,
1117
+ "truncated": 0,
1118
+ "non_truncated": 719,
1119
+ "padded": 0,
1120
+ "non_padded": 719,
1121
+ "fewshots_truncated": 0,
1122
+ "mean_seq_length": 1037.6175243393602,
1123
+ "min_seq_length": 782,
1124
+ "max_seq_length": 1563,
1125
+ "max_ctx_length": 2016,
1126
+ "max_gen_toks": 32,
1127
+ "mean_original_fewshots_size": 3.0,
1128
+ "mean_effective_fewshot_size": 3.0
1129
+ },
1130
+ "enem_challenge": {
1131
+ "sample_size": 1429,
1132
+ "truncated": 1,
1133
+ "non_truncated": 1428,
1134
+ "padded": 0,
1135
+ "non_padded": 1429,
1136
+ "fewshots_truncated": 1,
1137
+ "mean_seq_length": 1081.372988103569,
1138
+ "min_seq_length": 906,
1139
+ "max_seq_length": 2031,
1140
+ "max_ctx_length": 2016,
1141
+ "max_gen_toks": 32,
1142
+ "mean_original_fewshots_size": 3.0,
1143
+ "mean_effective_fewshot_size": 2.9993002099370187
1144
+ },
1145
+ "faquad_nli": {
1146
+ "sample_size": 650,
1147
+ "truncated": 0,
1148
+ "non_truncated": 650,
1149
+ "padded": 0,
1150
+ "non_padded": 650,
1151
+ "fewshots_truncated": 0,
1152
+ "mean_seq_length": 994.5061538461539,
1153
+ "min_seq_length": 959,
1154
+ "max_seq_length": 1065,
1155
+ "max_ctx_length": 2016,
1156
+ "max_gen_toks": 32,
1157
+ "mean_original_fewshots_size": 15.0,
1158
+ "mean_effective_fewshot_size": 15.0
1159
+ },
1160
+ "oab_exams": {
1161
+ "sample_size": 2195,
1162
+ "truncated": 0,
1163
+ "non_truncated": 2195,
1164
+ "padded": 0,
1165
+ "non_padded": 2195,
1166
+ "fewshots_truncated": 0,
1167
+ "mean_seq_length": 870.5266514806378,
1168
+ "min_seq_length": 691,
1169
+ "max_seq_length": 1156,
1170
+ "max_ctx_length": 2016,
1171
+ "max_gen_toks": 32,
1172
+ "mean_original_fewshots_size": 3.0,
1173
+ "mean_effective_fewshot_size": 3.0
1174
+ },
1175
+ "sparrow_emotion-2021-cortiz-por": {
1176
+ "sample_size": 500,
1177
+ "truncated": 0,
1178
+ "non_truncated": 500,
1179
+ "padded": 0,
1180
+ "non_padded": 500,
1181
+ "fewshots_truncated": 0,
1182
+ "mean_seq_length": 1221.068,
1183
+ "min_seq_length": 1205,
1184
+ "max_seq_length": 1247,
1185
+ "max_ctx_length": 2016,
1186
+ "max_gen_toks": 32,
1187
+ "mean_original_fewshots_size": 25.0,
1188
+ "mean_effective_fewshot_size": 25.0
1189
+ },
1190
+ "sparrow_hate-2019-fortuna-por": {
1191
+ "sample_size": 500,
1192
+ "truncated": 0,
1193
+ "non_truncated": 500,
1194
+ "padded": 0,
1195
+ "non_padded": 500,
1196
+ "fewshots_truncated": 0,
1197
+ "mean_seq_length": 1147.222,
1198
+ "min_seq_length": 1129,
1199
+ "max_seq_length": 1183,
1200
+ "max_ctx_length": 2016,
1201
+ "max_gen_toks": 32,
1202
+ "mean_original_fewshots_size": 25.0,
1203
+ "mean_effective_fewshot_size": 25.0
1204
+ },
1205
+ "sparrow_sentiment-2016-mozetic-por": {
1206
+ "sample_size": 500,
1207
+ "truncated": 0,
1208
+ "non_truncated": 500,
1209
+ "padded": 0,
1210
+ "non_padded": 500,
1211
+ "fewshots_truncated": 0,
1212
+ "mean_seq_length": 960.78,
1213
+ "min_seq_length": 949,
1214
+ "max_seq_length": 988,
1215
+ "max_ctx_length": 2016,
1216
+ "max_gen_toks": 32,
1217
+ "mean_original_fewshots_size": 25.0,
1218
+ "mean_effective_fewshot_size": 25.0
1219
+ },
1220
+ "sparrow_sentiment-2018-brum-por": {
1221
+ "sample_size": 500,
1222
+ "truncated": 0,
1223
+ "non_truncated": 500,
1224
+ "padded": 0,
1225
+ "non_padded": 500,
1226
+ "fewshots_truncated": 0,
1227
+ "mean_seq_length": 1074.998,
1228
+ "min_seq_length": 1063,
1229
+ "max_seq_length": 1099,
1230
+ "max_ctx_length": 2016,
1231
+ "max_gen_toks": 32,
1232
+ "mean_original_fewshots_size": 25.0,
1233
+ "mean_effective_fewshot_size": 25.0
1234
+ }
1235
+ },
1236
+ "config": {
1237
+ "model": "huggingface",
1238
+ "model_args": "pretrained=bigscience/bloom-1b7,dtype=float16,device=cuda:1,revision=main,trust_remote_code=True,starting_max_length=4096",
1239
+ "batch_size": "auto",
1240
+ "batch_sizes": [],
1241
+ "device": null,
1242
+ "use_cache": null,
1243
+ "limit": [
1244
+ null,
1245
+ null,
1246
+ null,
1247
+ null,
1248
+ null,
1249
+ null,
1250
+ 500.0,
1251
+ 500.0,
1252
+ 500.0,
1253
+ 500.0
1254
+ ],
1255
+ "bootstrap_iters": 0,
1256
+ "gen_kwargs": null
1257
+ },
1258
+ "git_hash": "15f86b5"
1259
+ }
bigscience/bloom-1b7/results_2024-02-07T01-17-14.803861.json ADDED
@@ -0,0 +1,346 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config_general": {
3
+ "start_date": "2024-02-07T01-17-14.803861",
4
+ "start_time": 1707268637.1606455,
5
+ "end_time": 1707270556.9489048,
6
+ "total_evaluation_time_seconds": 1919.788259267807,
7
+ "has_chat_template": false,
8
+ "chat_type": null,
9
+ "n_gpus": 1,
10
+ "accelerate_num_process": null,
11
+ "model_sha": "cc72a88036c2fb937d65efeacc57a0c2ef5d6fe5",
12
+ "model_dtype": "float16",
13
+ "model_memory_footprint": 3444817920,
14
+ "model_num_parameters": 1722408960,
15
+ "model_is_loaded_in_4bit": false,
16
+ "model_is_loaded_in_8bit": false,
17
+ "model_is_quantized": null,
18
+ "model_device": "cuda:1",
19
+ "batch_size": 16,
20
+ "max_length": 2048,
21
+ "max_ctx_length": 2016,
22
+ "max_gen_toks": 32,
23
+ "model_name": "bigscience/bloom-1b7",
24
+ "job_id": 26,
25
+ "model_id": "bigscience/bloom-1b7_eval_request_False_float16_Original",
26
+ "model_base_model": "",
27
+ "model_weight_type": "Original",
28
+ "model_revision": "main",
29
+ "model_private": false,
30
+ "model_type": "🟢 : pretrained",
31
+ "model_architectures": "BloomForCausalLM",
32
+ "submitted_time": "2024-02-05T23:04:30Z",
33
+ "lm_eval_model_type": "huggingface",
34
+ "eval_version": "1.0.0"
35
+ },
36
+ "results": {
37
+ "all_grouped_average": 0.23561558006238054,
38
+ "all_grouped_npm": -0.07152931770105994,
39
+ "all_grouped": {
40
+ "enem_challenge": 0.1966410076976907,
41
+ "bluex": 0.23087621696801114,
42
+ "oab_exams": 0.2428246013667426,
43
+ "assin2_rte": 0.3378561857199312,
44
+ "assin2_sts": 0.017927392336598496,
45
+ "faquad_nli": 0.4396551724137931,
46
+ "sparrow_pt": 0.1835284839338963
47
+ },
48
+ "all": {
49
+ "harness|enem_challenge|enem_challenge|None|3": 0.1966410076976907,
50
+ "harness|bluex|bluex|None|3": 0.23087621696801114,
51
+ "harness|oab_exams|oab_exams|None|3": 0.2428246013667426,
52
+ "harness|assin2_rte|assin2_rte|None|15": 0.3378561857199312,
53
+ "harness|assin2_sts|assin2_sts|None|15": 0.017927392336598496,
54
+ "harness|faquad_nli|faquad_nli|None|15": 0.4396551724137931,
55
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por|500|25": 0.048302336566881446,
56
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por|500|25": 0.3932038834951456,
57
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por|500|25": 0.09875428777757718,
58
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por|500|25": 0.19385342789598106
59
+ },
60
+ "harness|enem_challenge|enem_challenge|None|3": {
61
+ "acc,all": 0.1966410076976907,
62
+ "acc,exam_id__2016_2": 0.17073170731707318,
63
+ "acc,exam_id__2016": 0.24793388429752067,
64
+ "acc,exam_id__2010": 0.17094017094017094,
65
+ "acc,exam_id__2014": 0.1743119266055046,
66
+ "acc,exam_id__2023": 0.2222222222222222,
67
+ "acc,exam_id__2017": 0.1724137931034483,
68
+ "acc,exam_id__2015": 0.21008403361344538,
69
+ "acc,exam_id__2012": 0.1206896551724138,
70
+ "acc,exam_id__2022": 0.21804511278195488,
71
+ "acc,exam_id__2013": 0.16666666666666666,
72
+ "acc,exam_id__2009": 0.19130434782608696,
73
+ "acc,exam_id__2011": 0.28205128205128205,
74
+ "main_score": 0.1966410076976907
75
+ },
76
+ "harness|bluex|bluex|None|3": {
77
+ "acc,all": 0.23087621696801114,
78
+ "acc,exam_id__USP_2019": 0.175,
79
+ "acc,exam_id__UNICAMP_2018": 0.24074074074074073,
80
+ "acc,exam_id__UNICAMP_2020": 0.23636363636363636,
81
+ "acc,exam_id__UNICAMP_2023": 0.23255813953488372,
82
+ "acc,exam_id__UNICAMP_2024": 0.2,
83
+ "acc,exam_id__UNICAMP_2019": 0.24,
84
+ "acc,exam_id__UNICAMP_2021_2": 0.19607843137254902,
85
+ "acc,exam_id__USP_2022": 0.30612244897959184,
86
+ "acc,exam_id__USP_2020": 0.30357142857142855,
87
+ "acc,exam_id__USP_2018": 0.2037037037037037,
88
+ "acc,exam_id__USP_2021": 0.19230769230769232,
89
+ "acc,exam_id__USP_2023": 0.25,
90
+ "acc,exam_id__USP_2024": 0.17073170731707318,
91
+ "acc,exam_id__UNICAMP_2021_1": 0.2608695652173913,
92
+ "acc,exam_id__UNICAMP_2022": 0.23076923076923078,
93
+ "main_score": 0.23087621696801114
94
+ },
95
+ "harness|oab_exams|oab_exams|None|3": {
96
+ "acc,all": 0.2428246013667426,
97
+ "acc,exam_id__2014-13": 0.2875,
98
+ "acc,exam_id__2012-07": 0.175,
99
+ "acc,exam_id__2012-09": 0.24675324675324675,
100
+ "acc,exam_id__2011-03": 0.2222222222222222,
101
+ "acc,exam_id__2012-08": 0.275,
102
+ "acc,exam_id__2015-17": 0.20512820512820512,
103
+ "acc,exam_id__2014-15": 0.2564102564102564,
104
+ "acc,exam_id__2017-24": 0.175,
105
+ "acc,exam_id__2015-16": 0.2375,
106
+ "acc,exam_id__2017-23": 0.25,
107
+ "acc,exam_id__2011-04": 0.25,
108
+ "acc,exam_id__2010-02": 0.23,
109
+ "acc,exam_id__2016-19": 0.2564102564102564,
110
+ "acc,exam_id__2012-06": 0.25,
111
+ "acc,exam_id__2012-06a": 0.225,
112
+ "acc,exam_id__2013-12": 0.1875,
113
+ "acc,exam_id__2017-22": 0.3125,
114
+ "acc,exam_id__2010-01": 0.27058823529411763,
115
+ "acc,exam_id__2013-10": 0.225,
116
+ "acc,exam_id__2014-14": 0.25,
117
+ "acc,exam_id__2018-25": 0.2875,
118
+ "acc,exam_id__2016-20a": 0.2875,
119
+ "acc,exam_id__2015-18": 0.2875,
120
+ "acc,exam_id__2011-05": 0.2,
121
+ "acc,exam_id__2013-11": 0.175,
122
+ "acc,exam_id__2016-20": 0.275,
123
+ "acc,exam_id__2016-21": 0.2625,
124
+ "main_score": 0.2428246013667426
125
+ },
126
+ "harness|assin2_rte|assin2_rte|None|15": {
127
+ "f1_macro,all": 0.3378561857199312,
128
+ "acc,all": 0.5020424836601307,
129
+ "main_score": 0.3378561857199312
130
+ },
131
+ "harness|assin2_sts|assin2_sts|None|15": {
132
+ "pearson,all": 0.017927392336598496,
133
+ "mse,all": 2.7831372549019617,
134
+ "main_score": 0.017927392336598496
135
+ },
136
+ "harness|faquad_nli|faquad_nli|None|15": {
137
+ "f1_macro,all": 0.4396551724137931,
138
+ "acc,all": 0.7846153846153846,
139
+ "main_score": 0.4396551724137931
140
+ },
141
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por|500|25": {
142
+ "f1_macro,all": 0.048302336566881446,
143
+ "acc,all": 0.088,
144
+ "main_score": 0.048302336566881446
145
+ },
146
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por|500|25": {
147
+ "f1_macro,all": 0.3932038834951456,
148
+ "acc,all": 0.648,
149
+ "main_score": 0.3932038834951456
150
+ },
151
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por|500|25": {
152
+ "f1_macro,all": 0.09875428777757718,
153
+ "acc,all": 0.152,
154
+ "main_score": 0.09875428777757718
155
+ },
156
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por|500|25": {
157
+ "f1_macro,all": 0.19385342789598106,
158
+ "acc,all": 0.41,
159
+ "main_score": 0.19385342789598106
160
+ }
161
+ },
162
+ "config_tasks": {
163
+ "harness|enem_challenge|enem_challenge": "LM Harness task",
164
+ "harness|bluex|bluex": "LM Harness task",
165
+ "harness|oab_exams|oab_exams": "LM Harness task",
166
+ "harness|assin2_rte|assin2_rte": "LM Harness task",
167
+ "harness|assin2_sts|assin2_sts": "LM Harness task",
168
+ "harness|faquad_nli|faquad_nli": "LM Harness task",
169
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por": "LM Harness task",
170
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por": "LM Harness task",
171
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por": "LM Harness task",
172
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por": "LM Harness task"
173
+ },
174
+ "versions": {
175
+ "all": 0,
176
+ "harness|enem_challenge|enem_challenge": 1.0,
177
+ "harness|bluex|bluex": 1.0,
178
+ "harness|oab_exams|oab_exams": 1.4,
179
+ "harness|assin2_rte|assin2_rte": 1.0,
180
+ "harness|assin2_sts|assin2_sts": 1.0,
181
+ "harness|faquad_nli|faquad_nli": 1.0,
182
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por": 1.0,
183
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por": 1.0,
184
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por": 1.0,
185
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por": 1.0
186
+ },
187
+ "summary_tasks": {
188
+ "harness|enem_challenge|enem_challenge|None|3": {
189
+ "sample_size": 1429,
190
+ "truncated": 1,
191
+ "non_truncated": 1428,
192
+ "padded": 0,
193
+ "non_padded": 1429,
194
+ "fewshots_truncated": 1,
195
+ "mean_seq_length": 1081.372988103569,
196
+ "min_seq_length": 906,
197
+ "max_seq_length": 2031,
198
+ "max_ctx_length": 2016,
199
+ "max_gen_toks": 32,
200
+ "mean_original_fewshots_size": 3.0,
201
+ "mean_effective_fewshot_size": 2.9993002099370187
202
+ },
203
+ "harness|bluex|bluex|None|3": {
204
+ "sample_size": 719,
205
+ "truncated": 0,
206
+ "non_truncated": 719,
207
+ "padded": 0,
208
+ "non_padded": 719,
209
+ "fewshots_truncated": 0,
210
+ "mean_seq_length": 1037.6175243393602,
211
+ "min_seq_length": 782,
212
+ "max_seq_length": 1563,
213
+ "max_ctx_length": 2016,
214
+ "max_gen_toks": 32,
215
+ "mean_original_fewshots_size": 3.0,
216
+ "mean_effective_fewshot_size": 3.0
217
+ },
218
+ "harness|oab_exams|oab_exams|None|3": {
219
+ "sample_size": 2195,
220
+ "truncated": 0,
221
+ "non_truncated": 2195,
222
+ "padded": 0,
223
+ "non_padded": 2195,
224
+ "fewshots_truncated": 0,
225
+ "mean_seq_length": 870.5266514806378,
226
+ "min_seq_length": 691,
227
+ "max_seq_length": 1156,
228
+ "max_ctx_length": 2016,
229
+ "max_gen_toks": 32,
230
+ "mean_original_fewshots_size": 3.0,
231
+ "mean_effective_fewshot_size": 3.0
232
+ },
233
+ "harness|assin2_rte|assin2_rte|None|15": {
234
+ "sample_size": 2448,
235
+ "truncated": 0,
236
+ "non_truncated": 2448,
237
+ "padded": 0,
238
+ "non_padded": 2448,
239
+ "fewshots_truncated": 0,
240
+ "mean_seq_length": 852.9178921568628,
241
+ "min_seq_length": 838,
242
+ "max_seq_length": 899,
243
+ "max_ctx_length": 2016,
244
+ "max_gen_toks": 32,
245
+ "mean_original_fewshots_size": 15.0,
246
+ "mean_effective_fewshot_size": 15.0
247
+ },
248
+ "harness|assin2_sts|assin2_sts|None|15": {
249
+ "sample_size": 2448,
250
+ "truncated": 0,
251
+ "non_truncated": 2448,
252
+ "padded": 0,
253
+ "non_padded": 2448,
254
+ "fewshots_truncated": 0,
255
+ "mean_seq_length": 1023.9178921568628,
256
+ "min_seq_length": 1009,
257
+ "max_seq_length": 1070,
258
+ "max_ctx_length": 2016,
259
+ "max_gen_toks": 32,
260
+ "mean_original_fewshots_size": 15.0,
261
+ "mean_effective_fewshot_size": 15.0
262
+ },
263
+ "harness|faquad_nli|faquad_nli|None|15": {
264
+ "sample_size": 650,
265
+ "truncated": 0,
266
+ "non_truncated": 650,
267
+ "padded": 0,
268
+ "non_padded": 650,
269
+ "fewshots_truncated": 0,
270
+ "mean_seq_length": 994.5061538461539,
271
+ "min_seq_length": 959,
272
+ "max_seq_length": 1065,
273
+ "max_ctx_length": 2016,
274
+ "max_gen_toks": 32,
275
+ "mean_original_fewshots_size": 15.0,
276
+ "mean_effective_fewshot_size": 15.0
277
+ },
278
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por|500|25": {
279
+ "sample_size": 500,
280
+ "truncated": 0,
281
+ "non_truncated": 500,
282
+ "padded": 0,
283
+ "non_padded": 500,
284
+ "fewshots_truncated": 0,
285
+ "mean_seq_length": 1221.068,
286
+ "min_seq_length": 1205,
287
+ "max_seq_length": 1247,
288
+ "max_ctx_length": 2016,
289
+ "max_gen_toks": 32,
290
+ "mean_original_fewshots_size": 25.0,
291
+ "mean_effective_fewshot_size": 25.0
292
+ },
293
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por|500|25": {
294
+ "sample_size": 500,
295
+ "truncated": 0,
296
+ "non_truncated": 500,
297
+ "padded": 0,
298
+ "non_padded": 500,
299
+ "fewshots_truncated": 0,
300
+ "mean_seq_length": 1147.222,
301
+ "min_seq_length": 1129,
302
+ "max_seq_length": 1183,
303
+ "max_ctx_length": 2016,
304
+ "max_gen_toks": 32,
305
+ "mean_original_fewshots_size": 25.0,
306
+ "mean_effective_fewshot_size": 25.0
307
+ },
308
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por|500|25": {
309
+ "sample_size": 500,
310
+ "truncated": 0,
311
+ "non_truncated": 500,
312
+ "padded": 0,
313
+ "non_padded": 500,
314
+ "fewshots_truncated": 0,
315
+ "mean_seq_length": 960.78,
316
+ "min_seq_length": 949,
317
+ "max_seq_length": 988,
318
+ "max_ctx_length": 2016,
319
+ "max_gen_toks": 32,
320
+ "mean_original_fewshots_size": 25.0,
321
+ "mean_effective_fewshot_size": 25.0
322
+ },
323
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por|500|25": {
324
+ "sample_size": 500,
325
+ "truncated": 0,
326
+ "non_truncated": 500,
327
+ "padded": 0,
328
+ "non_padded": 500,
329
+ "fewshots_truncated": 0,
330
+ "mean_seq_length": 1074.998,
331
+ "min_seq_length": 1063,
332
+ "max_seq_length": 1099,
333
+ "max_ctx_length": 2016,
334
+ "max_gen_toks": 32,
335
+ "mean_original_fewshots_size": 25.0,
336
+ "mean_effective_fewshot_size": 25.0
337
+ }
338
+ },
339
+ "summary_general": {
340
+ "truncated": 1,
341
+ "non_truncated": 11888,
342
+ "padded": 0,
343
+ "non_padded": 11889,
344
+ "fewshots_truncated": 1
345
+ }
346
+ }