eduagarcia commited on
Commit
1192833
1 Parent(s): 6d595fa

Uploading raw results for google/umt5-base

Browse files
Files changed (13) hide show
  1. .gitattributes +6 -0
  2. google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_rte.jsonl +3 -0
  3. google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_sts.jsonl +3 -0
  4. google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl +3 -0
  5. google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl +3 -0
  6. google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_faquad_nli.jsonl +3 -0
  7. google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl +3 -0
  8. google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_emotion-2021-cortiz-por.jsonl +0 -0
  9. google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_hate-2019-fortuna-por.jsonl +0 -0
  10. google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_sentiment-2016-mozetic-por.jsonl +0 -0
  11. google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_sentiment-2018-brum-por.jsonl +0 -0
  12. google/umt5-base/raw_2024-02-15T05-17-17.944550/results.json +1259 -0
  13. google/umt5-base/results_2024-02-15T05-17-17.944550.json +346 -0
.gitattributes CHANGED
@@ -599,3 +599,9 @@ openlm-research/open_llama_3b/raw_2024-02-16T07-23-49.395940/pretrained__openlm-
599
  openlm-research/open_llama_3b/raw_2024-02-16T07-23-49.395940/pretrained__openlm-research__open_llama_3b,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl filter=lfs diff=lfs merge=lfs -text
600
  openlm-research/open_llama_3b/raw_2024-02-16T07-23-49.395940/pretrained__openlm-research__open_llama_3b,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl filter=lfs diff=lfs merge=lfs -text
601
  openlm-research/open_llama_3b/raw_2024-02-16T07-23-49.395940/pretrained__openlm-research__open_llama_3b,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
599
  openlm-research/open_llama_3b/raw_2024-02-16T07-23-49.395940/pretrained__openlm-research__open_llama_3b,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl filter=lfs diff=lfs merge=lfs -text
600
  openlm-research/open_llama_3b/raw_2024-02-16T07-23-49.395940/pretrained__openlm-research__open_llama_3b,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl filter=lfs diff=lfs merge=lfs -text
601
  openlm-research/open_llama_3b/raw_2024-02-16T07-23-49.395940/pretrained__openlm-research__open_llama_3b,dtype__float16,device__cuda:1,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl filter=lfs diff=lfs merge=lfs -text
602
+ google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_rte.jsonl filter=lfs diff=lfs merge=lfs -text
603
+ google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_sts.jsonl filter=lfs diff=lfs merge=lfs -text
604
+ google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl filter=lfs diff=lfs merge=lfs -text
605
+ google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl filter=lfs diff=lfs merge=lfs -text
606
+ google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_faquad_nli.jsonl filter=lfs diff=lfs merge=lfs -text
607
+ google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl filter=lfs diff=lfs merge=lfs -text
google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_rte.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7363f4aeb05e8642e234a7d4f0f7e53869e6245a0a489f659c784f3df36b365c
3
+ size 30349200
google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_sts.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be954cdd914ce71343337229693fb2f834ab845bdeab066a37f201cb59cf6835
3
+ size 33386340
google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db62f84b7a146352a75c46129c09380ae9ea998d3ead41915afaefc0df26aa85
3
+ size 11429981
google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a6a343d54b7aebca982b1a5addcea920cd0be6d2daadf0f3de7915d950f661d
3
+ size 25788472
google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_faquad_nli.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61bdf02a951b2d13fd6bae68366a5ce89174821aaf8e5c371ddd062cf8c08108
3
+ size 11064384
google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc8ed4b366b1379990b57d4d610173842225a7f85a8ee0aaf39a0d65f280c10f
3
+ size 34167269
google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_emotion-2021-cortiz-por.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_hate-2019-fortuna-por.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_sentiment-2016-mozetic-por.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
google/umt5-base/raw_2024-02-15T05-17-17.944550/pretrained__google__umt5-base,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_sentiment-2018-brum-por.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
google/umt5-base/raw_2024-02-15T05-17-17.944550/results.json ADDED
@@ -0,0 +1,1259 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "assin2_rte": {
4
+ "f1_macro,all": 0.008049597487606372,
5
+ "acc,all": 0.006127450980392157,
6
+ "alias": "assin2_rte"
7
+ },
8
+ "assin2_sts": {
9
+ "pearson,all": 0.11702145458274152,
10
+ "mse,all": 4.536915664667917,
11
+ "alias": "assin2_sts"
12
+ },
13
+ "bluex": {
14
+ "acc,all": 0.022253129346314324,
15
+ "acc,exam_id__UNICAMP_2022": 0.0,
16
+ "acc,exam_id__UNICAMP_2019": 0.06,
17
+ "acc,exam_id__UNICAMP_2023": 0.023255813953488372,
18
+ "acc,exam_id__UNICAMP_2018": 0.0,
19
+ "acc,exam_id__UNICAMP_2020": 0.05454545454545454,
20
+ "acc,exam_id__USP_2024": 0.0,
21
+ "acc,exam_id__UNICAMP_2021_2": 0.0,
22
+ "acc,exam_id__USP_2022": 0.0,
23
+ "acc,exam_id__USP_2019": 0.0,
24
+ "acc,exam_id__USP_2021": 0.019230769230769232,
25
+ "acc,exam_id__USP_2023": 0.09090909090909091,
26
+ "acc,exam_id__UNICAMP_2021_1": 0.021739130434782608,
27
+ "acc,exam_id__UNICAMP_2024": 0.022222222222222223,
28
+ "acc,exam_id__USP_2020": 0.017857142857142856,
29
+ "acc,exam_id__USP_2018": 0.018518518518518517,
30
+ "alias": "bluex"
31
+ },
32
+ "enem_challenge": {
33
+ "alias": "enem",
34
+ "acc,all": 0.02939118264520644,
35
+ "acc,exam_id__2015": 0.03361344537815126,
36
+ "acc,exam_id__2012": 0.034482758620689655,
37
+ "acc,exam_id__2017": 0.02586206896551724,
38
+ "acc,exam_id__2009": 0.017391304347826087,
39
+ "acc,exam_id__2023": 0.037037037037037035,
40
+ "acc,exam_id__2016_2": 0.024390243902439025,
41
+ "acc,exam_id__2010": 0.05128205128205128,
42
+ "acc,exam_id__2014": 0.01834862385321101,
43
+ "acc,exam_id__2013": 0.018518518518518517,
44
+ "acc,exam_id__2022": 0.05263157894736842,
45
+ "acc,exam_id__2011": 0.02564102564102564,
46
+ "acc,exam_id__2016": 0.008264462809917356
47
+ },
48
+ "faquad_nli": {
49
+ "f1_macro,all": 0.005177993527508091,
50
+ "acc,all": 0.006153846153846154,
51
+ "alias": "faquad_nli"
52
+ },
53
+ "oab_exams": {
54
+ "acc,all": 0.08337129840546698,
55
+ "acc,exam_id__2012-06a": 0.1,
56
+ "acc,exam_id__2015-17": 0.10256410256410256,
57
+ "acc,exam_id__2012-06": 0.0625,
58
+ "acc,exam_id__2014-15": 0.07692307692307693,
59
+ "acc,exam_id__2016-20": 0.075,
60
+ "acc,exam_id__2013-10": 0.0875,
61
+ "acc,exam_id__2012-07": 0.025,
62
+ "acc,exam_id__2011-03": 0.10101010101010101,
63
+ "acc,exam_id__2011-05": 0.1125,
64
+ "acc,exam_id__2016-19": 0.038461538461538464,
65
+ "acc,exam_id__2017-23": 0.1,
66
+ "acc,exam_id__2017-22": 0.0625,
67
+ "acc,exam_id__2018-25": 0.1,
68
+ "acc,exam_id__2014-13": 0.05,
69
+ "acc,exam_id__2017-24": 0.0625,
70
+ "acc,exam_id__2010-01": 0.12941176470588237,
71
+ "acc,exam_id__2014-14": 0.125,
72
+ "acc,exam_id__2011-04": 0.0625,
73
+ "acc,exam_id__2013-12": 0.0625,
74
+ "acc,exam_id__2015-16": 0.1,
75
+ "acc,exam_id__2016-21": 0.075,
76
+ "acc,exam_id__2013-11": 0.075,
77
+ "acc,exam_id__2012-08": 0.025,
78
+ "acc,exam_id__2016-20a": 0.1375,
79
+ "acc,exam_id__2012-09": 0.07792207792207792,
80
+ "acc,exam_id__2010-02": 0.11,
81
+ "acc,exam_id__2015-18": 0.1,
82
+ "alias": "oab_exams"
83
+ },
84
+ "sparrow_emotion-2021-cortiz-por": {
85
+ "alias": "emotion-2021-cortiz-por",
86
+ "f1_macro,all": 0.0,
87
+ "acc,all": 0.0
88
+ },
89
+ "sparrow_hate-2019-fortuna-por": {
90
+ "alias": "hate-2019-fortuna-por",
91
+ "f1_macro,all": 0.24175824175824176,
92
+ "acc,all": 0.528
93
+ },
94
+ "sparrow_sentiment-2016-mozetic-por": {
95
+ "alias": "sentiment-2016-mozetic-por",
96
+ "f1_macro,all": 0.0,
97
+ "acc,all": 0.0
98
+ },
99
+ "sparrow_sentiment-2018-brum-por": {
100
+ "alias": "sentiment-2018-brum-por",
101
+ "f1_macro,all": 0.03459119496855346,
102
+ "acc,all": 0.022
103
+ }
104
+ },
105
+ "configs": {
106
+ "assin2_rte": {
107
+ "task": "assin2_rte",
108
+ "group": [
109
+ "pt_benchmark",
110
+ "assin2"
111
+ ],
112
+ "dataset_path": "assin2",
113
+ "test_split": "test",
114
+ "fewshot_split": "train",
115
+ "doc_to_text": "Premissa: {{premise}}\nHipótese: {{hypothesis}}\nPergunta: A hipótese pode ser inferida pela premissa?\nResposta:",
116
+ "doc_to_target": "{{['Não', 'Sim'][entailment_judgment]}}",
117
+ "description": "Abaixo contém pares de premissa e hipótese, para cada par você deve julgar se a hipótese pode ser inferida a partir da premissa, responda apenas com Sim ou Não.\n\n",
118
+ "target_delimiter": " ",
119
+ "fewshot_delimiter": "\n\n",
120
+ "fewshot_config": {
121
+ "sampler": "id_sampler",
122
+ "sampler_config": {
123
+ "id_list": [
124
+ 1,
125
+ 3251,
126
+ 2,
127
+ 3252,
128
+ 3,
129
+ 4,
130
+ 5,
131
+ 6,
132
+ 3253,
133
+ 7,
134
+ 3254,
135
+ 3255,
136
+ 3256,
137
+ 8,
138
+ 9,
139
+ 10,
140
+ 3257,
141
+ 11,
142
+ 3258,
143
+ 12,
144
+ 13,
145
+ 14,
146
+ 15,
147
+ 3259,
148
+ 3260,
149
+ 3261,
150
+ 3262,
151
+ 3263,
152
+ 16,
153
+ 17,
154
+ 3264,
155
+ 18,
156
+ 3265,
157
+ 3266,
158
+ 3267,
159
+ 19,
160
+ 20,
161
+ 3268,
162
+ 3269,
163
+ 21,
164
+ 3270,
165
+ 3271,
166
+ 22,
167
+ 3272,
168
+ 3273,
169
+ 23,
170
+ 3274,
171
+ 24,
172
+ 25,
173
+ 3275
174
+ ],
175
+ "id_column": "sentence_pair_id"
176
+ }
177
+ },
178
+ "num_fewshot": 15,
179
+ "metric_list": [
180
+ {
181
+ "metric": "f1_macro",
182
+ "aggregation": "f1_macro",
183
+ "higher_is_better": true
184
+ },
185
+ {
186
+ "metric": "acc",
187
+ "aggregation": "acc",
188
+ "higher_is_better": true
189
+ }
190
+ ],
191
+ "output_type": "generate_until",
192
+ "generation_kwargs": {
193
+ "max_gen_toks": 32,
194
+ "do_sample": false,
195
+ "temperature": 0.0,
196
+ "top_k": null,
197
+ "top_p": null,
198
+ "until": [
199
+ "\n\n"
200
+ ]
201
+ },
202
+ "repeats": 1,
203
+ "filter_list": [
204
+ {
205
+ "name": "all",
206
+ "filter": [
207
+ {
208
+ "function": "find_similar_label",
209
+ "labels": [
210
+ "Sim",
211
+ "Não"
212
+ ]
213
+ },
214
+ {
215
+ "function": "take_first"
216
+ }
217
+ ]
218
+ }
219
+ ],
220
+ "should_decontaminate": false,
221
+ "metadata": {
222
+ "version": 1.0
223
+ }
224
+ },
225
+ "assin2_sts": {
226
+ "task": "assin2_sts",
227
+ "group": [
228
+ "pt_benchmark",
229
+ "assin2"
230
+ ],
231
+ "dataset_path": "assin2",
232
+ "test_split": "test",
233
+ "fewshot_split": "train",
234
+ "doc_to_text": "Frase 1: {{premise}}\nFrase 2: {{hypothesis}}\nPergunta: Qual o grau de similaridade entre as duas frases de 1,0 a 5,0?\nResposta:",
235
+ "doc_to_target": "<function assin2_float_to_pt_str at 0x7f96ddc11800>",
236
+ "description": "Abaixo contém pares de frases, para cada par você deve julgar o grau de similaridade de 1,0 a 5,0, responda apenas com o número.\n\n",
237
+ "target_delimiter": " ",
238
+ "fewshot_delimiter": "\n\n",
239
+ "fewshot_config": {
240
+ "sampler": "id_sampler",
241
+ "sampler_config": {
242
+ "id_list": [
243
+ 1,
244
+ 3251,
245
+ 2,
246
+ 3252,
247
+ 3,
248
+ 4,
249
+ 5,
250
+ 6,
251
+ 3253,
252
+ 7,
253
+ 3254,
254
+ 3255,
255
+ 3256,
256
+ 8,
257
+ 9,
258
+ 10,
259
+ 3257,
260
+ 11,
261
+ 3258,
262
+ 12,
263
+ 13,
264
+ 14,
265
+ 15,
266
+ 3259,
267
+ 3260,
268
+ 3261,
269
+ 3262,
270
+ 3263,
271
+ 16,
272
+ 17,
273
+ 3264,
274
+ 18,
275
+ 3265,
276
+ 3266,
277
+ 3267,
278
+ 19,
279
+ 20,
280
+ 3268,
281
+ 3269,
282
+ 21,
283
+ 3270,
284
+ 3271,
285
+ 22,
286
+ 3272,
287
+ 3273,
288
+ 23,
289
+ 3274,
290
+ 24,
291
+ 25,
292
+ 3275
293
+ ],
294
+ "id_column": "sentence_pair_id"
295
+ }
296
+ },
297
+ "num_fewshot": 15,
298
+ "metric_list": [
299
+ {
300
+ "metric": "pearson",
301
+ "aggregation": "pearsonr",
302
+ "higher_is_better": true
303
+ },
304
+ {
305
+ "metric": "mse",
306
+ "aggregation": "mean_squared_error",
307
+ "higher_is_better": false
308
+ }
309
+ ],
310
+ "output_type": "generate_until",
311
+ "generation_kwargs": {
312
+ "max_gen_toks": 32,
313
+ "do_sample": false,
314
+ "temperature": 0.0,
315
+ "top_k": null,
316
+ "top_p": null,
317
+ "until": [
318
+ "\n\n"
319
+ ]
320
+ },
321
+ "repeats": 1,
322
+ "filter_list": [
323
+ {
324
+ "name": "all",
325
+ "filter": [
326
+ {
327
+ "function": "number_filter",
328
+ "type": "float",
329
+ "range_min": 1.0,
330
+ "range_max": 5.0,
331
+ "on_outside_range": "clip",
332
+ "fallback": 5.0
333
+ },
334
+ {
335
+ "function": "take_first"
336
+ }
337
+ ]
338
+ }
339
+ ],
340
+ "should_decontaminate": false,
341
+ "metadata": {
342
+ "version": 1.0
343
+ }
344
+ },
345
+ "bluex": {
346
+ "task": "bluex",
347
+ "group": [
348
+ "pt_benchmark",
349
+ "vestibular"
350
+ ],
351
+ "dataset_path": "eduagarcia-temp/BLUEX_without_images",
352
+ "test_split": "train",
353
+ "fewshot_split": "train",
354
+ "doc_to_text": "<function enem_doc_to_text at 0x7f96ddc111c0>",
355
+ "doc_to_target": "{{answerKey}}",
356
+ "description": "As perguntas a seguir são questões de multipla escolha de provas de vestibular de Universidades Brasileiras, reponda apenas com as letras A, B, C, D ou E.\n\n",
357
+ "target_delimiter": " ",
358
+ "fewshot_delimiter": "\n\n",
359
+ "fewshot_config": {
360
+ "sampler": "id_sampler",
361
+ "sampler_config": {
362
+ "id_list": [
363
+ "USP_2018_3",
364
+ "UNICAMP_2018_2",
365
+ "USP_2018_35",
366
+ "UNICAMP_2018_16",
367
+ "USP_2018_89"
368
+ ],
369
+ "id_column": "id",
370
+ "exclude_from_task": true
371
+ }
372
+ },
373
+ "num_fewshot": 3,
374
+ "metric_list": [
375
+ {
376
+ "metric": "acc",
377
+ "aggregation": "acc",
378
+ "higher_is_better": true
379
+ }
380
+ ],
381
+ "output_type": "generate_until",
382
+ "generation_kwargs": {
383
+ "max_gen_toks": 32,
384
+ "do_sample": false,
385
+ "temperature": 0.0,
386
+ "top_k": null,
387
+ "top_p": null,
388
+ "until": [
389
+ "\n\n"
390
+ ]
391
+ },
392
+ "repeats": 1,
393
+ "filter_list": [
394
+ {
395
+ "name": "all",
396
+ "filter": [
397
+ {
398
+ "function": "normalize_spaces"
399
+ },
400
+ {
401
+ "function": "remove_accents"
402
+ },
403
+ {
404
+ "function": "find_choices",
405
+ "choices": [
406
+ "A",
407
+ "B",
408
+ "C",
409
+ "D",
410
+ "E"
411
+ ],
412
+ "regex_patterns": [
413
+ "(?:[Ll]etra|[Aa]lternativa|[Rr]esposta|[Rr]esposta [Cc]orreta|[Rr]esposta[Cc]orreta e|[Oo]pcao):? ([ABCDE])\\b",
414
+ "\\b([ABCDE])\\.",
415
+ "\\b([ABCDE]) ?[.):-]",
416
+ "\\b([ABCDE])$",
417
+ "\\b([ABCDE])\\b"
418
+ ]
419
+ },
420
+ {
421
+ "function": "take_first"
422
+ }
423
+ ],
424
+ "group_by": {
425
+ "column": "exam_id"
426
+ }
427
+ }
428
+ ],
429
+ "should_decontaminate": true,
430
+ "doc_to_decontamination_query": "<function enem_doc_to_text at 0x7f96ddc11440>",
431
+ "metadata": {
432
+ "version": 1.0
433
+ }
434
+ },
435
+ "enem_challenge": {
436
+ "task": "enem_challenge",
437
+ "task_alias": "enem",
438
+ "group": [
439
+ "pt_benchmark",
440
+ "vestibular"
441
+ ],
442
+ "dataset_path": "eduagarcia/enem_challenge",
443
+ "test_split": "train",
444
+ "fewshot_split": "train",
445
+ "doc_to_text": "<function enem_doc_to_text at 0x7f96ddc119e0>",
446
+ "doc_to_target": "{{answerKey}}",
447
+ "description": "As perguntas a seguir são questões de multipla escolha do Exame Nacional do Ensino Médio (ENEM), reponda apenas com as letras A, B, C, D ou E.\n\n",
448
+ "target_delimiter": " ",
449
+ "fewshot_delimiter": "\n\n",
450
+ "fewshot_config": {
451
+ "sampler": "id_sampler",
452
+ "sampler_config": {
453
+ "id_list": [
454
+ "2022_21",
455
+ "2022_88",
456
+ "2022_143"
457
+ ],
458
+ "id_column": "id",
459
+ "exclude_from_task": true
460
+ }
461
+ },
462
+ "num_fewshot": 3,
463
+ "metric_list": [
464
+ {
465
+ "metric": "acc",
466
+ "aggregation": "acc",
467
+ "higher_is_better": true
468
+ }
469
+ ],
470
+ "output_type": "generate_until",
471
+ "generation_kwargs": {
472
+ "max_gen_toks": 32,
473
+ "do_sample": false,
474
+ "temperature": 0.0,
475
+ "top_k": null,
476
+ "top_p": null,
477
+ "until": [
478
+ "\n\n"
479
+ ]
480
+ },
481
+ "repeats": 1,
482
+ "filter_list": [
483
+ {
484
+ "name": "all",
485
+ "filter": [
486
+ {
487
+ "function": "normalize_spaces"
488
+ },
489
+ {
490
+ "function": "remove_accents"
491
+ },
492
+ {
493
+ "function": "find_choices",
494
+ "choices": [
495
+ "A",
496
+ "B",
497
+ "C",
498
+ "D",
499
+ "E"
500
+ ],
501
+ "regex_patterns": [
502
+ "(?:[Ll]etra|[Aa]lternativa|[Rr]esposta|[Rr]esposta [Cc]orreta|[Rr]esposta[Cc]orreta e|[Oo]pcao):? ([ABCDE])\\b",
503
+ "\\b([ABCDE])\\.",
504
+ "\\b([ABCDE]) ?[.):-]",
505
+ "\\b([ABCDE])$",
506
+ "\\b([ABCDE])\\b"
507
+ ]
508
+ },
509
+ {
510
+ "function": "take_first"
511
+ }
512
+ ],
513
+ "group_by": {
514
+ "column": "exam_id"
515
+ }
516
+ }
517
+ ],
518
+ "should_decontaminate": true,
519
+ "doc_to_decontamination_query": "<function enem_doc_to_text at 0x7f96ddc11c60>",
520
+ "metadata": {
521
+ "version": 1.0
522
+ }
523
+ },
524
+ "faquad_nli": {
525
+ "task": "faquad_nli",
526
+ "group": [
527
+ "pt_benchmark"
528
+ ],
529
+ "dataset_path": "ruanchaves/faquad-nli",
530
+ "test_split": "test",
531
+ "fewshot_split": "train",
532
+ "doc_to_text": "Pergunta: {{question}}\nResposta: {{answer}}\nA resposta satisfaz a pergunta? Sim ou Não?",
533
+ "doc_to_target": "{{['Não', 'Sim'][label]}}",
534
+ "description": "Abaixo contém pares de pergunta e reposta, para cada par você deve julgar resposta responde a pergunta de maneira satisfatória e aparenta estar correta, escreva apenas Sim ou Não.\n\n",
535
+ "target_delimiter": " ",
536
+ "fewshot_delimiter": "\n\n",
537
+ "fewshot_config": {
538
+ "sampler": "first_n",
539
+ "sampler_config": {
540
+ "fewshot_indices": [
541
+ 1893,
542
+ 949,
543
+ 663,
544
+ 105,
545
+ 1169,
546
+ 2910,
547
+ 2227,
548
+ 2813,
549
+ 974,
550
+ 558,
551
+ 1503,
552
+ 1958,
553
+ 2918,
554
+ 601,
555
+ 1560,
556
+ 984,
557
+ 2388,
558
+ 995,
559
+ 2233,
560
+ 1982,
561
+ 165,
562
+ 2788,
563
+ 1312,
564
+ 2285,
565
+ 522,
566
+ 1113,
567
+ 1670,
568
+ 323,
569
+ 236,
570
+ 1263,
571
+ 1562,
572
+ 2519,
573
+ 1049,
574
+ 432,
575
+ 1167,
576
+ 1394,
577
+ 2022,
578
+ 2551,
579
+ 2194,
580
+ 2187,
581
+ 2282,
582
+ 2816,
583
+ 108,
584
+ 301,
585
+ 1185,
586
+ 1315,
587
+ 1420,
588
+ 2436,
589
+ 2322,
590
+ 766
591
+ ]
592
+ }
593
+ },
594
+ "num_fewshot": 15,
595
+ "metric_list": [
596
+ {
597
+ "metric": "f1_macro",
598
+ "aggregation": "f1_macro",
599
+ "higher_is_better": true
600
+ },
601
+ {
602
+ "metric": "acc",
603
+ "aggregation": "acc",
604
+ "higher_is_better": true
605
+ }
606
+ ],
607
+ "output_type": "generate_until",
608
+ "generation_kwargs": {
609
+ "max_gen_toks": 32,
610
+ "do_sample": false,
611
+ "temperature": 0.0,
612
+ "top_k": null,
613
+ "top_p": null,
614
+ "until": [
615
+ "\n\n"
616
+ ]
617
+ },
618
+ "repeats": 1,
619
+ "filter_list": [
620
+ {
621
+ "name": "all",
622
+ "filter": [
623
+ {
624
+ "function": "find_similar_label",
625
+ "labels": [
626
+ "Sim",
627
+ "Não"
628
+ ]
629
+ },
630
+ {
631
+ "function": "take_first"
632
+ }
633
+ ]
634
+ }
635
+ ],
636
+ "should_decontaminate": false,
637
+ "metadata": {
638
+ "version": 1.0
639
+ }
640
+ },
641
+ "oab_exams": {
642
+ "task": "oab_exams",
643
+ "group": [
644
+ "legal_benchmark",
645
+ "pt_benchmark"
646
+ ],
647
+ "dataset_path": "eduagarcia/oab_exams",
648
+ "test_split": "train",
649
+ "fewshot_split": "train",
650
+ "doc_to_text": "<function doc_to_text at 0x7f96ddc10b80>",
651
+ "doc_to_target": "{{answerKey}}",
652
+ "description": "As perguntas a seguir são questões de multipla escolha do Exame de Ordem da Ordem dos Advogados do Brasil (OAB), reponda apenas com as letras A, B, C ou D.\n\n",
653
+ "target_delimiter": " ",
654
+ "fewshot_delimiter": "\n\n",
655
+ "fewshot_config": {
656
+ "sampler": "id_sampler",
657
+ "sampler_config": {
658
+ "id_list": [
659
+ "2010-01_1",
660
+ "2010-01_11",
661
+ "2010-01_13",
662
+ "2010-01_23",
663
+ "2010-01_26",
664
+ "2010-01_28",
665
+ "2010-01_38",
666
+ "2010-01_48",
667
+ "2010-01_58",
668
+ "2010-01_68",
669
+ "2010-01_76",
670
+ "2010-01_83",
671
+ "2010-01_85",
672
+ "2010-01_91",
673
+ "2010-01_99"
674
+ ],
675
+ "id_column": "id",
676
+ "exclude_from_task": true
677
+ }
678
+ },
679
+ "num_fewshot": 3,
680
+ "metric_list": [
681
+ {
682
+ "metric": "acc",
683
+ "aggregation": "acc",
684
+ "higher_is_better": true
685
+ }
686
+ ],
687
+ "output_type": "generate_until",
688
+ "generation_kwargs": {
689
+ "max_gen_toks": 32,
690
+ "do_sample": false,
691
+ "temperature": 0.0,
692
+ "top_k": null,
693
+ "top_p": null,
694
+ "until": [
695
+ "\n\n"
696
+ ]
697
+ },
698
+ "repeats": 1,
699
+ "filter_list": [
700
+ {
701
+ "name": "all",
702
+ "filter": [
703
+ {
704
+ "function": "normalize_spaces"
705
+ },
706
+ {
707
+ "function": "remove_accents"
708
+ },
709
+ {
710
+ "function": "find_choices",
711
+ "choices": [
712
+ "A",
713
+ "B",
714
+ "C",
715
+ "D"
716
+ ],
717
+ "regex_patterns": [
718
+ "(?:[Ll]etra|[Aa]lternativa|[Rr]esposta|[Rr]esposta [Cc]orreta|[Rr]esposta[Cc]orreta e|[Oo]pcao):? ([ABCD])\\b",
719
+ "\\b([ABCD])\\)",
720
+ "\\b([ABCD]) ?[.):-]",
721
+ "\\b([ABCD])$",
722
+ "\\b([ABCD])\\b"
723
+ ]
724
+ },
725
+ {
726
+ "function": "take_first"
727
+ }
728
+ ],
729
+ "group_by": {
730
+ "column": "exam_id"
731
+ }
732
+ }
733
+ ],
734
+ "should_decontaminate": true,
735
+ "doc_to_decontamination_query": "<function doc_to_text at 0x7f96ddc10e00>",
736
+ "metadata": {
737
+ "version": 1.4
738
+ }
739
+ },
740
+ "sparrow_emotion-2021-cortiz-por": {
741
+ "task": "sparrow_emotion-2021-cortiz-por",
742
+ "task_alias": "emotion-2021-cortiz-por",
743
+ "group": [
744
+ "pt_benchmark",
745
+ "sparrow"
746
+ ],
747
+ "dataset_path": "UBC-NLP/sparrow",
748
+ "dataset_name": "emotion-2021-cortiz-por",
749
+ "test_split": "validation",
750
+ "fewshot_split": "train",
751
+ "doc_to_text": "Texto: {{content}}\nPergunta: Qual a principal emoção apresentada no texto?\nResposta:",
752
+ "doc_to_target": "<function sparrow_emotion_por_trans_label at 0x7f96ddc11080>",
753
+ "description": "Abaixo contém o conteúdo de tweets de usuarios do Twitter em português, sua tarefa é extrair qual a principal emoção dos textos. Responda com apenas uma das seguintes opções:\n Admiração, Diversão, Raiva, Aborrecimento, Aprovação, Compaixão, Confusão, Curiosidade, Desejo, Decepção, Desaprovação, Nojo, Vergonha, Inveja, Entusiasmo, Medo, Gratidão, Luto, Alegria, Saudade, Amor, Nervosismo, Otimismo, Orgulho, Alívio, Remorso, Tristeza ou Surpresa.\n\n",
754
+ "target_delimiter": " ",
755
+ "fewshot_delimiter": "\n\n",
756
+ "fewshot_config": {
757
+ "sampler": "first_n"
758
+ },
759
+ "num_fewshot": 25,
760
+ "metric_list": [
761
+ {
762
+ "metric": "f1_macro",
763
+ "aggregation": "f1_macro",
764
+ "higher_is_better": true
765
+ },
766
+ {
767
+ "metric": "acc",
768
+ "aggregation": "acc",
769
+ "higher_is_better": true
770
+ }
771
+ ],
772
+ "output_type": "generate_until",
773
+ "generation_kwargs": {
774
+ "max_gen_toks": 32,
775
+ "do_sample": false,
776
+ "temperature": 0.0,
777
+ "top_k": null,
778
+ "top_p": null,
779
+ "until": [
780
+ "\n\n"
781
+ ]
782
+ },
783
+ "repeats": 1,
784
+ "filter_list": [
785
+ {
786
+ "name": "all",
787
+ "filter": [
788
+ {
789
+ "function": "find_similar_label",
790
+ "labels": [
791
+ "Admiração",
792
+ "Diversão",
793
+ "Raiva",
794
+ "Aborrecimento",
795
+ "Aprovação",
796
+ "Compaixão",
797
+ "Confusão",
798
+ "Curiosidade",
799
+ "Desejo",
800
+ "Decepção",
801
+ "Desaprovação",
802
+ "Nojo",
803
+ " Vergonha",
804
+ "Inveja",
805
+ "Entusiasmo",
806
+ "Medo",
807
+ "Gratidão",
808
+ "Luto",
809
+ "Alegria",
810
+ "Saudade",
811
+ "Amor",
812
+ "Nervosismo",
813
+ "Otimismo",
814
+ "Orgulho",
815
+ "Alívio",
816
+ "Remorso",
817
+ "Tristeza",
818
+ "Surpresa"
819
+ ]
820
+ },
821
+ {
822
+ "function": "take_first"
823
+ }
824
+ ]
825
+ }
826
+ ],
827
+ "should_decontaminate": false,
828
+ "limit": 500,
829
+ "metadata": {
830
+ "version": 1.0
831
+ }
832
+ },
833
+ "sparrow_hate-2019-fortuna-por": {
834
+ "task": "sparrow_hate-2019-fortuna-por",
835
+ "task_alias": "hate-2019-fortuna-por",
836
+ "group": [
837
+ "pt_benchmark",
838
+ "sparrow"
839
+ ],
840
+ "dataset_path": "UBC-NLP/sparrow",
841
+ "dataset_name": "hate-2019-fortuna-por",
842
+ "test_split": "validation",
843
+ "fewshot_split": "train",
844
+ "doc_to_text": "Texto: {{content}}\nPergunta: O texto contém discurso de ódio?\nResposta:",
845
+ "doc_to_target": "{{'Sim' if label == 'Hate' else 'Não'}}",
846
+ "description": "Abaixo contém o conteúdo de tweets de usuarios do Twitter em português, sua tarefa é classificar se o texto contem discurso de ódio our não. Responda apenas com Sim ou Não.\n\n",
847
+ "target_delimiter": " ",
848
+ "fewshot_delimiter": "\n\n",
849
+ "fewshot_config": {
850
+ "sampler": "first_n"
851
+ },
852
+ "num_fewshot": 25,
853
+ "metric_list": [
854
+ {
855
+ "metric": "f1_macro",
856
+ "aggregation": "f1_macro",
857
+ "higher_is_better": true
858
+ },
859
+ {
860
+ "metric": "acc",
861
+ "aggregation": "acc",
862
+ "higher_is_better": true
863
+ }
864
+ ],
865
+ "output_type": "generate_until",
866
+ "generation_kwargs": {
867
+ "max_gen_toks": 32,
868
+ "do_sample": false,
869
+ "temperature": 0.0,
870
+ "top_k": null,
871
+ "top_p": null,
872
+ "until": [
873
+ "\n\n"
874
+ ]
875
+ },
876
+ "repeats": 1,
877
+ "filter_list": [
878
+ {
879
+ "name": "all",
880
+ "filter": [
881
+ {
882
+ "function": "find_similar_label",
883
+ "labels": [
884
+ "Sim",
885
+ "Não"
886
+ ]
887
+ },
888
+ {
889
+ "function": "take_first"
890
+ }
891
+ ]
892
+ }
893
+ ],
894
+ "should_decontaminate": false,
895
+ "limit": 500,
896
+ "metadata": {
897
+ "version": 1.0
898
+ }
899
+ },
900
+ "sparrow_sentiment-2016-mozetic-por": {
901
+ "task": "sparrow_sentiment-2016-mozetic-por",
902
+ "task_alias": "sentiment-2016-mozetic-por",
903
+ "group": [
904
+ "pt_benchmark",
905
+ "sparrow"
906
+ ],
907
+ "dataset_path": "UBC-NLP/sparrow",
908
+ "dataset_name": "sentiment-2016-mozetic-por",
909
+ "test_split": "validation",
910
+ "fewshot_split": "train",
911
+ "doc_to_text": "Texto: {{content}}\nPergunta: O sentimento do texto é Positivo, Neutro ou Negativo?\nResposta:",
912
+ "doc_to_target": "{{'Positivo' if label == 'Positive' else ('Negativo' if label == 'Negative' else 'Neutro')}}",
913
+ "description": "Abaixo contém o conteúdo de tweets de usuarios do Twitter em português, sua tarefa é classificar se o sentimento do texto é Positivo, Neutro ou Negativo. Responda apenas com uma das opções.\n\n",
914
+ "target_delimiter": " ",
915
+ "fewshot_delimiter": "\n\n",
916
+ "fewshot_config": {
917
+ "sampler": "first_n"
918
+ },
919
+ "num_fewshot": 25,
920
+ "metric_list": [
921
+ {
922
+ "metric": "f1_macro",
923
+ "aggregation": "f1_macro",
924
+ "higher_is_better": true
925
+ },
926
+ {
927
+ "metric": "acc",
928
+ "aggregation": "acc",
929
+ "higher_is_better": true
930
+ }
931
+ ],
932
+ "output_type": "generate_until",
933
+ "generation_kwargs": {
934
+ "max_gen_toks": 32,
935
+ "do_sample": false,
936
+ "temperature": 0.0,
937
+ "top_k": null,
938
+ "top_p": null,
939
+ "until": [
940
+ "\n\n"
941
+ ]
942
+ },
943
+ "repeats": 1,
944
+ "filter_list": [
945
+ {
946
+ "name": "all",
947
+ "filter": [
948
+ {
949
+ "function": "find_similar_label",
950
+ "labels": [
951
+ "Positivo",
952
+ "Neutro",
953
+ "Negativo"
954
+ ]
955
+ },
956
+ {
957
+ "function": "take_first"
958
+ }
959
+ ]
960
+ }
961
+ ],
962
+ "should_decontaminate": false,
963
+ "limit": 500,
964
+ "metadata": {
965
+ "version": 1.0
966
+ }
967
+ },
968
+ "sparrow_sentiment-2018-brum-por": {
969
+ "task": "sparrow_sentiment-2018-brum-por",
970
+ "task_alias": "sentiment-2018-brum-por",
971
+ "group": [
972
+ "pt_benchmark",
973
+ "sparrow"
974
+ ],
975
+ "dataset_path": "UBC-NLP/sparrow",
976
+ "dataset_name": "sentiment-2018-brum-por",
977
+ "test_split": "validation",
978
+ "fewshot_split": "train",
979
+ "doc_to_text": "Texto: {{content}}\nPergunta: O sentimento do texto é Positivo, Neutro ou Negativo?\nResposta:",
980
+ "doc_to_target": "{{'Positivo' if label == 'Positive' else ('Negativo' if label == 'Negative' else 'Neutro')}}",
981
+ "description": "Abaixo contém o conteúdo de tweets de usuarios do Twitter em português, sua tarefa é classificar se o sentimento do texto é Positivo, Neutro ou Negativo. Responda apenas com uma das opções.\n\n",
982
+ "target_delimiter": " ",
983
+ "fewshot_delimiter": "\n\n",
984
+ "fewshot_config": {
985
+ "sampler": "first_n"
986
+ },
987
+ "num_fewshot": 25,
988
+ "metric_list": [
989
+ {
990
+ "metric": "f1_macro",
991
+ "aggregation": "f1_macro",
992
+ "higher_is_better": true
993
+ },
994
+ {
995
+ "metric": "acc",
996
+ "aggregation": "acc",
997
+ "higher_is_better": true
998
+ }
999
+ ],
1000
+ "output_type": "generate_until",
1001
+ "generation_kwargs": {
1002
+ "max_gen_toks": 32,
1003
+ "do_sample": false,
1004
+ "temperature": 0.0,
1005
+ "top_k": null,
1006
+ "top_p": null,
1007
+ "until": [
1008
+ "\n\n"
1009
+ ]
1010
+ },
1011
+ "repeats": 1,
1012
+ "filter_list": [
1013
+ {
1014
+ "name": "all",
1015
+ "filter": [
1016
+ {
1017
+ "function": "find_similar_label",
1018
+ "labels": [
1019
+ "Positivo",
1020
+ "Neutro",
1021
+ "Negativo"
1022
+ ]
1023
+ },
1024
+ {
1025
+ "function": "take_first"
1026
+ }
1027
+ ]
1028
+ }
1029
+ ],
1030
+ "should_decontaminate": false,
1031
+ "limit": 500,
1032
+ "metadata": {
1033
+ "version": 1.0
1034
+ }
1035
+ }
1036
+ },
1037
+ "versions": {
1038
+ "assin2_rte": 1.0,
1039
+ "assin2_sts": 1.0,
1040
+ "bluex": 1.0,
1041
+ "enem_challenge": 1.0,
1042
+ "faquad_nli": 1.0,
1043
+ "oab_exams": 1.4,
1044
+ "sparrow_emotion-2021-cortiz-por": 1.0,
1045
+ "sparrow_hate-2019-fortuna-por": 1.0,
1046
+ "sparrow_sentiment-2016-mozetic-por": 1.0,
1047
+ "sparrow_sentiment-2018-brum-por": 1.0
1048
+ },
1049
+ "n-shot": {
1050
+ "assin2_rte": 15,
1051
+ "assin2_sts": 15,
1052
+ "bluex": 3,
1053
+ "enem_challenge": 3,
1054
+ "faquad_nli": 15,
1055
+ "oab_exams": 3,
1056
+ "sparrow_emotion-2021-cortiz-por": 25,
1057
+ "sparrow_hate-2019-fortuna-por": 25,
1058
+ "sparrow_sentiment-2016-mozetic-por": 25,
1059
+ "sparrow_sentiment-2018-brum-por": 25
1060
+ },
1061
+ "model_meta": {
1062
+ "truncated": 2,
1063
+ "non_truncated": 11887,
1064
+ "padded": 0,
1065
+ "non_padded": 11889,
1066
+ "fewshots_truncated": 2,
1067
+ "has_chat_template": false,
1068
+ "chat_type": null,
1069
+ "n_gpus": 1,
1070
+ "accelerate_num_process": null,
1071
+ "model_sha": "0de9394d54f8975e71838d309de1cb496c894ab9",
1072
+ "model_dtype": "torch.bfloat16",
1073
+ "model_memory_footprint": 1184087040,
1074
+ "model_num_parameters": 592043520,
1075
+ "model_is_loaded_in_4bit": false,
1076
+ "model_is_loaded_in_8bit": false,
1077
+ "model_is_quantized": null,
1078
+ "model_device": "cuda:0",
1079
+ "batch_size": 16,
1080
+ "max_length": 2048,
1081
+ "max_ctx_length": 2048,
1082
+ "max_gen_toks": 32
1083
+ },
1084
+ "task_model_meta": {
1085
+ "assin2_rte": {
1086
+ "sample_size": 2448,
1087
+ "truncated": 0,
1088
+ "non_truncated": 2448,
1089
+ "padded": 0,
1090
+ "non_padded": 2448,
1091
+ "fewshots_truncated": 0,
1092
+ "mean_seq_length": 928.1200980392157,
1093
+ "min_seq_length": 912,
1094
+ "max_seq_length": 976,
1095
+ "max_ctx_length": 2048,
1096
+ "max_gen_toks": 32,
1097
+ "mean_original_fewshots_size": 15.0,
1098
+ "mean_effective_fewshot_size": 15.0
1099
+ },
1100
+ "assin2_sts": {
1101
+ "sample_size": 2448,
1102
+ "truncated": 0,
1103
+ "non_truncated": 2448,
1104
+ "padded": 0,
1105
+ "non_padded": 2448,
1106
+ "fewshots_truncated": 0,
1107
+ "mean_seq_length": 1196.1200980392157,
1108
+ "min_seq_length": 1180,
1109
+ "max_seq_length": 1244,
1110
+ "max_ctx_length": 2048,
1111
+ "max_gen_toks": 32,
1112
+ "mean_original_fewshots_size": 15.0,
1113
+ "mean_effective_fewshot_size": 15.0
1114
+ },
1115
+ "bluex": {
1116
+ "sample_size": 719,
1117
+ "truncated": 0,
1118
+ "non_truncated": 719,
1119
+ "padded": 0,
1120
+ "non_padded": 719,
1121
+ "fewshots_truncated": 0,
1122
+ "mean_seq_length": 1164.269819193324,
1123
+ "min_seq_length": 890,
1124
+ "max_seq_length": 1696,
1125
+ "max_ctx_length": 2048,
1126
+ "max_gen_toks": 32,
1127
+ "mean_original_fewshots_size": 3.0,
1128
+ "mean_effective_fewshot_size": 3.0
1129
+ },
1130
+ "enem_challenge": {
1131
+ "sample_size": 1429,
1132
+ "truncated": 2,
1133
+ "non_truncated": 1427,
1134
+ "padded": 0,
1135
+ "non_padded": 1429,
1136
+ "fewshots_truncated": 2,
1137
+ "mean_seq_length": 1207.7872638208537,
1138
+ "min_seq_length": 1012,
1139
+ "max_seq_length": 2293,
1140
+ "max_ctx_length": 2048,
1141
+ "max_gen_toks": 32,
1142
+ "mean_original_fewshots_size": 3.0,
1143
+ "mean_effective_fewshot_size": 2.998600419874038
1144
+ },
1145
+ "faquad_nli": {
1146
+ "sample_size": 650,
1147
+ "truncated": 0,
1148
+ "non_truncated": 650,
1149
+ "padded": 0,
1150
+ "non_padded": 650,
1151
+ "fewshots_truncated": 0,
1152
+ "mean_seq_length": 1074.863076923077,
1153
+ "min_seq_length": 1040,
1154
+ "max_seq_length": 1144,
1155
+ "max_ctx_length": 2048,
1156
+ "max_gen_toks": 32,
1157
+ "mean_original_fewshots_size": 15.0,
1158
+ "mean_effective_fewshot_size": 15.0
1159
+ },
1160
+ "oab_exams": {
1161
+ "sample_size": 2195,
1162
+ "truncated": 0,
1163
+ "non_truncated": 2195,
1164
+ "padded": 0,
1165
+ "non_padded": 2195,
1166
+ "fewshots_truncated": 0,
1167
+ "mean_seq_length": 953.780410022779,
1168
+ "min_seq_length": 762,
1169
+ "max_seq_length": 1264,
1170
+ "max_ctx_length": 2048,
1171
+ "max_gen_toks": 32,
1172
+ "mean_original_fewshots_size": 3.0,
1173
+ "mean_effective_fewshot_size": 3.0
1174
+ },
1175
+ "sparrow_emotion-2021-cortiz-por": {
1176
+ "sample_size": 500,
1177
+ "truncated": 0,
1178
+ "non_truncated": 500,
1179
+ "padded": 0,
1180
+ "non_padded": 500,
1181
+ "fewshots_truncated": 0,
1182
+ "mean_seq_length": 1321.456,
1183
+ "min_seq_length": 1306,
1184
+ "max_seq_length": 1343,
1185
+ "max_ctx_length": 2048,
1186
+ "max_gen_toks": 32,
1187
+ "mean_original_fewshots_size": 25.0,
1188
+ "mean_effective_fewshot_size": 25.0
1189
+ },
1190
+ "sparrow_hate-2019-fortuna-por": {
1191
+ "sample_size": 500,
1192
+ "truncated": 0,
1193
+ "non_truncated": 500,
1194
+ "padded": 0,
1195
+ "non_padded": 500,
1196
+ "fewshots_truncated": 0,
1197
+ "mean_seq_length": 1305.35,
1198
+ "min_seq_length": 1286,
1199
+ "max_seq_length": 1338,
1200
+ "max_ctx_length": 2048,
1201
+ "max_gen_toks": 32,
1202
+ "mean_original_fewshots_size": 25.0,
1203
+ "mean_effective_fewshot_size": 25.0
1204
+ },
1205
+ "sparrow_sentiment-2016-mozetic-por": {
1206
+ "sample_size": 500,
1207
+ "truncated": 0,
1208
+ "non_truncated": 500,
1209
+ "padded": 0,
1210
+ "non_padded": 500,
1211
+ "fewshots_truncated": 0,
1212
+ "mean_seq_length": 1178.136,
1213
+ "min_seq_length": 1164,
1214
+ "max_seq_length": 1215,
1215
+ "max_ctx_length": 2048,
1216
+ "max_gen_toks": 32,
1217
+ "mean_original_fewshots_size": 25.0,
1218
+ "mean_effective_fewshot_size": 25.0
1219
+ },
1220
+ "sparrow_sentiment-2018-brum-por": {
1221
+ "sample_size": 500,
1222
+ "truncated": 0,
1223
+ "non_truncated": 500,
1224
+ "padded": 0,
1225
+ "non_padded": 500,
1226
+ "fewshots_truncated": 0,
1227
+ "mean_seq_length": 1291.334,
1228
+ "min_seq_length": 1277,
1229
+ "max_seq_length": 1322,
1230
+ "max_ctx_length": 2048,
1231
+ "max_gen_toks": 32,
1232
+ "mean_original_fewshots_size": 25.0,
1233
+ "mean_effective_fewshot_size": 25.0
1234
+ }
1235
+ },
1236
+ "config": {
1237
+ "model": "huggingface",
1238
+ "model_args": "pretrained=google/umt5-base,dtype=bfloat16,device=cuda:0,revision=main,trust_remote_code=True,starting_max_length=4096",
1239
+ "batch_size": "auto",
1240
+ "batch_sizes": [],
1241
+ "device": null,
1242
+ "use_cache": null,
1243
+ "limit": [
1244
+ null,
1245
+ null,
1246
+ null,
1247
+ null,
1248
+ null,
1249
+ null,
1250
+ 500.0,
1251
+ 500.0,
1252
+ 500.0,
1253
+ 500.0
1254
+ ],
1255
+ "bootstrap_iters": 0,
1256
+ "gen_kwargs": null
1257
+ },
1258
+ "git_hash": "15f86b5"
1259
+ }
google/umt5-base/results_2024-02-15T05-17-17.944550.json ADDED
@@ -0,0 +1,346 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config_general": {
3
+ "start_date": "2024-02-15T05-17-17.944550",
4
+ "start_time": 1707974238.5539687,
5
+ "end_time": 1708079048.7124906,
6
+ "total_evaluation_time_seconds": 104810.15852189064,
7
+ "has_chat_template": false,
8
+ "chat_type": null,
9
+ "n_gpus": 1,
10
+ "accelerate_num_process": null,
11
+ "model_sha": "0de9394d54f8975e71838d309de1cb496c894ab9",
12
+ "model_dtype": "bfloat16",
13
+ "model_memory_footprint": 1184087040,
14
+ "model_num_parameters": 592043520,
15
+ "model_is_loaded_in_4bit": false,
16
+ "model_is_loaded_in_8bit": false,
17
+ "model_is_quantized": null,
18
+ "model_device": "cuda:0",
19
+ "batch_size": 16,
20
+ "max_length": 2048,
21
+ "max_ctx_length": 2048,
22
+ "max_gen_toks": 32,
23
+ "model_name": "google/umt5-base",
24
+ "job_id": 206,
25
+ "model_id": "google/umt5-base_eval_request_False_bfloat16_Original",
26
+ "model_base_model": "",
27
+ "model_weight_type": "Original",
28
+ "model_revision": "main",
29
+ "model_private": false,
30
+ "model_type": "🟢 : pretrained",
31
+ "model_architectures": "UMT5ForConditionalGeneration",
32
+ "submitted_time": "2024-02-05T23:18:55Z",
33
+ "lm_eval_model_type": "huggingface",
34
+ "eval_version": "1.0.0"
35
+ },
36
+ "results": {
37
+ "all_grouped_average": 0.04776457359664892,
38
+ "all_grouped_npm": -0.38758303220535195,
39
+ "all_grouped": {
40
+ "enem_challenge": 0.02939118264520644,
41
+ "bluex": 0.022253129346314324,
42
+ "oab_exams": 0.08337129840546698,
43
+ "assin2_rte": 0.008049597487606372,
44
+ "assin2_sts": 0.11702145458274152,
45
+ "faquad_nli": 0.005177993527508091,
46
+ "sparrow_pt": 0.06908735918169881
47
+ },
48
+ "all": {
49
+ "harness|enem_challenge|enem_challenge|None|3": 0.02939118264520644,
50
+ "harness|bluex|bluex|None|3": 0.022253129346314324,
51
+ "harness|oab_exams|oab_exams|None|3": 0.08337129840546698,
52
+ "harness|assin2_rte|assin2_rte|None|15": 0.008049597487606372,
53
+ "harness|assin2_sts|assin2_sts|None|15": 0.11702145458274152,
54
+ "harness|faquad_nli|faquad_nli|None|15": 0.005177993527508091,
55
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por|500|25": 0.0,
56
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por|500|25": 0.24175824175824176,
57
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por|500|25": 0.0,
58
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por|500|25": 0.03459119496855346
59
+ },
60
+ "harness|enem_challenge|enem_challenge|None|3": {
61
+ "acc,all": 0.02939118264520644,
62
+ "acc,exam_id__2015": 0.03361344537815126,
63
+ "acc,exam_id__2012": 0.034482758620689655,
64
+ "acc,exam_id__2017": 0.02586206896551724,
65
+ "acc,exam_id__2009": 0.017391304347826087,
66
+ "acc,exam_id__2023": 0.037037037037037035,
67
+ "acc,exam_id__2016_2": 0.024390243902439025,
68
+ "acc,exam_id__2010": 0.05128205128205128,
69
+ "acc,exam_id__2014": 0.01834862385321101,
70
+ "acc,exam_id__2013": 0.018518518518518517,
71
+ "acc,exam_id__2022": 0.05263157894736842,
72
+ "acc,exam_id__2011": 0.02564102564102564,
73
+ "acc,exam_id__2016": 0.008264462809917356,
74
+ "main_score": 0.02939118264520644
75
+ },
76
+ "harness|bluex|bluex|None|3": {
77
+ "acc,all": 0.022253129346314324,
78
+ "acc,exam_id__UNICAMP_2022": 0.0,
79
+ "acc,exam_id__UNICAMP_2019": 0.06,
80
+ "acc,exam_id__UNICAMP_2023": 0.023255813953488372,
81
+ "acc,exam_id__UNICAMP_2018": 0.0,
82
+ "acc,exam_id__UNICAMP_2020": 0.05454545454545454,
83
+ "acc,exam_id__USP_2024": 0.0,
84
+ "acc,exam_id__UNICAMP_2021_2": 0.0,
85
+ "acc,exam_id__USP_2022": 0.0,
86
+ "acc,exam_id__USP_2019": 0.0,
87
+ "acc,exam_id__USP_2021": 0.019230769230769232,
88
+ "acc,exam_id__USP_2023": 0.09090909090909091,
89
+ "acc,exam_id__UNICAMP_2021_1": 0.021739130434782608,
90
+ "acc,exam_id__UNICAMP_2024": 0.022222222222222223,
91
+ "acc,exam_id__USP_2020": 0.017857142857142856,
92
+ "acc,exam_id__USP_2018": 0.018518518518518517,
93
+ "main_score": 0.022253129346314324
94
+ },
95
+ "harness|oab_exams|oab_exams|None|3": {
96
+ "acc,all": 0.08337129840546698,
97
+ "acc,exam_id__2012-06a": 0.1,
98
+ "acc,exam_id__2015-17": 0.10256410256410256,
99
+ "acc,exam_id__2012-06": 0.0625,
100
+ "acc,exam_id__2014-15": 0.07692307692307693,
101
+ "acc,exam_id__2016-20": 0.075,
102
+ "acc,exam_id__2013-10": 0.0875,
103
+ "acc,exam_id__2012-07": 0.025,
104
+ "acc,exam_id__2011-03": 0.10101010101010101,
105
+ "acc,exam_id__2011-05": 0.1125,
106
+ "acc,exam_id__2016-19": 0.038461538461538464,
107
+ "acc,exam_id__2017-23": 0.1,
108
+ "acc,exam_id__2017-22": 0.0625,
109
+ "acc,exam_id__2018-25": 0.1,
110
+ "acc,exam_id__2014-13": 0.05,
111
+ "acc,exam_id__2017-24": 0.0625,
112
+ "acc,exam_id__2010-01": 0.12941176470588237,
113
+ "acc,exam_id__2014-14": 0.125,
114
+ "acc,exam_id__2011-04": 0.0625,
115
+ "acc,exam_id__2013-12": 0.0625,
116
+ "acc,exam_id__2015-16": 0.1,
117
+ "acc,exam_id__2016-21": 0.075,
118
+ "acc,exam_id__2013-11": 0.075,
119
+ "acc,exam_id__2012-08": 0.025,
120
+ "acc,exam_id__2016-20a": 0.1375,
121
+ "acc,exam_id__2012-09": 0.07792207792207792,
122
+ "acc,exam_id__2010-02": 0.11,
123
+ "acc,exam_id__2015-18": 0.1,
124
+ "main_score": 0.08337129840546698
125
+ },
126
+ "harness|assin2_rte|assin2_rte|None|15": {
127
+ "f1_macro,all": 0.008049597487606372,
128
+ "acc,all": 0.006127450980392157,
129
+ "main_score": 0.008049597487606372
130
+ },
131
+ "harness|assin2_sts|assin2_sts|None|15": {
132
+ "pearson,all": 0.11702145458274152,
133
+ "mse,all": 4.536915664667917,
134
+ "main_score": 0.11702145458274152
135
+ },
136
+ "harness|faquad_nli|faquad_nli|None|15": {
137
+ "f1_macro,all": 0.005177993527508091,
138
+ "acc,all": 0.006153846153846154,
139
+ "main_score": 0.005177993527508091
140
+ },
141
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por|500|25": {
142
+ "f1_macro,all": 0.0,
143
+ "acc,all": 0.0,
144
+ "main_score": 0.0
145
+ },
146
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por|500|25": {
147
+ "f1_macro,all": 0.24175824175824176,
148
+ "acc,all": 0.528,
149
+ "main_score": 0.24175824175824176
150
+ },
151
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por|500|25": {
152
+ "f1_macro,all": 0.0,
153
+ "acc,all": 0.0,
154
+ "main_score": 0.0
155
+ },
156
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por|500|25": {
157
+ "f1_macro,all": 0.03459119496855346,
158
+ "acc,all": 0.022,
159
+ "main_score": 0.03459119496855346
160
+ }
161
+ },
162
+ "config_tasks": {
163
+ "harness|enem_challenge|enem_challenge": "LM Harness task",
164
+ "harness|bluex|bluex": "LM Harness task",
165
+ "harness|oab_exams|oab_exams": "LM Harness task",
166
+ "harness|assin2_rte|assin2_rte": "LM Harness task",
167
+ "harness|assin2_sts|assin2_sts": "LM Harness task",
168
+ "harness|faquad_nli|faquad_nli": "LM Harness task",
169
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por": "LM Harness task",
170
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por": "LM Harness task",
171
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por": "LM Harness task",
172
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por": "LM Harness task"
173
+ },
174
+ "versions": {
175
+ "all": 0,
176
+ "harness|enem_challenge|enem_challenge": 1.0,
177
+ "harness|bluex|bluex": 1.0,
178
+ "harness|oab_exams|oab_exams": 1.4,
179
+ "harness|assin2_rte|assin2_rte": 1.0,
180
+ "harness|assin2_sts|assin2_sts": 1.0,
181
+ "harness|faquad_nli|faquad_nli": 1.0,
182
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por": 1.0,
183
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por": 1.0,
184
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por": 1.0,
185
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por": 1.0
186
+ },
187
+ "summary_tasks": {
188
+ "harness|enem_challenge|enem_challenge|None|3": {
189
+ "sample_size": 1429,
190
+ "truncated": 2,
191
+ "non_truncated": 1427,
192
+ "padded": 0,
193
+ "non_padded": 1429,
194
+ "fewshots_truncated": 2,
195
+ "mean_seq_length": 1207.7872638208537,
196
+ "min_seq_length": 1012,
197
+ "max_seq_length": 2293,
198
+ "max_ctx_length": 2048,
199
+ "max_gen_toks": 32,
200
+ "mean_original_fewshots_size": 3.0,
201
+ "mean_effective_fewshot_size": 2.998600419874038
202
+ },
203
+ "harness|bluex|bluex|None|3": {
204
+ "sample_size": 719,
205
+ "truncated": 0,
206
+ "non_truncated": 719,
207
+ "padded": 0,
208
+ "non_padded": 719,
209
+ "fewshots_truncated": 0,
210
+ "mean_seq_length": 1164.269819193324,
211
+ "min_seq_length": 890,
212
+ "max_seq_length": 1696,
213
+ "max_ctx_length": 2048,
214
+ "max_gen_toks": 32,
215
+ "mean_original_fewshots_size": 3.0,
216
+ "mean_effective_fewshot_size": 3.0
217
+ },
218
+ "harness|oab_exams|oab_exams|None|3": {
219
+ "sample_size": 2195,
220
+ "truncated": 0,
221
+ "non_truncated": 2195,
222
+ "padded": 0,
223
+ "non_padded": 2195,
224
+ "fewshots_truncated": 0,
225
+ "mean_seq_length": 953.780410022779,
226
+ "min_seq_length": 762,
227
+ "max_seq_length": 1264,
228
+ "max_ctx_length": 2048,
229
+ "max_gen_toks": 32,
230
+ "mean_original_fewshots_size": 3.0,
231
+ "mean_effective_fewshot_size": 3.0
232
+ },
233
+ "harness|assin2_rte|assin2_rte|None|15": {
234
+ "sample_size": 2448,
235
+ "truncated": 0,
236
+ "non_truncated": 2448,
237
+ "padded": 0,
238
+ "non_padded": 2448,
239
+ "fewshots_truncated": 0,
240
+ "mean_seq_length": 928.1200980392157,
241
+ "min_seq_length": 912,
242
+ "max_seq_length": 976,
243
+ "max_ctx_length": 2048,
244
+ "max_gen_toks": 32,
245
+ "mean_original_fewshots_size": 15.0,
246
+ "mean_effective_fewshot_size": 15.0
247
+ },
248
+ "harness|assin2_sts|assin2_sts|None|15": {
249
+ "sample_size": 2448,
250
+ "truncated": 0,
251
+ "non_truncated": 2448,
252
+ "padded": 0,
253
+ "non_padded": 2448,
254
+ "fewshots_truncated": 0,
255
+ "mean_seq_length": 1196.1200980392157,
256
+ "min_seq_length": 1180,
257
+ "max_seq_length": 1244,
258
+ "max_ctx_length": 2048,
259
+ "max_gen_toks": 32,
260
+ "mean_original_fewshots_size": 15.0,
261
+ "mean_effective_fewshot_size": 15.0
262
+ },
263
+ "harness|faquad_nli|faquad_nli|None|15": {
264
+ "sample_size": 650,
265
+ "truncated": 0,
266
+ "non_truncated": 650,
267
+ "padded": 0,
268
+ "non_padded": 650,
269
+ "fewshots_truncated": 0,
270
+ "mean_seq_length": 1074.863076923077,
271
+ "min_seq_length": 1040,
272
+ "max_seq_length": 1144,
273
+ "max_ctx_length": 2048,
274
+ "max_gen_toks": 32,
275
+ "mean_original_fewshots_size": 15.0,
276
+ "mean_effective_fewshot_size": 15.0
277
+ },
278
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por|500|25": {
279
+ "sample_size": 500,
280
+ "truncated": 0,
281
+ "non_truncated": 500,
282
+ "padded": 0,
283
+ "non_padded": 500,
284
+ "fewshots_truncated": 0,
285
+ "mean_seq_length": 1321.456,
286
+ "min_seq_length": 1306,
287
+ "max_seq_length": 1343,
288
+ "max_ctx_length": 2048,
289
+ "max_gen_toks": 32,
290
+ "mean_original_fewshots_size": 25.0,
291
+ "mean_effective_fewshot_size": 25.0
292
+ },
293
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por|500|25": {
294
+ "sample_size": 500,
295
+ "truncated": 0,
296
+ "non_truncated": 500,
297
+ "padded": 0,
298
+ "non_padded": 500,
299
+ "fewshots_truncated": 0,
300
+ "mean_seq_length": 1305.35,
301
+ "min_seq_length": 1286,
302
+ "max_seq_length": 1338,
303
+ "max_ctx_length": 2048,
304
+ "max_gen_toks": 32,
305
+ "mean_original_fewshots_size": 25.0,
306
+ "mean_effective_fewshot_size": 25.0
307
+ },
308
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por|500|25": {
309
+ "sample_size": 500,
310
+ "truncated": 0,
311
+ "non_truncated": 500,
312
+ "padded": 0,
313
+ "non_padded": 500,
314
+ "fewshots_truncated": 0,
315
+ "mean_seq_length": 1178.136,
316
+ "min_seq_length": 1164,
317
+ "max_seq_length": 1215,
318
+ "max_ctx_length": 2048,
319
+ "max_gen_toks": 32,
320
+ "mean_original_fewshots_size": 25.0,
321
+ "mean_effective_fewshot_size": 25.0
322
+ },
323
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por|500|25": {
324
+ "sample_size": 500,
325
+ "truncated": 0,
326
+ "non_truncated": 500,
327
+ "padded": 0,
328
+ "non_padded": 500,
329
+ "fewshots_truncated": 0,
330
+ "mean_seq_length": 1291.334,
331
+ "min_seq_length": 1277,
332
+ "max_seq_length": 1322,
333
+ "max_ctx_length": 2048,
334
+ "max_gen_toks": 32,
335
+ "mean_original_fewshots_size": 25.0,
336
+ "mean_effective_fewshot_size": 25.0
337
+ }
338
+ },
339
+ "summary_general": {
340
+ "truncated": 2,
341
+ "non_truncated": 11887,
342
+ "padded": 0,
343
+ "non_padded": 11889,
344
+ "fewshots_truncated": 2
345
+ }
346
+ }