eduagarcia commited on
Commit
862d0d4
1 Parent(s): 4e9ef69

Uploading raw results for mistralai/Mixtral-8x7B-v0.1

Browse files
Files changed (13) hide show
  1. .gitattributes +5 -0
  2. mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_rte.jsonl +3 -0
  3. mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_sts.jsonl +3 -0
  4. mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl +3 -0
  5. mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl +3 -0
  6. mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_faquad_nli.jsonl +0 -0
  7. mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl +3 -0
  8. mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_emotion-2021-cortiz-por.jsonl +0 -0
  9. mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_hate-2019-fortuna-por.jsonl +0 -0
  10. mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_sentiment-2016-mozetic-por.jsonl +0 -0
  11. mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_sentiment-2018-brum-por.jsonl +0 -0
  12. mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/results.json +1255 -0
  13. mistralai/Mixtral-8x7B-v0.1/results_2024-02-10T03-25-53.024286.json +346 -0
.gitattributes CHANGED
@@ -390,3 +390,8 @@ Qwen/Qwen-7B/raw_2024-02-09T20-57-09.165430/pretrained__Qwen__Qwen-7B,dtype__bfl
390
  Qwen/Qwen-7B/raw_2024-02-09T20-57-09.165430/pretrained__Qwen__Qwen-7B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl filter=lfs diff=lfs merge=lfs -text
391
  Qwen/Qwen-7B/raw_2024-02-09T20-57-09.165430/pretrained__Qwen__Qwen-7B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl filter=lfs diff=lfs merge=lfs -text
392
  Qwen/Qwen-7B/raw_2024-02-09T20-57-09.165430/pretrained__Qwen__Qwen-7B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
390
  Qwen/Qwen-7B/raw_2024-02-09T20-57-09.165430/pretrained__Qwen__Qwen-7B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl filter=lfs diff=lfs merge=lfs -text
391
  Qwen/Qwen-7B/raw_2024-02-09T20-57-09.165430/pretrained__Qwen__Qwen-7B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl filter=lfs diff=lfs merge=lfs -text
392
  Qwen/Qwen-7B/raw_2024-02-09T20-57-09.165430/pretrained__Qwen__Qwen-7B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl filter=lfs diff=lfs merge=lfs -text
393
+ mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_rte.jsonl filter=lfs diff=lfs merge=lfs -text
394
+ mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_sts.jsonl filter=lfs diff=lfs merge=lfs -text
395
+ mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl filter=lfs diff=lfs merge=lfs -text
396
+ mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl filter=lfs diff=lfs merge=lfs -text
397
+ mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl filter=lfs diff=lfs merge=lfs -text
mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_rte.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fedd3c9b131c12de52f518e6b5fde8acec454cccb488afa504f950d6cc5abf9e
3
+ size 29358689
mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_assin2_sts.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5ed542f1626a3cadfe3a588efbb4a1a9d10fae9d193df444fbfb2d1b3ef86e8
3
+ size 30839037
mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_bluex.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a119845af402a3774d1b6713fecfa97fc27acacbb4cbc2ab367e42f89ba7b8b
3
+ size 10578534
mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_enem_challenge.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08fe562e75121f7f736f34a630fbef20083b89e70bb167a905b89fcf93081b43
3
+ size 23295140
mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_faquad_nli.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_oab_exams.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77c228f19411b1b5c8f27efb21aacc675365ed929ba013a16d44cfb400a469c1
3
+ size 32556330
mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_emotion-2021-cortiz-por.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_hate-2019-fortuna-por.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_sentiment-2016-mozetic-por.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/pretrained__mistralai__Mixtral-8x7B-v0.1,dtype__bfloat16,parallelize__True,revision__main,trust_remote_code__True,starting_max_length__4096_sparrow_sentiment-2018-brum-por.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
mistralai/Mixtral-8x7B-v0.1/raw_2024-02-10T03-25-53.024286/results.json ADDED
@@ -0,0 +1,1255 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "assin2_rte": {
4
+ "f1_macro,all": 0.6646069030555132,
5
+ "acc,all": 0.6956699346405228,
6
+ "alias": "assin2_rte"
7
+ },
8
+ "assin2_sts": {
9
+ "pearson,all": 0.49872071984375066,
10
+ "mse,all": 1.9470588235294124,
11
+ "alias": "assin2_sts"
12
+ },
13
+ "bluex": {
14
+ "acc,all": 0.6369958275382476,
15
+ "acc,exam_id__USP_2019": 0.6,
16
+ "acc,exam_id__USP_2018": 0.5740740740740741,
17
+ "acc,exam_id__UNICAMP_2018": 0.5555555555555556,
18
+ "acc,exam_id__UNICAMP_2023": 0.7674418604651163,
19
+ "acc,exam_id__USP_2024": 0.7073170731707317,
20
+ "acc,exam_id__UNICAMP_2021_1": 0.5869565217391305,
21
+ "acc,exam_id__UNICAMP_2019": 0.68,
22
+ "acc,exam_id__USP_2020": 0.6607142857142857,
23
+ "acc,exam_id__UNICAMP_2020": 0.6545454545454545,
24
+ "acc,exam_id__USP_2022": 0.6122448979591837,
25
+ "acc,exam_id__USP_2023": 0.75,
26
+ "acc,exam_id__UNICAMP_2021_2": 0.5882352941176471,
27
+ "acc,exam_id__UNICAMP_2024": 0.5555555555555556,
28
+ "acc,exam_id__USP_2021": 0.6153846153846154,
29
+ "acc,exam_id__UNICAMP_2022": 0.6923076923076923,
30
+ "alias": "bluex"
31
+ },
32
+ "enem_challenge": {
33
+ "alias": "enem",
34
+ "acc,all": 0.7249825052484254,
35
+ "acc,exam_id__2015": 0.773109243697479,
36
+ "acc,exam_id__2009": 0.7217391304347827,
37
+ "acc,exam_id__2014": 0.7431192660550459,
38
+ "acc,exam_id__2010": 0.7008547008547008,
39
+ "acc,exam_id__2011": 0.7863247863247863,
40
+ "acc,exam_id__2023": 0.725925925925926,
41
+ "acc,exam_id__2016_2": 0.7317073170731707,
42
+ "acc,exam_id__2016": 0.7107438016528925,
43
+ "acc,exam_id__2013": 0.75,
44
+ "acc,exam_id__2022": 0.6466165413533834,
45
+ "acc,exam_id__2017": 0.6982758620689655,
46
+ "acc,exam_id__2012": 0.7241379310344828
47
+ },
48
+ "faquad_nli": {
49
+ "f1_macro,all": 0.5026086956521739,
50
+ "acc,all": 0.796923076923077,
51
+ "alias": "faquad_nli"
52
+ },
53
+ "oab_exams": {
54
+ "acc,all": 0.5553530751708429,
55
+ "acc,exam_id__2016-20": 0.5875,
56
+ "acc,exam_id__2016-20a": 0.45,
57
+ "acc,exam_id__2017-22": 0.5875,
58
+ "acc,exam_id__2014-14": 0.6,
59
+ "acc,exam_id__2010-01": 0.49411764705882355,
60
+ "acc,exam_id__2017-23": 0.5125,
61
+ "acc,exam_id__2014-15": 0.717948717948718,
62
+ "acc,exam_id__2011-04": 0.525,
63
+ "acc,exam_id__2015-17": 0.6923076923076923,
64
+ "acc,exam_id__2013-10": 0.5625,
65
+ "acc,exam_id__2012-07": 0.5,
66
+ "acc,exam_id__2011-03": 0.5151515151515151,
67
+ "acc,exam_id__2012-06a": 0.575,
68
+ "acc,exam_id__2012-09": 0.45454545454545453,
69
+ "acc,exam_id__2014-13": 0.525,
70
+ "acc,exam_id__2015-18": 0.65,
71
+ "acc,exam_id__2011-05": 0.5625,
72
+ "acc,exam_id__2012-08": 0.5625,
73
+ "acc,exam_id__2018-25": 0.5375,
74
+ "acc,exam_id__2016-21": 0.5,
75
+ "acc,exam_id__2013-11": 0.5625,
76
+ "acc,exam_id__2010-02": 0.6,
77
+ "acc,exam_id__2015-16": 0.5875,
78
+ "acc,exam_id__2013-12": 0.55,
79
+ "acc,exam_id__2016-19": 0.5769230769230769,
80
+ "acc,exam_id__2017-24": 0.4625,
81
+ "acc,exam_id__2012-06": 0.55,
82
+ "alias": "oab_exams"
83
+ },
84
+ "sparrow_emotion-2021-cortiz-por": {
85
+ "alias": "emotion-2021-cortiz-por",
86
+ "f1_macro,all": 0.10240832053715039,
87
+ "acc,all": 0.148
88
+ },
89
+ "sparrow_hate-2019-fortuna-por": {
90
+ "alias": "hate-2019-fortuna-por",
91
+ "f1_macro,all": 0.4876514802883878,
92
+ "acc,all": 0.666
93
+ },
94
+ "sparrow_sentiment-2016-mozetic-por": {
95
+ "alias": "sentiment-2016-mozetic-por",
96
+ "f1_macro,all": 0.47227979759408134,
97
+ "acc,all": 0.516
98
+ },
99
+ "sparrow_sentiment-2018-brum-por": {
100
+ "alias": "sentiment-2018-brum-por",
101
+ "f1_macro,all": 0.4124146819549119,
102
+ "acc,all": 0.414
103
+ }
104
+ },
105
+ "configs": {
106
+ "assin2_rte": {
107
+ "task": "assin2_rte",
108
+ "group": [
109
+ "pt_benchmark",
110
+ "assin2"
111
+ ],
112
+ "dataset_path": "assin2",
113
+ "test_split": "test",
114
+ "fewshot_split": "train",
115
+ "doc_to_text": "Premissa: {{premise}}\nHipótese: {{hypothesis}}\nPergunta: A hipótese pode ser inferida pela premissa?\nResposta:",
116
+ "doc_to_target": "{{['Não', 'Sim'][entailment_judgment]}}",
117
+ "description": "Abaixo contém pares de premissa e hipótese, para cada par você deve julgar se a hipótese pode ser inferida a partir da premissa, responda apenas com Sim ou Não.\n\n",
118
+ "target_delimiter": " ",
119
+ "fewshot_delimiter": "\n\n",
120
+ "fewshot_config": {
121
+ "sampler": "id_sampler",
122
+ "sampler_config": {
123
+ "id_list": [
124
+ 1,
125
+ 3251,
126
+ 2,
127
+ 3252,
128
+ 3,
129
+ 4,
130
+ 5,
131
+ 6,
132
+ 3253,
133
+ 7,
134
+ 3254,
135
+ 3255,
136
+ 3256,
137
+ 8,
138
+ 9,
139
+ 10,
140
+ 3257,
141
+ 11,
142
+ 3258,
143
+ 12,
144
+ 13,
145
+ 14,
146
+ 15,
147
+ 3259,
148
+ 3260,
149
+ 3261,
150
+ 3262,
151
+ 3263,
152
+ 16,
153
+ 17,
154
+ 3264,
155
+ 18,
156
+ 3265,
157
+ 3266,
158
+ 3267,
159
+ 19,
160
+ 20,
161
+ 3268,
162
+ 3269,
163
+ 21,
164
+ 3270,
165
+ 3271,
166
+ 22,
167
+ 3272,
168
+ 3273,
169
+ 23,
170
+ 3274,
171
+ 24,
172
+ 25,
173
+ 3275
174
+ ],
175
+ "id_column": "sentence_pair_id"
176
+ }
177
+ },
178
+ "num_fewshot": 15,
179
+ "metric_list": [
180
+ {
181
+ "metric": "f1_macro",
182
+ "aggregation": "f1_macro",
183
+ "higher_is_better": true
184
+ },
185
+ {
186
+ "metric": "acc",
187
+ "aggregation": "acc",
188
+ "higher_is_better": true
189
+ }
190
+ ],
191
+ "output_type": "generate_until",
192
+ "generation_kwargs": {
193
+ "max_gen_toks": 32,
194
+ "do_sample": false,
195
+ "temperature": 0.0,
196
+ "top_k": null,
197
+ "top_p": null,
198
+ "until": [
199
+ "\n\n"
200
+ ]
201
+ },
202
+ "repeats": 1,
203
+ "filter_list": [
204
+ {
205
+ "name": "all",
206
+ "filter": [
207
+ {
208
+ "function": "find_similar_label",
209
+ "labels": [
210
+ "Sim",
211
+ "Não"
212
+ ]
213
+ },
214
+ {
215
+ "function": "take_first"
216
+ }
217
+ ]
218
+ }
219
+ ],
220
+ "should_decontaminate": false,
221
+ "metadata": {
222
+ "version": 1.0
223
+ }
224
+ },
225
+ "assin2_sts": {
226
+ "task": "assin2_sts",
227
+ "group": [
228
+ "pt_benchmark",
229
+ "assin2"
230
+ ],
231
+ "dataset_path": "assin2",
232
+ "test_split": "test",
233
+ "fewshot_split": "train",
234
+ "doc_to_text": "Frase 1: {{premise}}\nFrase 2: {{hypothesis}}\nPergunta: Qual o grau de similaridade entre as duas frases de 1,0 a 5,0?\nResposta:",
235
+ "doc_to_target": "<function assin2_float_to_pt_str at 0x7f71b0035800>",
236
+ "description": "Abaixo contém pares de frases, para cada par você deve julgar o grau de similaridade de 1,0 a 5,0, responda apenas com o número.\n\n",
237
+ "target_delimiter": " ",
238
+ "fewshot_delimiter": "\n\n",
239
+ "fewshot_config": {
240
+ "sampler": "id_sampler",
241
+ "sampler_config": {
242
+ "id_list": [
243
+ 1,
244
+ 3251,
245
+ 2,
246
+ 3252,
247
+ 3,
248
+ 4,
249
+ 5,
250
+ 6,
251
+ 3253,
252
+ 7,
253
+ 3254,
254
+ 3255,
255
+ 3256,
256
+ 8,
257
+ 9,
258
+ 10,
259
+ 3257,
260
+ 11,
261
+ 3258,
262
+ 12,
263
+ 13,
264
+ 14,
265
+ 15,
266
+ 3259,
267
+ 3260,
268
+ 3261,
269
+ 3262,
270
+ 3263,
271
+ 16,
272
+ 17,
273
+ 3264,
274
+ 18,
275
+ 3265,
276
+ 3266,
277
+ 3267,
278
+ 19,
279
+ 20,
280
+ 3268,
281
+ 3269,
282
+ 21,
283
+ 3270,
284
+ 3271,
285
+ 22,
286
+ 3272,
287
+ 3273,
288
+ 23,
289
+ 3274,
290
+ 24,
291
+ 25,
292
+ 3275
293
+ ],
294
+ "id_column": "sentence_pair_id"
295
+ }
296
+ },
297
+ "num_fewshot": 15,
298
+ "metric_list": [
299
+ {
300
+ "metric": "pearson",
301
+ "aggregation": "pearsonr",
302
+ "higher_is_better": true
303
+ },
304
+ {
305
+ "metric": "mse",
306
+ "aggregation": "mean_squared_error",
307
+ "higher_is_better": false
308
+ }
309
+ ],
310
+ "output_type": "generate_until",
311
+ "generation_kwargs": {
312
+ "max_gen_toks": 32,
313
+ "do_sample": false,
314
+ "temperature": 0.0,
315
+ "top_k": null,
316
+ "top_p": null,
317
+ "until": [
318
+ "\n\n"
319
+ ]
320
+ },
321
+ "repeats": 1,
322
+ "filter_list": [
323
+ {
324
+ "name": "all",
325
+ "filter": [
326
+ {
327
+ "function": "number_filter",
328
+ "type": "float",
329
+ "range_min": 1.0,
330
+ "range_max": 5.0,
331
+ "on_outside_range": "clip",
332
+ "fallback": 5.0
333
+ },
334
+ {
335
+ "function": "take_first"
336
+ }
337
+ ]
338
+ }
339
+ ],
340
+ "should_decontaminate": false,
341
+ "metadata": {
342
+ "version": 1.0
343
+ }
344
+ },
345
+ "bluex": {
346
+ "task": "bluex",
347
+ "group": [
348
+ "pt_benchmark",
349
+ "vestibular"
350
+ ],
351
+ "dataset_path": "eduagarcia-temp/BLUEX_without_images",
352
+ "test_split": "train",
353
+ "fewshot_split": "train",
354
+ "doc_to_text": "<function enem_doc_to_text at 0x7f71b00351c0>",
355
+ "doc_to_target": "{{answerKey}}",
356
+ "description": "As perguntas a seguir são questões de multipla escolha de provas de vestibular de Universidades Brasileiras, reponda apenas com as letras A, B, C, D ou E.\n\n",
357
+ "target_delimiter": " ",
358
+ "fewshot_delimiter": "\n\n",
359
+ "fewshot_config": {
360
+ "sampler": "id_sampler",
361
+ "sampler_config": {
362
+ "id_list": [
363
+ "USP_2018_3",
364
+ "UNICAMP_2018_2",
365
+ "USP_2018_35",
366
+ "UNICAMP_2018_16",
367
+ "USP_2018_89"
368
+ ],
369
+ "id_column": "id",
370
+ "exclude_from_task": true
371
+ }
372
+ },
373
+ "num_fewshot": 3,
374
+ "metric_list": [
375
+ {
376
+ "metric": "acc",
377
+ "aggregation": "acc",
378
+ "higher_is_better": true
379
+ }
380
+ ],
381
+ "output_type": "generate_until",
382
+ "generation_kwargs": {
383
+ "max_gen_toks": 32,
384
+ "do_sample": false,
385
+ "temperature": 0.0,
386
+ "top_k": null,
387
+ "top_p": null,
388
+ "until": [
389
+ "\n\n"
390
+ ]
391
+ },
392
+ "repeats": 1,
393
+ "filter_list": [
394
+ {
395
+ "name": "all",
396
+ "filter": [
397
+ {
398
+ "function": "normalize_spaces"
399
+ },
400
+ {
401
+ "function": "remove_accents"
402
+ },
403
+ {
404
+ "function": "find_choices",
405
+ "choices": [
406
+ "A",
407
+ "B",
408
+ "C",
409
+ "D",
410
+ "E"
411
+ ],
412
+ "regex_patterns": [
413
+ "(?:[Ll]etra|[Aa]lternativa|[Rr]esposta|[Rr]esposta [Cc]orreta|[Rr]esposta[Cc]orreta e|[Oo]pcao):? ([ABCDE])\\b",
414
+ "\\b([ABCDE])\\.",
415
+ "\\b([ABCDE]) ?[.):-]",
416
+ "\\b([ABCDE])$",
417
+ "\\b([ABCDE])\\b"
418
+ ]
419
+ },
420
+ {
421
+ "function": "take_first"
422
+ }
423
+ ],
424
+ "group_by": {
425
+ "column": "exam_id"
426
+ }
427
+ }
428
+ ],
429
+ "should_decontaminate": true,
430
+ "doc_to_decontamination_query": "<function enem_doc_to_text at 0x7f71b0035440>",
431
+ "metadata": {
432
+ "version": 1.0
433
+ }
434
+ },
435
+ "enem_challenge": {
436
+ "task": "enem_challenge",
437
+ "task_alias": "enem",
438
+ "group": [
439
+ "pt_benchmark",
440
+ "vestibular"
441
+ ],
442
+ "dataset_path": "eduagarcia/enem_challenge",
443
+ "test_split": "train",
444
+ "fewshot_split": "train",
445
+ "doc_to_text": "<function enem_doc_to_text at 0x7f71b00359e0>",
446
+ "doc_to_target": "{{answerKey}}",
447
+ "description": "As perguntas a seguir são questões de multipla escolha do Exame Nacional do Ensino Médio (ENEM), reponda apenas com as letras A, B, C, D ou E.\n\n",
448
+ "target_delimiter": " ",
449
+ "fewshot_delimiter": "\n\n",
450
+ "fewshot_config": {
451
+ "sampler": "id_sampler",
452
+ "sampler_config": {
453
+ "id_list": [
454
+ "2022_21",
455
+ "2022_88",
456
+ "2022_143"
457
+ ],
458
+ "id_column": "id",
459
+ "exclude_from_task": true
460
+ }
461
+ },
462
+ "num_fewshot": 3,
463
+ "metric_list": [
464
+ {
465
+ "metric": "acc",
466
+ "aggregation": "acc",
467
+ "higher_is_better": true
468
+ }
469
+ ],
470
+ "output_type": "generate_until",
471
+ "generation_kwargs": {
472
+ "max_gen_toks": 32,
473
+ "do_sample": false,
474
+ "temperature": 0.0,
475
+ "top_k": null,
476
+ "top_p": null,
477
+ "until": [
478
+ "\n\n"
479
+ ]
480
+ },
481
+ "repeats": 1,
482
+ "filter_list": [
483
+ {
484
+ "name": "all",
485
+ "filter": [
486
+ {
487
+ "function": "normalize_spaces"
488
+ },
489
+ {
490
+ "function": "remove_accents"
491
+ },
492
+ {
493
+ "function": "find_choices",
494
+ "choices": [
495
+ "A",
496
+ "B",
497
+ "C",
498
+ "D",
499
+ "E"
500
+ ],
501
+ "regex_patterns": [
502
+ "(?:[Ll]etra|[Aa]lternativa|[Rr]esposta|[Rr]esposta [Cc]orreta|[Rr]esposta[Cc]orreta e|[Oo]pcao):? ([ABCDE])\\b",
503
+ "\\b([ABCDE])\\.",
504
+ "\\b([ABCDE]) ?[.):-]",
505
+ "\\b([ABCDE])$",
506
+ "\\b([ABCDE])\\b"
507
+ ]
508
+ },
509
+ {
510
+ "function": "take_first"
511
+ }
512
+ ],
513
+ "group_by": {
514
+ "column": "exam_id"
515
+ }
516
+ }
517
+ ],
518
+ "should_decontaminate": true,
519
+ "doc_to_decontamination_query": "<function enem_doc_to_text at 0x7f71b0035c60>",
520
+ "metadata": {
521
+ "version": 1.0
522
+ }
523
+ },
524
+ "faquad_nli": {
525
+ "task": "faquad_nli",
526
+ "group": [
527
+ "pt_benchmark"
528
+ ],
529
+ "dataset_path": "ruanchaves/faquad-nli",
530
+ "test_split": "test",
531
+ "fewshot_split": "train",
532
+ "doc_to_text": "Pergunta: {{question}}\nResposta: {{answer}}\nA resposta satisfaz a pergunta? Sim ou Não?",
533
+ "doc_to_target": "{{['Não', 'Sim'][label]}}",
534
+ "description": "Abaixo contém pares de pergunta e reposta, para cada par você deve julgar resposta responde a pergunta de maneira satisfatória e aparenta estar correta, escreva apenas Sim ou Não.\n\n",
535
+ "target_delimiter": " ",
536
+ "fewshot_delimiter": "\n\n",
537
+ "fewshot_config": {
538
+ "sampler": "first_n",
539
+ "sampler_config": {
540
+ "fewshot_indices": [
541
+ 1893,
542
+ 949,
543
+ 663,
544
+ 105,
545
+ 1169,
546
+ 2910,
547
+ 2227,
548
+ 2813,
549
+ 974,
550
+ 558,
551
+ 1503,
552
+ 1958,
553
+ 2918,
554
+ 601,
555
+ 1560,
556
+ 984,
557
+ 2388,
558
+ 995,
559
+ 2233,
560
+ 1982,
561
+ 165,
562
+ 2788,
563
+ 1312,
564
+ 2285,
565
+ 522,
566
+ 1113,
567
+ 1670,
568
+ 323,
569
+ 236,
570
+ 1263,
571
+ 1562,
572
+ 2519,
573
+ 1049,
574
+ 432,
575
+ 1167,
576
+ 1394,
577
+ 2022,
578
+ 2551,
579
+ 2194,
580
+ 2187,
581
+ 2282,
582
+ 2816,
583
+ 108,
584
+ 301,
585
+ 1185,
586
+ 1315,
587
+ 1420,
588
+ 2436,
589
+ 2322,
590
+ 766
591
+ ]
592
+ }
593
+ },
594
+ "num_fewshot": 15,
595
+ "metric_list": [
596
+ {
597
+ "metric": "f1_macro",
598
+ "aggregation": "f1_macro",
599
+ "higher_is_better": true
600
+ },
601
+ {
602
+ "metric": "acc",
603
+ "aggregation": "acc",
604
+ "higher_is_better": true
605
+ }
606
+ ],
607
+ "output_type": "generate_until",
608
+ "generation_kwargs": {
609
+ "max_gen_toks": 32,
610
+ "do_sample": false,
611
+ "temperature": 0.0,
612
+ "top_k": null,
613
+ "top_p": null,
614
+ "until": [
615
+ "\n\n"
616
+ ]
617
+ },
618
+ "repeats": 1,
619
+ "filter_list": [
620
+ {
621
+ "name": "all",
622
+ "filter": [
623
+ {
624
+ "function": "find_similar_label",
625
+ "labels": [
626
+ "Sim",
627
+ "Não"
628
+ ]
629
+ },
630
+ {
631
+ "function": "take_first"
632
+ }
633
+ ]
634
+ }
635
+ ],
636
+ "should_decontaminate": false,
637
+ "metadata": {
638
+ "version": 1.0
639
+ }
640
+ },
641
+ "oab_exams": {
642
+ "task": "oab_exams",
643
+ "group": [
644
+ "legal_benchmark",
645
+ "pt_benchmark"
646
+ ],
647
+ "dataset_path": "eduagarcia/oab_exams",
648
+ "test_split": "train",
649
+ "fewshot_split": "train",
650
+ "doc_to_text": "<function doc_to_text at 0x7f71b0034b80>",
651
+ "doc_to_target": "{{answerKey}}",
652
+ "description": "As perguntas a seguir são questões de multipla escolha do Exame de Ordem da Ordem dos Advogados do Brasil (OAB), reponda apenas com as letras A, B, C ou D.\n\n",
653
+ "target_delimiter": " ",
654
+ "fewshot_delimiter": "\n\n",
655
+ "fewshot_config": {
656
+ "sampler": "id_sampler",
657
+ "sampler_config": {
658
+ "id_list": [
659
+ "2010-01_1",
660
+ "2010-01_11",
661
+ "2010-01_13",
662
+ "2010-01_23",
663
+ "2010-01_26",
664
+ "2010-01_28",
665
+ "2010-01_38",
666
+ "2010-01_48",
667
+ "2010-01_58",
668
+ "2010-01_68",
669
+ "2010-01_76",
670
+ "2010-01_83",
671
+ "2010-01_85",
672
+ "2010-01_91",
673
+ "2010-01_99"
674
+ ],
675
+ "id_column": "id",
676
+ "exclude_from_task": true
677
+ }
678
+ },
679
+ "num_fewshot": 3,
680
+ "metric_list": [
681
+ {
682
+ "metric": "acc",
683
+ "aggregation": "acc",
684
+ "higher_is_better": true
685
+ }
686
+ ],
687
+ "output_type": "generate_until",
688
+ "generation_kwargs": {
689
+ "max_gen_toks": 32,
690
+ "do_sample": false,
691
+ "temperature": 0.0,
692
+ "top_k": null,
693
+ "top_p": null,
694
+ "until": [
695
+ "\n\n"
696
+ ]
697
+ },
698
+ "repeats": 1,
699
+ "filter_list": [
700
+ {
701
+ "name": "all",
702
+ "filter": [
703
+ {
704
+ "function": "normalize_spaces"
705
+ },
706
+ {
707
+ "function": "remove_accents"
708
+ },
709
+ {
710
+ "function": "find_choices",
711
+ "choices": [
712
+ "A",
713
+ "B",
714
+ "C",
715
+ "D"
716
+ ],
717
+ "regex_patterns": [
718
+ "(?:[Ll]etra|[Aa]lternativa|[Rr]esposta|[Rr]esposta [Cc]orreta|[Rr]esposta[Cc]orreta e|[Oo]pcao):? ([ABCD])\\b",
719
+ "\\b([ABCD])\\)",
720
+ "\\b([ABCD]) ?[.):-]",
721
+ "\\b([ABCD])$",
722
+ "\\b([ABCD])\\b"
723
+ ]
724
+ },
725
+ {
726
+ "function": "take_first"
727
+ }
728
+ ],
729
+ "group_by": {
730
+ "column": "exam_id"
731
+ }
732
+ }
733
+ ],
734
+ "should_decontaminate": true,
735
+ "doc_to_decontamination_query": "<function doc_to_text at 0x7f71b0034e00>",
736
+ "metadata": {
737
+ "version": 1.4
738
+ }
739
+ },
740
+ "sparrow_emotion-2021-cortiz-por": {
741
+ "task": "sparrow_emotion-2021-cortiz-por",
742
+ "task_alias": "emotion-2021-cortiz-por",
743
+ "group": [
744
+ "pt_benchmark",
745
+ "sparrow"
746
+ ],
747
+ "dataset_path": "UBC-NLP/sparrow",
748
+ "dataset_name": "emotion-2021-cortiz-por",
749
+ "test_split": "validation",
750
+ "fewshot_split": "train",
751
+ "doc_to_text": "Texto: {{content}}\nPergunta: Qual a principal emoção apresentada no texto?\nResposta:",
752
+ "doc_to_target": "<function sparrow_emotion_por_trans_label at 0x7f71b0035080>",
753
+ "description": "Abaixo contém o conteúdo de tweets de usuarios do Twitter em português, sua tarefa é extrair qual a principal emoção dos textos. Responda com apenas uma das seguintes opções:\n Admiração, Diversão, Raiva, Aborrecimento, Aprovação, Compaixão, Confusão, Curiosidade, Desejo, Decepção, Desaprovação, Nojo, Vergonha, Inveja, Entusiasmo, Medo, Gratidão, Luto, Alegria, Saudade, Amor, Nervosismo, Otimismo, Orgulho, Alívio, Remorso, Tristeza ou Surpresa.\n\n",
754
+ "target_delimiter": " ",
755
+ "fewshot_delimiter": "\n\n",
756
+ "fewshot_config": {
757
+ "sampler": "first_n"
758
+ },
759
+ "num_fewshot": 25,
760
+ "metric_list": [
761
+ {
762
+ "metric": "f1_macro",
763
+ "aggregation": "f1_macro",
764
+ "higher_is_better": true
765
+ },
766
+ {
767
+ "metric": "acc",
768
+ "aggregation": "acc",
769
+ "higher_is_better": true
770
+ }
771
+ ],
772
+ "output_type": "generate_until",
773
+ "generation_kwargs": {
774
+ "max_gen_toks": 32,
775
+ "do_sample": false,
776
+ "temperature": 0.0,
777
+ "top_k": null,
778
+ "top_p": null,
779
+ "until": [
780
+ "\n\n"
781
+ ]
782
+ },
783
+ "repeats": 1,
784
+ "filter_list": [
785
+ {
786
+ "name": "all",
787
+ "filter": [
788
+ {
789
+ "function": "find_similar_label",
790
+ "labels": [
791
+ "Admiração",
792
+ "Diversão",
793
+ "Raiva",
794
+ "Aborrecimento",
795
+ "Aprovação",
796
+ "Compaixão",
797
+ "Confusão",
798
+ "Curiosidade",
799
+ "Desejo",
800
+ "Decepção",
801
+ "Desaprovação",
802
+ "Nojo",
803
+ " Vergonha",
804
+ "Inveja",
805
+ "Entusiasmo",
806
+ "Medo",
807
+ "Gratidão",
808
+ "Luto",
809
+ "Alegria",
810
+ "Saudade",
811
+ "Amor",
812
+ "Nervosismo",
813
+ "Otimismo",
814
+ "Orgulho",
815
+ "Alívio",
816
+ "Remorso",
817
+ "Tristeza",
818
+ "Surpresa"
819
+ ]
820
+ },
821
+ {
822
+ "function": "take_first"
823
+ }
824
+ ]
825
+ }
826
+ ],
827
+ "should_decontaminate": false,
828
+ "metadata": {
829
+ "version": 1.0
830
+ }
831
+ },
832
+ "sparrow_hate-2019-fortuna-por": {
833
+ "task": "sparrow_hate-2019-fortuna-por",
834
+ "task_alias": "hate-2019-fortuna-por",
835
+ "group": [
836
+ "pt_benchmark",
837
+ "sparrow"
838
+ ],
839
+ "dataset_path": "UBC-NLP/sparrow",
840
+ "dataset_name": "hate-2019-fortuna-por",
841
+ "test_split": "validation",
842
+ "fewshot_split": "train",
843
+ "doc_to_text": "Texto: {{content}}\nPergunta: O texto contém discurso de ódio?\nResposta:",
844
+ "doc_to_target": "{{'Sim' if label == 'Hate' else 'Não'}}",
845
+ "description": "Abaixo contém o conteúdo de tweets de usuarios do Twitter em português, sua tarefa é classificar se o texto contem discurso de ódio our não. Responda apenas com Sim ou Não.\n\n",
846
+ "target_delimiter": " ",
847
+ "fewshot_delimiter": "\n\n",
848
+ "fewshot_config": {
849
+ "sampler": "first_n"
850
+ },
851
+ "num_fewshot": 25,
852
+ "metric_list": [
853
+ {
854
+ "metric": "f1_macro",
855
+ "aggregation": "f1_macro",
856
+ "higher_is_better": true
857
+ },
858
+ {
859
+ "metric": "acc",
860
+ "aggregation": "acc",
861
+ "higher_is_better": true
862
+ }
863
+ ],
864
+ "output_type": "generate_until",
865
+ "generation_kwargs": {
866
+ "max_gen_toks": 32,
867
+ "do_sample": false,
868
+ "temperature": 0.0,
869
+ "top_k": null,
870
+ "top_p": null,
871
+ "until": [
872
+ "\n\n"
873
+ ]
874
+ },
875
+ "repeats": 1,
876
+ "filter_list": [
877
+ {
878
+ "name": "all",
879
+ "filter": [
880
+ {
881
+ "function": "find_similar_label",
882
+ "labels": [
883
+ "Sim",
884
+ "Não"
885
+ ]
886
+ },
887
+ {
888
+ "function": "take_first"
889
+ }
890
+ ]
891
+ }
892
+ ],
893
+ "should_decontaminate": false,
894
+ "metadata": {
895
+ "version": 1.0
896
+ }
897
+ },
898
+ "sparrow_sentiment-2016-mozetic-por": {
899
+ "task": "sparrow_sentiment-2016-mozetic-por",
900
+ "task_alias": "sentiment-2016-mozetic-por",
901
+ "group": [
902
+ "pt_benchmark",
903
+ "sparrow"
904
+ ],
905
+ "dataset_path": "UBC-NLP/sparrow",
906
+ "dataset_name": "sentiment-2016-mozetic-por",
907
+ "test_split": "validation",
908
+ "fewshot_split": "train",
909
+ "doc_to_text": "Texto: {{content}}\nPergunta: O sentimento do texto é Positivo, Neutro ou Negativo?\nResposta:",
910
+ "doc_to_target": "{{'Positivo' if label == 'Positive' else ('Negativo' if label == 'Negative' else 'Neutro')}}",
911
+ "description": "Abaixo contém o conteúdo de tweets de usuarios do Twitter em português, sua tarefa é classificar se o sentimento do texto é Positivo, Neutro ou Negativo. Responda apenas com uma das opções.\n\n",
912
+ "target_delimiter": " ",
913
+ "fewshot_delimiter": "\n\n",
914
+ "fewshot_config": {
915
+ "sampler": "first_n"
916
+ },
917
+ "num_fewshot": 25,
918
+ "metric_list": [
919
+ {
920
+ "metric": "f1_macro",
921
+ "aggregation": "f1_macro",
922
+ "higher_is_better": true
923
+ },
924
+ {
925
+ "metric": "acc",
926
+ "aggregation": "acc",
927
+ "higher_is_better": true
928
+ }
929
+ ],
930
+ "output_type": "generate_until",
931
+ "generation_kwargs": {
932
+ "max_gen_toks": 32,
933
+ "do_sample": false,
934
+ "temperature": 0.0,
935
+ "top_k": null,
936
+ "top_p": null,
937
+ "until": [
938
+ "\n\n"
939
+ ]
940
+ },
941
+ "repeats": 1,
942
+ "filter_list": [
943
+ {
944
+ "name": "all",
945
+ "filter": [
946
+ {
947
+ "function": "find_similar_label",
948
+ "labels": [
949
+ "Positivo",
950
+ "Neutro",
951
+ "Negativo"
952
+ ]
953
+ },
954
+ {
955
+ "function": "take_first"
956
+ }
957
+ ]
958
+ }
959
+ ],
960
+ "should_decontaminate": false,
961
+ "metadata": {
962
+ "version": 1.0
963
+ }
964
+ },
965
+ "sparrow_sentiment-2018-brum-por": {
966
+ "task": "sparrow_sentiment-2018-brum-por",
967
+ "task_alias": "sentiment-2018-brum-por",
968
+ "group": [
969
+ "pt_benchmark",
970
+ "sparrow"
971
+ ],
972
+ "dataset_path": "UBC-NLP/sparrow",
973
+ "dataset_name": "sentiment-2018-brum-por",
974
+ "test_split": "validation",
975
+ "fewshot_split": "train",
976
+ "doc_to_text": "Texto: {{content}}\nPergunta: O sentimento do texto é Positivo, Neutro ou Negativo?\nResposta:",
977
+ "doc_to_target": "{{'Positivo' if label == 'Positive' else ('Negativo' if label == 'Negative' else 'Neutro')}}",
978
+ "description": "Abaixo contém o conteúdo de tweets de usuarios do Twitter em português, sua tarefa é classificar se o sentimento do texto é Positivo, Neutro ou Negativo. Responda apenas com uma das opções.\n\n",
979
+ "target_delimiter": " ",
980
+ "fewshot_delimiter": "\n\n",
981
+ "fewshot_config": {
982
+ "sampler": "first_n"
983
+ },
984
+ "num_fewshot": 25,
985
+ "metric_list": [
986
+ {
987
+ "metric": "f1_macro",
988
+ "aggregation": "f1_macro",
989
+ "higher_is_better": true
990
+ },
991
+ {
992
+ "metric": "acc",
993
+ "aggregation": "acc",
994
+ "higher_is_better": true
995
+ }
996
+ ],
997
+ "output_type": "generate_until",
998
+ "generation_kwargs": {
999
+ "max_gen_toks": 32,
1000
+ "do_sample": false,
1001
+ "temperature": 0.0,
1002
+ "top_k": null,
1003
+ "top_p": null,
1004
+ "until": [
1005
+ "\n\n"
1006
+ ]
1007
+ },
1008
+ "repeats": 1,
1009
+ "filter_list": [
1010
+ {
1011
+ "name": "all",
1012
+ "filter": [
1013
+ {
1014
+ "function": "find_similar_label",
1015
+ "labels": [
1016
+ "Positivo",
1017
+ "Neutro",
1018
+ "Negativo"
1019
+ ]
1020
+ },
1021
+ {
1022
+ "function": "take_first"
1023
+ }
1024
+ ]
1025
+ }
1026
+ ],
1027
+ "should_decontaminate": false,
1028
+ "metadata": {
1029
+ "version": 1.0
1030
+ }
1031
+ }
1032
+ },
1033
+ "versions": {
1034
+ "assin2_rte": 1.0,
1035
+ "assin2_sts": 1.0,
1036
+ "bluex": 1.0,
1037
+ "enem_challenge": 1.0,
1038
+ "faquad_nli": 1.0,
1039
+ "oab_exams": 1.4,
1040
+ "sparrow_emotion-2021-cortiz-por": 1.0,
1041
+ "sparrow_hate-2019-fortuna-por": 1.0,
1042
+ "sparrow_sentiment-2016-mozetic-por": 1.0,
1043
+ "sparrow_sentiment-2018-brum-por": 1.0
1044
+ },
1045
+ "n-shot": {
1046
+ "assin2_rte": 15,
1047
+ "assin2_sts": 15,
1048
+ "bluex": 3,
1049
+ "enem_challenge": 3,
1050
+ "faquad_nli": 15,
1051
+ "oab_exams": 3,
1052
+ "sparrow_emotion-2021-cortiz-por": 25,
1053
+ "sparrow_hate-2019-fortuna-por": 25,
1054
+ "sparrow_sentiment-2016-mozetic-por": 25,
1055
+ "sparrow_sentiment-2018-brum-por": 25
1056
+ },
1057
+ "model_meta": {
1058
+ "truncated": 0,
1059
+ "non_truncated": 11889,
1060
+ "padded": 0,
1061
+ "non_padded": 11889,
1062
+ "fewshots_truncated": 0,
1063
+ "has_chat_template": false,
1064
+ "chat_type": null,
1065
+ "n_gpus": 2,
1066
+ "accelerate_num_process": null,
1067
+ "model_sha": "985aa055896a8f943d4a9f2572e6ea1341823841",
1068
+ "model_dtype": "torch.bfloat16",
1069
+ "model_memory_footprint": 93942464512,
1070
+ "model_num_parameters": 46702792704,
1071
+ "model_is_loaded_in_4bit": false,
1072
+ "model_is_loaded_in_8bit": false,
1073
+ "model_is_quantized": null,
1074
+ "model_device": "cuda:0",
1075
+ "batch_size": 2,
1076
+ "max_length": 4096,
1077
+ "max_ctx_length": 4064,
1078
+ "max_gen_toks": 32
1079
+ },
1080
+ "task_model_meta": {
1081
+ "assin2_rte": {
1082
+ "sample_size": 2448,
1083
+ "truncated": 0,
1084
+ "non_truncated": 2448,
1085
+ "padded": 0,
1086
+ "non_padded": 2448,
1087
+ "fewshots_truncated": 0,
1088
+ "mean_seq_length": 1244.7455065359477,
1089
+ "min_seq_length": 1221,
1090
+ "max_seq_length": 1311,
1091
+ "max_ctx_length": 4064,
1092
+ "max_gen_toks": 32,
1093
+ "mean_original_fewshots_size": 15.0,
1094
+ "mean_effective_fewshot_size": 15.0
1095
+ },
1096
+ "assin2_sts": {
1097
+ "sample_size": 2448,
1098
+ "truncated": 0,
1099
+ "non_truncated": 2448,
1100
+ "padded": 0,
1101
+ "non_padded": 2448,
1102
+ "fewshots_truncated": 0,
1103
+ "mean_seq_length": 1459.7455065359477,
1104
+ "min_seq_length": 1436,
1105
+ "max_seq_length": 1526,
1106
+ "max_ctx_length": 4064,
1107
+ "max_gen_toks": 32,
1108
+ "mean_original_fewshots_size": 15.0,
1109
+ "mean_effective_fewshot_size": 15.0
1110
+ },
1111
+ "bluex": {
1112
+ "sample_size": 719,
1113
+ "truncated": 0,
1114
+ "non_truncated": 719,
1115
+ "padded": 0,
1116
+ "non_padded": 719,
1117
+ "fewshots_truncated": 0,
1118
+ "mean_seq_length": 1394.9262865090404,
1119
+ "min_seq_length": 1018,
1120
+ "max_seq_length": 2195,
1121
+ "max_ctx_length": 4064,
1122
+ "max_gen_toks": 32,
1123
+ "mean_original_fewshots_size": 3.0,
1124
+ "mean_effective_fewshot_size": 3.0
1125
+ },
1126
+ "enem_challenge": {
1127
+ "sample_size": 1429,
1128
+ "truncated": 0,
1129
+ "non_truncated": 1429,
1130
+ "padded": 0,
1131
+ "non_padded": 1429,
1132
+ "fewshots_truncated": 0,
1133
+ "mean_seq_length": 1605.039188243527,
1134
+ "min_seq_length": 1339,
1135
+ "max_seq_length": 2603,
1136
+ "max_ctx_length": 4064,
1137
+ "max_gen_toks": 32,
1138
+ "mean_original_fewshots_size": 3.0,
1139
+ "mean_effective_fewshot_size": 3.0
1140
+ },
1141
+ "faquad_nli": {
1142
+ "sample_size": 650,
1143
+ "truncated": 0,
1144
+ "non_truncated": 650,
1145
+ "padded": 0,
1146
+ "non_padded": 650,
1147
+ "fewshots_truncated": 0,
1148
+ "mean_seq_length": 1555.9876923076922,
1149
+ "min_seq_length": 1500,
1150
+ "max_seq_length": 1676,
1151
+ "max_ctx_length": 4064,
1152
+ "max_gen_toks": 32,
1153
+ "mean_original_fewshots_size": 15.0,
1154
+ "mean_effective_fewshot_size": 15.0
1155
+ },
1156
+ "oab_exams": {
1157
+ "sample_size": 2195,
1158
+ "truncated": 0,
1159
+ "non_truncated": 2195,
1160
+ "padded": 0,
1161
+ "non_padded": 2195,
1162
+ "fewshots_truncated": 0,
1163
+ "mean_seq_length": 1351.764464692483,
1164
+ "min_seq_length": 1085,
1165
+ "max_seq_length": 1854,
1166
+ "max_ctx_length": 4064,
1167
+ "max_gen_toks": 32,
1168
+ "mean_original_fewshots_size": 3.0,
1169
+ "mean_effective_fewshot_size": 3.0
1170
+ },
1171
+ "sparrow_emotion-2021-cortiz-por": {
1172
+ "sample_size": 500,
1173
+ "truncated": 0,
1174
+ "non_truncated": 500,
1175
+ "padded": 0,
1176
+ "non_padded": 500,
1177
+ "fewshots_truncated": 0,
1178
+ "mean_seq_length": 1723.336,
1179
+ "min_seq_length": 1701,
1180
+ "max_seq_length": 1757,
1181
+ "max_ctx_length": 4064,
1182
+ "max_gen_toks": 32,
1183
+ "mean_original_fewshots_size": 25.0,
1184
+ "mean_effective_fewshot_size": 25.0
1185
+ },
1186
+ "sparrow_hate-2019-fortuna-por": {
1187
+ "sample_size": 500,
1188
+ "truncated": 0,
1189
+ "non_truncated": 500,
1190
+ "padded": 0,
1191
+ "non_padded": 500,
1192
+ "fewshots_truncated": 0,
1193
+ "mean_seq_length": 1746.15,
1194
+ "min_seq_length": 1722,
1195
+ "max_seq_length": 1791,
1196
+ "max_ctx_length": 4064,
1197
+ "max_gen_toks": 32,
1198
+ "mean_original_fewshots_size": 25.0,
1199
+ "mean_effective_fewshot_size": 25.0
1200
+ },
1201
+ "sparrow_sentiment-2016-mozetic-por": {
1202
+ "sample_size": 500,
1203
+ "truncated": 0,
1204
+ "non_truncated": 500,
1205
+ "padded": 0,
1206
+ "non_padded": 500,
1207
+ "fewshots_truncated": 0,
1208
+ "mean_seq_length": 1458.088,
1209
+ "min_seq_length": 1441,
1210
+ "max_seq_length": 1494,
1211
+ "max_ctx_length": 4064,
1212
+ "max_gen_toks": 32,
1213
+ "mean_original_fewshots_size": 25.0,
1214
+ "mean_effective_fewshot_size": 25.0
1215
+ },
1216
+ "sparrow_sentiment-2018-brum-por": {
1217
+ "sample_size": 500,
1218
+ "truncated": 0,
1219
+ "non_truncated": 500,
1220
+ "padded": 0,
1221
+ "non_padded": 500,
1222
+ "fewshots_truncated": 0,
1223
+ "mean_seq_length": 1623.55,
1224
+ "min_seq_length": 1606,
1225
+ "max_seq_length": 1656,
1226
+ "max_ctx_length": 4064,
1227
+ "max_gen_toks": 32,
1228
+ "mean_original_fewshots_size": 25.0,
1229
+ "mean_effective_fewshot_size": 25.0
1230
+ }
1231
+ },
1232
+ "config": {
1233
+ "model": "huggingface",
1234
+ "model_args": "pretrained=mistralai/Mixtral-8x7B-v0.1,dtype=bfloat16,parallelize=True,revision=main,trust_remote_code=True,starting_max_length=4096",
1235
+ "batch_size": "auto",
1236
+ "batch_sizes": [],
1237
+ "device": null,
1238
+ "use_cache": null,
1239
+ "limit": [
1240
+ null,
1241
+ null,
1242
+ null,
1243
+ null,
1244
+ null,
1245
+ null,
1246
+ 500.0,
1247
+ 500.0,
1248
+ 500.0,
1249
+ 500.0
1250
+ ],
1251
+ "bootstrap_iters": 0,
1252
+ "gen_kwargs": null
1253
+ },
1254
+ "git_hash": "15f86b5"
1255
+ }
mistralai/Mixtral-8x7B-v0.1/results_2024-02-10T03-25-53.024286.json ADDED
@@ -0,0 +1,346 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config_general": {
3
+ "start_date": "2024-02-10T03-25-53.024286",
4
+ "start_time": 1707535554.3225698,
5
+ "end_time": 1707544616.880482,
6
+ "total_evaluation_time_seconds": 9062.557912111282,
7
+ "has_chat_template": false,
8
+ "chat_type": null,
9
+ "n_gpus": 2,
10
+ "accelerate_num_process": null,
11
+ "model_sha": "985aa055896a8f943d4a9f2572e6ea1341823841",
12
+ "model_dtype": "bfloat16",
13
+ "model_memory_footprint": 93942464512,
14
+ "model_num_parameters": 46702792704,
15
+ "model_is_loaded_in_4bit": false,
16
+ "model_is_loaded_in_8bit": false,
17
+ "model_is_quantized": null,
18
+ "model_device": "cuda:0",
19
+ "batch_size": 2,
20
+ "max_length": 4096,
21
+ "max_ctx_length": 4064,
22
+ "max_gen_toks": 32,
23
+ "model_name": "mistralai/Mixtral-8x7B-v0.1",
24
+ "job_id": 176,
25
+ "model_id": "mistralai/Mixtral-8x7B-v0.1_eval_request_False_bfloat16_Original",
26
+ "model_base_model": "",
27
+ "model_weight_type": "Original",
28
+ "model_revision": "main",
29
+ "model_private": false,
30
+ "model_type": "🟢 : pretrained",
31
+ "model_architectures": "MixtralForCausalLM",
32
+ "submitted_time": "2024-02-05T23:05:20Z",
33
+ "lm_eval_model_type": "huggingface",
34
+ "eval_version": "1.0.0"
35
+ },
36
+ "results": {
37
+ "all_grouped_average": 0.564565185228941,
38
+ "all_grouped_npm": 0.37330116490004156,
39
+ "all_grouped": {
40
+ "enem_challenge": 0.7249825052484254,
41
+ "bluex": 0.6369958275382476,
42
+ "oab_exams": 0.5553530751708429,
43
+ "assin2_rte": 0.6646069030555132,
44
+ "assin2_sts": 0.49872071984375066,
45
+ "faquad_nli": 0.5026086956521739,
46
+ "sparrow_pt": 0.3686885700936329
47
+ },
48
+ "all": {
49
+ "harness|enem_challenge|enem_challenge|None|3": 0.7249825052484254,
50
+ "harness|bluex|bluex|None|3": 0.6369958275382476,
51
+ "harness|oab_exams|oab_exams|None|3": 0.5553530751708429,
52
+ "harness|assin2_rte|assin2_rte|None|15": 0.6646069030555132,
53
+ "harness|assin2_sts|assin2_sts|None|15": 0.49872071984375066,
54
+ "harness|faquad_nli|faquad_nli|None|15": 0.5026086956521739,
55
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por|500|25": 0.10240832053715039,
56
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por|500|25": 0.4876514802883878,
57
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por|500|25": 0.47227979759408134,
58
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por|500|25": 0.4124146819549119
59
+ },
60
+ "harness|enem_challenge|enem_challenge|None|3": {
61
+ "acc,all": 0.7249825052484254,
62
+ "acc,exam_id__2015": 0.773109243697479,
63
+ "acc,exam_id__2009": 0.7217391304347827,
64
+ "acc,exam_id__2014": 0.7431192660550459,
65
+ "acc,exam_id__2010": 0.7008547008547008,
66
+ "acc,exam_id__2011": 0.7863247863247863,
67
+ "acc,exam_id__2023": 0.725925925925926,
68
+ "acc,exam_id__2016_2": 0.7317073170731707,
69
+ "acc,exam_id__2016": 0.7107438016528925,
70
+ "acc,exam_id__2013": 0.75,
71
+ "acc,exam_id__2022": 0.6466165413533834,
72
+ "acc,exam_id__2017": 0.6982758620689655,
73
+ "acc,exam_id__2012": 0.7241379310344828,
74
+ "main_score": 0.7249825052484254
75
+ },
76
+ "harness|bluex|bluex|None|3": {
77
+ "acc,all": 0.6369958275382476,
78
+ "acc,exam_id__USP_2019": 0.6,
79
+ "acc,exam_id__USP_2018": 0.5740740740740741,
80
+ "acc,exam_id__UNICAMP_2018": 0.5555555555555556,
81
+ "acc,exam_id__UNICAMP_2023": 0.7674418604651163,
82
+ "acc,exam_id__USP_2024": 0.7073170731707317,
83
+ "acc,exam_id__UNICAMP_2021_1": 0.5869565217391305,
84
+ "acc,exam_id__UNICAMP_2019": 0.68,
85
+ "acc,exam_id__USP_2020": 0.6607142857142857,
86
+ "acc,exam_id__UNICAMP_2020": 0.6545454545454545,
87
+ "acc,exam_id__USP_2022": 0.6122448979591837,
88
+ "acc,exam_id__USP_2023": 0.75,
89
+ "acc,exam_id__UNICAMP_2021_2": 0.5882352941176471,
90
+ "acc,exam_id__UNICAMP_2024": 0.5555555555555556,
91
+ "acc,exam_id__USP_2021": 0.6153846153846154,
92
+ "acc,exam_id__UNICAMP_2022": 0.6923076923076923,
93
+ "main_score": 0.6369958275382476
94
+ },
95
+ "harness|oab_exams|oab_exams|None|3": {
96
+ "acc,all": 0.5553530751708429,
97
+ "acc,exam_id__2016-20": 0.5875,
98
+ "acc,exam_id__2016-20a": 0.45,
99
+ "acc,exam_id__2017-22": 0.5875,
100
+ "acc,exam_id__2014-14": 0.6,
101
+ "acc,exam_id__2010-01": 0.49411764705882355,
102
+ "acc,exam_id__2017-23": 0.5125,
103
+ "acc,exam_id__2014-15": 0.717948717948718,
104
+ "acc,exam_id__2011-04": 0.525,
105
+ "acc,exam_id__2015-17": 0.6923076923076923,
106
+ "acc,exam_id__2013-10": 0.5625,
107
+ "acc,exam_id__2012-07": 0.5,
108
+ "acc,exam_id__2011-03": 0.5151515151515151,
109
+ "acc,exam_id__2012-06a": 0.575,
110
+ "acc,exam_id__2012-09": 0.45454545454545453,
111
+ "acc,exam_id__2014-13": 0.525,
112
+ "acc,exam_id__2015-18": 0.65,
113
+ "acc,exam_id__2011-05": 0.5625,
114
+ "acc,exam_id__2012-08": 0.5625,
115
+ "acc,exam_id__2018-25": 0.5375,
116
+ "acc,exam_id__2016-21": 0.5,
117
+ "acc,exam_id__2013-11": 0.5625,
118
+ "acc,exam_id__2010-02": 0.6,
119
+ "acc,exam_id__2015-16": 0.5875,
120
+ "acc,exam_id__2013-12": 0.55,
121
+ "acc,exam_id__2016-19": 0.5769230769230769,
122
+ "acc,exam_id__2017-24": 0.4625,
123
+ "acc,exam_id__2012-06": 0.55,
124
+ "main_score": 0.5553530751708429
125
+ },
126
+ "harness|assin2_rte|assin2_rte|None|15": {
127
+ "f1_macro,all": 0.6646069030555132,
128
+ "acc,all": 0.6956699346405228,
129
+ "main_score": 0.6646069030555132
130
+ },
131
+ "harness|assin2_sts|assin2_sts|None|15": {
132
+ "pearson,all": 0.49872071984375066,
133
+ "mse,all": 1.9470588235294124,
134
+ "main_score": 0.49872071984375066
135
+ },
136
+ "harness|faquad_nli|faquad_nli|None|15": {
137
+ "f1_macro,all": 0.5026086956521739,
138
+ "acc,all": 0.796923076923077,
139
+ "main_score": 0.5026086956521739
140
+ },
141
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por|500|25": {
142
+ "f1_macro,all": 0.10240832053715039,
143
+ "acc,all": 0.148,
144
+ "main_score": 0.10240832053715039
145
+ },
146
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por|500|25": {
147
+ "f1_macro,all": 0.4876514802883878,
148
+ "acc,all": 0.666,
149
+ "main_score": 0.4876514802883878
150
+ },
151
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por|500|25": {
152
+ "f1_macro,all": 0.47227979759408134,
153
+ "acc,all": 0.516,
154
+ "main_score": 0.47227979759408134
155
+ },
156
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por|500|25": {
157
+ "f1_macro,all": 0.4124146819549119,
158
+ "acc,all": 0.414,
159
+ "main_score": 0.4124146819549119
160
+ }
161
+ },
162
+ "config_tasks": {
163
+ "harness|enem_challenge|enem_challenge": "LM Harness task",
164
+ "harness|bluex|bluex": "LM Harness task",
165
+ "harness|oab_exams|oab_exams": "LM Harness task",
166
+ "harness|assin2_rte|assin2_rte": "LM Harness task",
167
+ "harness|assin2_sts|assin2_sts": "LM Harness task",
168
+ "harness|faquad_nli|faquad_nli": "LM Harness task",
169
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por": "LM Harness task",
170
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por": "LM Harness task",
171
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por": "LM Harness task",
172
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por": "LM Harness task"
173
+ },
174
+ "versions": {
175
+ "all": 0,
176
+ "harness|enem_challenge|enem_challenge": 1.0,
177
+ "harness|bluex|bluex": 1.0,
178
+ "harness|oab_exams|oab_exams": 1.4,
179
+ "harness|assin2_rte|assin2_rte": 1.0,
180
+ "harness|assin2_sts|assin2_sts": 1.0,
181
+ "harness|faquad_nli|faquad_nli": 1.0,
182
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por": 1.0,
183
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por": 1.0,
184
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por": 1.0,
185
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por": 1.0
186
+ },
187
+ "summary_tasks": {
188
+ "harness|enem_challenge|enem_challenge|None|3": {
189
+ "sample_size": 1429,
190
+ "truncated": 0,
191
+ "non_truncated": 1429,
192
+ "padded": 0,
193
+ "non_padded": 1429,
194
+ "fewshots_truncated": 0,
195
+ "mean_seq_length": 1605.039188243527,
196
+ "min_seq_length": 1339,
197
+ "max_seq_length": 2603,
198
+ "max_ctx_length": 4064,
199
+ "max_gen_toks": 32,
200
+ "mean_original_fewshots_size": 3.0,
201
+ "mean_effective_fewshot_size": 3.0
202
+ },
203
+ "harness|bluex|bluex|None|3": {
204
+ "sample_size": 719,
205
+ "truncated": 0,
206
+ "non_truncated": 719,
207
+ "padded": 0,
208
+ "non_padded": 719,
209
+ "fewshots_truncated": 0,
210
+ "mean_seq_length": 1394.9262865090404,
211
+ "min_seq_length": 1018,
212
+ "max_seq_length": 2195,
213
+ "max_ctx_length": 4064,
214
+ "max_gen_toks": 32,
215
+ "mean_original_fewshots_size": 3.0,
216
+ "mean_effective_fewshot_size": 3.0
217
+ },
218
+ "harness|oab_exams|oab_exams|None|3": {
219
+ "sample_size": 2195,
220
+ "truncated": 0,
221
+ "non_truncated": 2195,
222
+ "padded": 0,
223
+ "non_padded": 2195,
224
+ "fewshots_truncated": 0,
225
+ "mean_seq_length": 1351.764464692483,
226
+ "min_seq_length": 1085,
227
+ "max_seq_length": 1854,
228
+ "max_ctx_length": 4064,
229
+ "max_gen_toks": 32,
230
+ "mean_original_fewshots_size": 3.0,
231
+ "mean_effective_fewshot_size": 3.0
232
+ },
233
+ "harness|assin2_rte|assin2_rte|None|15": {
234
+ "sample_size": 2448,
235
+ "truncated": 0,
236
+ "non_truncated": 2448,
237
+ "padded": 0,
238
+ "non_padded": 2448,
239
+ "fewshots_truncated": 0,
240
+ "mean_seq_length": 1244.7455065359477,
241
+ "min_seq_length": 1221,
242
+ "max_seq_length": 1311,
243
+ "max_ctx_length": 4064,
244
+ "max_gen_toks": 32,
245
+ "mean_original_fewshots_size": 15.0,
246
+ "mean_effective_fewshot_size": 15.0
247
+ },
248
+ "harness|assin2_sts|assin2_sts|None|15": {
249
+ "sample_size": 2448,
250
+ "truncated": 0,
251
+ "non_truncated": 2448,
252
+ "padded": 0,
253
+ "non_padded": 2448,
254
+ "fewshots_truncated": 0,
255
+ "mean_seq_length": 1459.7455065359477,
256
+ "min_seq_length": 1436,
257
+ "max_seq_length": 1526,
258
+ "max_ctx_length": 4064,
259
+ "max_gen_toks": 32,
260
+ "mean_original_fewshots_size": 15.0,
261
+ "mean_effective_fewshot_size": 15.0
262
+ },
263
+ "harness|faquad_nli|faquad_nli|None|15": {
264
+ "sample_size": 650,
265
+ "truncated": 0,
266
+ "non_truncated": 650,
267
+ "padded": 0,
268
+ "non_padded": 650,
269
+ "fewshots_truncated": 0,
270
+ "mean_seq_length": 1555.9876923076922,
271
+ "min_seq_length": 1500,
272
+ "max_seq_length": 1676,
273
+ "max_ctx_length": 4064,
274
+ "max_gen_toks": 32,
275
+ "mean_original_fewshots_size": 15.0,
276
+ "mean_effective_fewshot_size": 15.0
277
+ },
278
+ "harness|sparrow_pt|sparrow_emotion-2021-cortiz-por|500|25": {
279
+ "sample_size": 500,
280
+ "truncated": 0,
281
+ "non_truncated": 500,
282
+ "padded": 0,
283
+ "non_padded": 500,
284
+ "fewshots_truncated": 0,
285
+ "mean_seq_length": 1723.336,
286
+ "min_seq_length": 1701,
287
+ "max_seq_length": 1757,
288
+ "max_ctx_length": 4064,
289
+ "max_gen_toks": 32,
290
+ "mean_original_fewshots_size": 25.0,
291
+ "mean_effective_fewshot_size": 25.0
292
+ },
293
+ "harness|sparrow_pt|sparrow_hate-2019-fortuna-por|500|25": {
294
+ "sample_size": 500,
295
+ "truncated": 0,
296
+ "non_truncated": 500,
297
+ "padded": 0,
298
+ "non_padded": 500,
299
+ "fewshots_truncated": 0,
300
+ "mean_seq_length": 1746.15,
301
+ "min_seq_length": 1722,
302
+ "max_seq_length": 1791,
303
+ "max_ctx_length": 4064,
304
+ "max_gen_toks": 32,
305
+ "mean_original_fewshots_size": 25.0,
306
+ "mean_effective_fewshot_size": 25.0
307
+ },
308
+ "harness|sparrow_pt|sparrow_sentiment-2016-mozetic-por|500|25": {
309
+ "sample_size": 500,
310
+ "truncated": 0,
311
+ "non_truncated": 500,
312
+ "padded": 0,
313
+ "non_padded": 500,
314
+ "fewshots_truncated": 0,
315
+ "mean_seq_length": 1458.088,
316
+ "min_seq_length": 1441,
317
+ "max_seq_length": 1494,
318
+ "max_ctx_length": 4064,
319
+ "max_gen_toks": 32,
320
+ "mean_original_fewshots_size": 25.0,
321
+ "mean_effective_fewshot_size": 25.0
322
+ },
323
+ "harness|sparrow_pt|sparrow_sentiment-2018-brum-por|500|25": {
324
+ "sample_size": 500,
325
+ "truncated": 0,
326
+ "non_truncated": 500,
327
+ "padded": 0,
328
+ "non_padded": 500,
329
+ "fewshots_truncated": 0,
330
+ "mean_seq_length": 1623.55,
331
+ "min_seq_length": 1606,
332
+ "max_seq_length": 1656,
333
+ "max_ctx_length": 4064,
334
+ "max_gen_toks": 32,
335
+ "mean_original_fewshots_size": 25.0,
336
+ "mean_effective_fewshot_size": 25.0
337
+ }
338
+ },
339
+ "summary_general": {
340
+ "truncated": 0,
341
+ "non_truncated": 11889,
342
+ "padded": 0,
343
+ "non_padded": 11889,
344
+ "fewshots_truncated": 0
345
+ }
346
+ }