eduagarcia commited on
Commit
8fcd70f
·
verified ·
1 Parent(s): e170594

Uploading raw results for BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B

Browse files
Files changed (12) hide show
  1. .gitattributes +9 -0
  2. BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_assin2_rte.jsonl +3 -0
  3. BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_assin2_sts.jsonl +3 -0
  4. BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_bluex.jsonl +3 -0
  5. BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_enem_challenge.jsonl +3 -0
  6. BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_faquad_nli.jsonl +3 -0
  7. BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_hatebr_offensive.jsonl +3 -0
  8. BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_oab_exams.jsonl +3 -0
  9. BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_portuguese_hate_speech.jsonl +3 -0
  10. BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_tweetsentbr.jsonl +3 -0
  11. BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/results.json +1237 -0
  12. BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/results_2025-05-04T10-35-32.472663.json +326 -0
.gitattributes CHANGED
@@ -10824,3 +10824,12 @@ TIGER-Lab/Qwen2.5-Math-7B-CFT/raw_2025-04-19T01-35-24.572463/pretrained__TIGER-L
10824
  TIGER-Lab/Qwen2.5-Math-7B-CFT/raw_2025-04-19T01-35-24.572463/pretrained__TIGER-Lab__Qwen2.5-Math-7B-CFT,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_oab_exams.jsonl filter=lfs diff=lfs merge=lfs -text
10825
  TIGER-Lab/Qwen2.5-Math-7B-CFT/raw_2025-04-19T01-35-24.572463/pretrained__TIGER-Lab__Qwen2.5-Math-7B-CFT,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_portuguese_hate_speech.jsonl filter=lfs diff=lfs merge=lfs -text
10826
  TIGER-Lab/Qwen2.5-Math-7B-CFT/raw_2025-04-19T01-35-24.572463/pretrained__TIGER-Lab__Qwen2.5-Math-7B-CFT,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_tweetsentbr.jsonl filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
10824
  TIGER-Lab/Qwen2.5-Math-7B-CFT/raw_2025-04-19T01-35-24.572463/pretrained__TIGER-Lab__Qwen2.5-Math-7B-CFT,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_oab_exams.jsonl filter=lfs diff=lfs merge=lfs -text
10825
  TIGER-Lab/Qwen2.5-Math-7B-CFT/raw_2025-04-19T01-35-24.572463/pretrained__TIGER-Lab__Qwen2.5-Math-7B-CFT,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_portuguese_hate_speech.jsonl filter=lfs diff=lfs merge=lfs -text
10826
  TIGER-Lab/Qwen2.5-Math-7B-CFT/raw_2025-04-19T01-35-24.572463/pretrained__TIGER-Lab__Qwen2.5-Math-7B-CFT,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_tweetsentbr.jsonl filter=lfs diff=lfs merge=lfs -text
10827
+ BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_assin2_rte.jsonl filter=lfs diff=lfs merge=lfs -text
10828
+ BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_assin2_sts.jsonl filter=lfs diff=lfs merge=lfs -text
10829
+ BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_bluex.jsonl filter=lfs diff=lfs merge=lfs -text
10830
+ BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_enem_challenge.jsonl filter=lfs diff=lfs merge=lfs -text
10831
+ BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_faquad_nli.jsonl filter=lfs diff=lfs merge=lfs -text
10832
+ BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_hatebr_offensive.jsonl filter=lfs diff=lfs merge=lfs -text
10833
+ BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_oab_exams.jsonl filter=lfs diff=lfs merge=lfs -text
10834
+ BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_portuguese_hate_speech.jsonl filter=lfs diff=lfs merge=lfs -text
10835
+ BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_tweetsentbr.jsonl filter=lfs diff=lfs merge=lfs -text
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_assin2_rte.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8882a79ebae8a32edabccc7da0901865855cf8b09da42ff5e71977819d365140
3
+ size 37136219
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_assin2_sts.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66230f327741b70a228b338eb3701ea4e1f3022b6c46d379daaaec4441824af9
3
+ size 38718441
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_bluex.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:369bc59df455e7ecaa3ca267747bc1e900928d06fddef5b8525af06e389b7f42
3
+ size 12378447
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_enem_challenge.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f023cab9a5162c6c7992a5a59a810ba181f46aa245faf858c381e6ae5b816929
3
+ size 24811809
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_faquad_nli.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15f16a46299b434cdf5b9fd2c5ef45e80b77ac2e3a68de73c6ea3a5aa87e1be6
3
+ size 11827895
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_hatebr_offensive.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25fb0a7e304d4e79db616482bf93569278a95ba6202d7d3874c0829b6008c910
3
+ size 20008302
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_oab_exams.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d63c8bcd252f4ed4c4daff7b1d7be610cbf3eb0c152b3e9ccc3732bbef7262f6
3
+ size 34651291
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_portuguese_hate_speech.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:47557c204c14bfc44d656d0d10d1c7c8309ccbb78ee63e35a9e5719f40a53802
3
+ size 15392991
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/pretrained__BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype__bfloat16,device__cuda:0,revision__main,trust_remote_code__True,starting_max_length__2560_tweetsentbr.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e63bf2474245df092c6ddab38c75d51343543d827004fa710b0ada068928443b
3
+ size 34276775
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/raw_2025-05-04T10-35-32.472663/results.json ADDED
@@ -0,0 +1,1237 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "assin2_rte": {
4
+ "f1_macro,all": 0.8902591771846999,
5
+ "acc,all": 0.8905228758169934,
6
+ "alias": "assin2_rte"
7
+ },
8
+ "assin2_sts": {
9
+ "pearson,all": 0.7417032662621263,
10
+ "mse,all": 0.4840727124183006,
11
+ "alias": "assin2_sts"
12
+ },
13
+ "bluex": {
14
+ "acc,all": 0.5841446453407511,
15
+ "acc,exam_id__USP_2021": 0.5384615384615384,
16
+ "acc,exam_id__USP_2020": 0.5535714285714286,
17
+ "acc,exam_id__USP_2019": 0.55,
18
+ "acc,exam_id__UNICAMP_2019": 0.6,
19
+ "acc,exam_id__UNICAMP_2021_2": 0.5686274509803921,
20
+ "acc,exam_id__USP_2023": 0.75,
21
+ "acc,exam_id__UNICAMP_2024": 0.4666666666666667,
22
+ "acc,exam_id__UNICAMP_2018": 0.5,
23
+ "acc,exam_id__USP_2024": 0.7560975609756098,
24
+ "acc,exam_id__USP_2018": 0.48148148148148145,
25
+ "acc,exam_id__USP_2022": 0.5714285714285714,
26
+ "acc,exam_id__UNICAMP_2022": 0.5641025641025641,
27
+ "acc,exam_id__UNICAMP_2023": 0.7209302325581395,
28
+ "acc,exam_id__UNICAMP_2020": 0.5636363636363636,
29
+ "acc,exam_id__UNICAMP_2021_1": 0.6521739130434783,
30
+ "alias": "bluex"
31
+ },
32
+ "enem_challenge": {
33
+ "alias": "enem",
34
+ "acc,all": 0.6913925822253324,
35
+ "acc,exam_id__2011": 0.7521367521367521,
36
+ "acc,exam_id__2012": 0.6896551724137931,
37
+ "acc,exam_id__2014": 0.7064220183486238,
38
+ "acc,exam_id__2023": 0.6518518518518519,
39
+ "acc,exam_id__2010": 0.7094017094017094,
40
+ "acc,exam_id__2009": 0.6956521739130435,
41
+ "acc,exam_id__2013": 0.6851851851851852,
42
+ "acc,exam_id__2017": 0.6810344827586207,
43
+ "acc,exam_id__2016_2": 0.6422764227642277,
44
+ "acc,exam_id__2015": 0.680672268907563,
45
+ "acc,exam_id__2022": 0.6917293233082706,
46
+ "acc,exam_id__2016": 0.71900826446281
47
+ },
48
+ "faquad_nli": {
49
+ "f1_macro,all": 0.7691782381060779,
50
+ "acc,all": 0.82,
51
+ "alias": "faquad_nli"
52
+ },
53
+ "hatebr_offensive": {
54
+ "alias": "hatebr_offensive_binary",
55
+ "f1_macro,all": 0.8633798389731095,
56
+ "acc,all": 0.8642857142857143
57
+ },
58
+ "oab_exams": {
59
+ "acc,all": 0.475626423690205,
60
+ "acc,exam_id__2017-23": 0.4875,
61
+ "acc,exam_id__2017-22": 0.525,
62
+ "acc,exam_id__2014-14": 0.55,
63
+ "acc,exam_id__2012-06": 0.45,
64
+ "acc,exam_id__2011-05": 0.5,
65
+ "acc,exam_id__2012-09": 0.4025974025974026,
66
+ "acc,exam_id__2011-03": 0.41414141414141414,
67
+ "acc,exam_id__2012-08": 0.525,
68
+ "acc,exam_id__2010-02": 0.52,
69
+ "acc,exam_id__2012-06a": 0.4875,
70
+ "acc,exam_id__2014-13": 0.3875,
71
+ "acc,exam_id__2014-15": 0.5641025641025641,
72
+ "acc,exam_id__2015-16": 0.4375,
73
+ "acc,exam_id__2016-20a": 0.4125,
74
+ "acc,exam_id__2013-10": 0.4625,
75
+ "acc,exam_id__2010-01": 0.4588235294117647,
76
+ "acc,exam_id__2015-17": 0.6025641025641025,
77
+ "acc,exam_id__2013-12": 0.5625,
78
+ "acc,exam_id__2011-04": 0.5,
79
+ "acc,exam_id__2018-25": 0.475,
80
+ "acc,exam_id__2015-18": 0.3375,
81
+ "acc,exam_id__2016-21": 0.45,
82
+ "acc,exam_id__2012-07": 0.375,
83
+ "acc,exam_id__2017-24": 0.5,
84
+ "acc,exam_id__2016-19": 0.5256410256410257,
85
+ "acc,exam_id__2013-11": 0.5,
86
+ "acc,exam_id__2016-20": 0.4375,
87
+ "alias": "oab_exams"
88
+ },
89
+ "portuguese_hate_speech": {
90
+ "alias": "portuguese_hate_speech_binary",
91
+ "f1_macro,all": 0.6466753105050977,
92
+ "acc,all": 0.6556991774383079
93
+ },
94
+ "tweetsentbr": {
95
+ "f1_macro,all": 0.6968077814532404,
96
+ "acc,all": 0.7179104477611941,
97
+ "alias": "tweetsentbr"
98
+ }
99
+ },
100
+ "configs": {
101
+ "assin2_rte": {
102
+ "task": "assin2_rte",
103
+ "group": [
104
+ "pt_benchmark",
105
+ "assin2"
106
+ ],
107
+ "dataset_path": "assin2",
108
+ "test_split": "test",
109
+ "fewshot_split": "train",
110
+ "doc_to_text": "Premissa: {{premise}}\nHipótese: {{hypothesis}}\nPergunta: A hipótese pode ser inferida pela premissa? Sim ou Não?\nResposta:",
111
+ "doc_to_target": "{{['Não', 'Sim'][entailment_judgment]}}",
112
+ "description": "Abaixo estão pares de premissa e hipótese. Para cada par, indique se a hipótese pode ser inferida a partir da premissa, responda apenas com \"Sim\" ou \"Não\".\n\n",
113
+ "target_delimiter": " ",
114
+ "fewshot_delimiter": "\n\n",
115
+ "fewshot_config": {
116
+ "sampler": "id_sampler",
117
+ "sampler_config": {
118
+ "id_list": [
119
+ 1,
120
+ 3251,
121
+ 2,
122
+ 3252,
123
+ 3,
124
+ 4,
125
+ 5,
126
+ 6,
127
+ 3253,
128
+ 7,
129
+ 3254,
130
+ 3255,
131
+ 3256,
132
+ 8,
133
+ 9,
134
+ 10,
135
+ 3257,
136
+ 11,
137
+ 3258,
138
+ 12,
139
+ 13,
140
+ 14,
141
+ 15,
142
+ 3259,
143
+ 3260,
144
+ 3261,
145
+ 3262,
146
+ 3263,
147
+ 16,
148
+ 17,
149
+ 3264,
150
+ 18,
151
+ 3265,
152
+ 3266,
153
+ 3267,
154
+ 19,
155
+ 20,
156
+ 3268,
157
+ 3269,
158
+ 21,
159
+ 3270,
160
+ 3271,
161
+ 22,
162
+ 3272,
163
+ 3273,
164
+ 23,
165
+ 3274,
166
+ 24,
167
+ 25,
168
+ 3275
169
+ ],
170
+ "id_column": "sentence_pair_id"
171
+ }
172
+ },
173
+ "num_fewshot": 15,
174
+ "metric_list": [
175
+ {
176
+ "metric": "f1_macro",
177
+ "aggregation": "f1_macro",
178
+ "higher_is_better": true
179
+ },
180
+ {
181
+ "metric": "acc",
182
+ "aggregation": "acc",
183
+ "higher_is_better": true
184
+ }
185
+ ],
186
+ "output_type": "generate_until",
187
+ "generation_kwargs": {
188
+ "do_sample": false,
189
+ "temperature": 0.0,
190
+ "top_k": null,
191
+ "top_p": null,
192
+ "until": [
193
+ "\n\n"
194
+ ]
195
+ },
196
+ "repeats": 1,
197
+ "filter_list": [
198
+ {
199
+ "name": "all",
200
+ "filter": [
201
+ {
202
+ "function": "find_similar_label",
203
+ "labels": [
204
+ "Sim",
205
+ "Não"
206
+ ]
207
+ },
208
+ {
209
+ "function": "take_first"
210
+ }
211
+ ]
212
+ }
213
+ ],
214
+ "should_decontaminate": false,
215
+ "metadata": {
216
+ "version": 1.1
217
+ }
218
+ },
219
+ "assin2_sts": {
220
+ "task": "assin2_sts",
221
+ "group": [
222
+ "pt_benchmark",
223
+ "assin2"
224
+ ],
225
+ "dataset_path": "assin2",
226
+ "test_split": "test",
227
+ "fewshot_split": "train",
228
+ "doc_to_text": "Frase 1: {{premise}}\nFrase 2: {{hypothesis}}\nPergunta: Quão similares são as duas frases? Dê uma pontuação entre 1,0 a 5,0.\nResposta:",
229
+ "doc_to_target": "<function assin2_float_to_pt_str at 0x7ff8900d40e0>",
230
+ "description": "Abaixo estão pares de frases que você deve avaliar o grau de similaridade. Dê uma pontuação entre 1,0 e 5,0, sendo 1,0 pouco similar e 5,0 muito similar.\n\n",
231
+ "target_delimiter": " ",
232
+ "fewshot_delimiter": "\n\n",
233
+ "fewshot_config": {
234
+ "sampler": "id_sampler",
235
+ "sampler_config": {
236
+ "id_list": [
237
+ 1,
238
+ 3251,
239
+ 2,
240
+ 3252,
241
+ 3,
242
+ 4,
243
+ 5,
244
+ 6,
245
+ 3253,
246
+ 7,
247
+ 3254,
248
+ 3255,
249
+ 3256,
250
+ 8,
251
+ 9,
252
+ 10,
253
+ 3257,
254
+ 11,
255
+ 3258,
256
+ 12,
257
+ 13,
258
+ 14,
259
+ 15,
260
+ 3259,
261
+ 3260,
262
+ 3261,
263
+ 3262,
264
+ 3263,
265
+ 16,
266
+ 17,
267
+ 3264,
268
+ 18,
269
+ 3265,
270
+ 3266,
271
+ 3267,
272
+ 19,
273
+ 20,
274
+ 3268,
275
+ 3269,
276
+ 21,
277
+ 3270,
278
+ 3271,
279
+ 22,
280
+ 3272,
281
+ 3273,
282
+ 23,
283
+ 3274,
284
+ 24,
285
+ 25,
286
+ 3275
287
+ ],
288
+ "id_column": "sentence_pair_id"
289
+ }
290
+ },
291
+ "num_fewshot": 15,
292
+ "metric_list": [
293
+ {
294
+ "metric": "pearson",
295
+ "aggregation": "pearsonr",
296
+ "higher_is_better": true
297
+ },
298
+ {
299
+ "metric": "mse",
300
+ "aggregation": "mean_squared_error",
301
+ "higher_is_better": false
302
+ }
303
+ ],
304
+ "output_type": "generate_until",
305
+ "generation_kwargs": {
306
+ "do_sample": false,
307
+ "temperature": 0.0,
308
+ "top_k": null,
309
+ "top_p": null,
310
+ "until": [
311
+ "\n\n"
312
+ ]
313
+ },
314
+ "repeats": 1,
315
+ "filter_list": [
316
+ {
317
+ "name": "all",
318
+ "filter": [
319
+ {
320
+ "function": "number_filter",
321
+ "type": "float",
322
+ "range_min": 1.0,
323
+ "range_max": 5.0,
324
+ "on_outside_range": "clip",
325
+ "fallback": 5.0
326
+ },
327
+ {
328
+ "function": "take_first"
329
+ }
330
+ ]
331
+ }
332
+ ],
333
+ "should_decontaminate": false,
334
+ "metadata": {
335
+ "version": 1.1
336
+ }
337
+ },
338
+ "bluex": {
339
+ "task": "bluex",
340
+ "group": [
341
+ "pt_benchmark",
342
+ "vestibular"
343
+ ],
344
+ "dataset_path": "eduagarcia-temp/BLUEX_without_images",
345
+ "test_split": "train",
346
+ "fewshot_split": "train",
347
+ "doc_to_text": "<function enem_doc_to_text at 0x7ff89009f9c0>",
348
+ "doc_to_target": "{{answerKey}}",
349
+ "description": "As perguntas a seguir são questões de múltipla escolha de provas de vestibular de universidades brasileiras, selecione a única alternativa correta e responda apenas com as letras \"A\", \"B\", \"C\", \"D\" ou \"E\".\n\n",
350
+ "target_delimiter": " ",
351
+ "fewshot_delimiter": "\n\n",
352
+ "fewshot_config": {
353
+ "sampler": "id_sampler",
354
+ "sampler_config": {
355
+ "id_list": [
356
+ "USP_2018_3",
357
+ "UNICAMP_2018_2",
358
+ "USP_2018_35",
359
+ "UNICAMP_2018_16",
360
+ "USP_2018_89"
361
+ ],
362
+ "id_column": "id",
363
+ "exclude_from_task": true
364
+ }
365
+ },
366
+ "num_fewshot": 3,
367
+ "metric_list": [
368
+ {
369
+ "metric": "acc",
370
+ "aggregation": "acc",
371
+ "higher_is_better": true
372
+ }
373
+ ],
374
+ "output_type": "generate_until",
375
+ "generation_kwargs": {
376
+ "do_sample": false,
377
+ "temperature": 0.0,
378
+ "top_k": null,
379
+ "top_p": null,
380
+ "until": [
381
+ "\n\n"
382
+ ]
383
+ },
384
+ "repeats": 1,
385
+ "filter_list": [
386
+ {
387
+ "name": "all",
388
+ "filter": [
389
+ {
390
+ "function": "normalize_spaces"
391
+ },
392
+ {
393
+ "function": "remove_accents"
394
+ },
395
+ {
396
+ "function": "find_choices",
397
+ "choices": [
398
+ "A",
399
+ "B",
400
+ "C",
401
+ "D",
402
+ "E"
403
+ ],
404
+ "regex_patterns": [
405
+ "(?:[Ll]etra|[Aa]lternativa|[Rr]esposta|[Rr]esposta [Cc]orreta|[Rr]esposta [Cc]orreta e|[Oo]pcao):? ([ABCDE])\\b",
406
+ "\\b([ABCDE])\\.",
407
+ "\\b([ABCDE]) ?[.):-]",
408
+ "\\b([ABCDE])$",
409
+ "\\b([ABCDE])\\b"
410
+ ]
411
+ },
412
+ {
413
+ "function": "take_first"
414
+ }
415
+ ],
416
+ "group_by": {
417
+ "column": "exam_id"
418
+ }
419
+ }
420
+ ],
421
+ "should_decontaminate": true,
422
+ "doc_to_decontamination_query": "<function enem_doc_to_text at 0x7ff89009fce0>",
423
+ "metadata": {
424
+ "version": 1.1
425
+ }
426
+ },
427
+ "enem_challenge": {
428
+ "task": "enem_challenge",
429
+ "task_alias": "enem",
430
+ "group": [
431
+ "pt_benchmark",
432
+ "vestibular"
433
+ ],
434
+ "dataset_path": "eduagarcia/enem_challenge",
435
+ "test_split": "train",
436
+ "fewshot_split": "train",
437
+ "doc_to_text": "<function enem_doc_to_text at 0x7ff8900d4540>",
438
+ "doc_to_target": "{{answerKey}}",
439
+ "description": "As perguntas a seguir são questões de múltipla escolha do Exame Nacional do Ensino Médio (ENEM), selecione a única alternativa correta e responda apenas com as letras \"A\", \"B\", \"C\", \"D\" ou \"E\".\n\n",
440
+ "target_delimiter": " ",
441
+ "fewshot_delimiter": "\n\n",
442
+ "fewshot_config": {
443
+ "sampler": "id_sampler",
444
+ "sampler_config": {
445
+ "id_list": [
446
+ "2022_21",
447
+ "2022_88",
448
+ "2022_143"
449
+ ],
450
+ "id_column": "id",
451
+ "exclude_from_task": true
452
+ }
453
+ },
454
+ "num_fewshot": 3,
455
+ "metric_list": [
456
+ {
457
+ "metric": "acc",
458
+ "aggregation": "acc",
459
+ "higher_is_better": true
460
+ }
461
+ ],
462
+ "output_type": "generate_until",
463
+ "generation_kwargs": {
464
+ "do_sample": false,
465
+ "temperature": 0.0,
466
+ "top_k": null,
467
+ "top_p": null,
468
+ "until": [
469
+ "\n\n"
470
+ ]
471
+ },
472
+ "repeats": 1,
473
+ "filter_list": [
474
+ {
475
+ "name": "all",
476
+ "filter": [
477
+ {
478
+ "function": "normalize_spaces"
479
+ },
480
+ {
481
+ "function": "remove_accents"
482
+ },
483
+ {
484
+ "function": "find_choices",
485
+ "choices": [
486
+ "A",
487
+ "B",
488
+ "C",
489
+ "D",
490
+ "E"
491
+ ],
492
+ "regex_patterns": [
493
+ "(?:[Ll]etra|[Aa]lternativa|[Rr]esposta|[Rr]esposta [Cc]orreta|[Rr]esposta [Cc]orreta e|[Oo]pcao):? ([ABCDE])\\b",
494
+ "\\b([ABCDE])\\.",
495
+ "\\b([ABCDE]) ?[.):-]",
496
+ "\\b([ABCDE])$",
497
+ "\\b([ABCDE])\\b"
498
+ ]
499
+ },
500
+ {
501
+ "function": "take_first"
502
+ }
503
+ ],
504
+ "group_by": {
505
+ "column": "exam_id"
506
+ }
507
+ }
508
+ ],
509
+ "should_decontaminate": true,
510
+ "doc_to_decontamination_query": "<function enem_doc_to_text at 0x7ff8900d4860>",
511
+ "metadata": {
512
+ "version": 1.1
513
+ }
514
+ },
515
+ "faquad_nli": {
516
+ "task": "faquad_nli",
517
+ "group": [
518
+ "pt_benchmark"
519
+ ],
520
+ "dataset_path": "ruanchaves/faquad-nli",
521
+ "test_split": "test",
522
+ "fewshot_split": "train",
523
+ "doc_to_text": "Pergunta: {{question}}\nResposta: {{answer}}\nA resposta dada satisfaz à pergunta? Sim ou Não?",
524
+ "doc_to_target": "{{['Não', 'Sim'][label]}}",
525
+ "description": "Abaixo estão pares de pergunta e resposta. Para cada par, você deve julgar se a resposta responde à pergunta de maneira satisfatória e aparenta estar correta. Escreva apenas \"Sim\" ou \"Não\".\n\n",
526
+ "target_delimiter": " ",
527
+ "fewshot_delimiter": "\n\n",
528
+ "fewshot_config": {
529
+ "sampler": "first_n",
530
+ "sampler_config": {
531
+ "fewshot_indices": [
532
+ 1893,
533
+ 949,
534
+ 663,
535
+ 105,
536
+ 1169,
537
+ 2910,
538
+ 2227,
539
+ 2813,
540
+ 974,
541
+ 558,
542
+ 1503,
543
+ 1958,
544
+ 2918,
545
+ 601,
546
+ 1560,
547
+ 984,
548
+ 2388,
549
+ 995,
550
+ 2233,
551
+ 1982,
552
+ 165,
553
+ 2788,
554
+ 1312,
555
+ 2285,
556
+ 522,
557
+ 1113,
558
+ 1670,
559
+ 323,
560
+ 236,
561
+ 1263,
562
+ 1562,
563
+ 2519,
564
+ 1049,
565
+ 432,
566
+ 1167,
567
+ 1394,
568
+ 2022,
569
+ 2551,
570
+ 2194,
571
+ 2187,
572
+ 2282,
573
+ 2816,
574
+ 108,
575
+ 301,
576
+ 1185,
577
+ 1315,
578
+ 1420,
579
+ 2436,
580
+ 2322,
581
+ 766
582
+ ]
583
+ }
584
+ },
585
+ "num_fewshot": 15,
586
+ "metric_list": [
587
+ {
588
+ "metric": "f1_macro",
589
+ "aggregation": "f1_macro",
590
+ "higher_is_better": true
591
+ },
592
+ {
593
+ "metric": "acc",
594
+ "aggregation": "acc",
595
+ "higher_is_better": true
596
+ }
597
+ ],
598
+ "output_type": "generate_until",
599
+ "generation_kwargs": {
600
+ "do_sample": false,
601
+ "temperature": 0.0,
602
+ "top_k": null,
603
+ "top_p": null,
604
+ "until": [
605
+ "\n\n"
606
+ ]
607
+ },
608
+ "repeats": 1,
609
+ "filter_list": [
610
+ {
611
+ "name": "all",
612
+ "filter": [
613
+ {
614
+ "function": "find_similar_label",
615
+ "labels": [
616
+ "Sim",
617
+ "Não"
618
+ ]
619
+ },
620
+ {
621
+ "function": "take_first"
622
+ }
623
+ ]
624
+ }
625
+ ],
626
+ "should_decontaminate": false,
627
+ "metadata": {
628
+ "version": 1.1
629
+ }
630
+ },
631
+ "hatebr_offensive": {
632
+ "task": "hatebr_offensive",
633
+ "task_alias": "hatebr_offensive_binary",
634
+ "group": [
635
+ "pt_benchmark"
636
+ ],
637
+ "dataset_path": "eduagarcia/portuguese_benchmark",
638
+ "dataset_name": "HateBR_offensive_binary",
639
+ "test_split": "test",
640
+ "fewshot_split": "train",
641
+ "doc_to_text": "Texto: {{sentence}}\nPergunta: O texto é ofensivo?\nResposta:",
642
+ "doc_to_target": "{{'Sim' if label == 1 else 'Não'}}",
643
+ "description": "Abaixo contém o texto de comentários de usuários do Instagram em português, sua tarefa é classificar se o texto é ofensivo ou não. Responda apenas com \"Sim\" ou \"Não\".\n\n",
644
+ "target_delimiter": " ",
645
+ "fewshot_delimiter": "\n\n",
646
+ "fewshot_config": {
647
+ "sampler": "id_sampler",
648
+ "sampler_config": {
649
+ "id_list": [
650
+ 48,
651
+ 44,
652
+ 36,
653
+ 20,
654
+ 3511,
655
+ 88,
656
+ 3555,
657
+ 16,
658
+ 56,
659
+ 3535,
660
+ 60,
661
+ 40,
662
+ 3527,
663
+ 4,
664
+ 76,
665
+ 3579,
666
+ 3523,
667
+ 3551,
668
+ 68,
669
+ 3503,
670
+ 84,
671
+ 3539,
672
+ 64,
673
+ 3599,
674
+ 80,
675
+ 3563,
676
+ 3559,
677
+ 3543,
678
+ 3547,
679
+ 3587,
680
+ 3595,
681
+ 3575,
682
+ 3567,
683
+ 3591,
684
+ 24,
685
+ 96,
686
+ 92,
687
+ 3507,
688
+ 52,
689
+ 72,
690
+ 8,
691
+ 3571,
692
+ 3515,
693
+ 3519,
694
+ 3531,
695
+ 28,
696
+ 32,
697
+ 0,
698
+ 12,
699
+ 3583
700
+ ],
701
+ "id_column": "idx"
702
+ }
703
+ },
704
+ "num_fewshot": 25,
705
+ "metric_list": [
706
+ {
707
+ "metric": "f1_macro",
708
+ "aggregation": "f1_macro",
709
+ "higher_is_better": true
710
+ },
711
+ {
712
+ "metric": "acc",
713
+ "aggregation": "acc",
714
+ "higher_is_better": true
715
+ }
716
+ ],
717
+ "output_type": "generate_until",
718
+ "generation_kwargs": {
719
+ "do_sample": false,
720
+ "temperature": 0.0,
721
+ "top_k": null,
722
+ "top_p": null,
723
+ "until": [
724
+ "\n\n"
725
+ ]
726
+ },
727
+ "repeats": 1,
728
+ "filter_list": [
729
+ {
730
+ "name": "all",
731
+ "filter": [
732
+ {
733
+ "function": "find_similar_label",
734
+ "labels": [
735
+ "Sim",
736
+ "Não"
737
+ ]
738
+ },
739
+ {
740
+ "function": "take_first"
741
+ }
742
+ ]
743
+ }
744
+ ],
745
+ "should_decontaminate": false,
746
+ "metadata": {
747
+ "version": 1.0
748
+ }
749
+ },
750
+ "oab_exams": {
751
+ "task": "oab_exams",
752
+ "group": [
753
+ "legal_benchmark",
754
+ "pt_benchmark"
755
+ ],
756
+ "dataset_path": "eduagarcia/oab_exams",
757
+ "test_split": "train",
758
+ "fewshot_split": "train",
759
+ "doc_to_text": "<function doc_to_text at 0x7ff89009f380>",
760
+ "doc_to_target": "{{answerKey}}",
761
+ "description": "As perguntas a seguir são questões de múltipla escolha do Exame de Ordem da Ordem dos Advogados do Brasil (OAB), selecione a única alternativa correta e responda apenas com as letras \"A\", \"B\", \"C\" ou \"D\".\n\n",
762
+ "target_delimiter": " ",
763
+ "fewshot_delimiter": "\n\n",
764
+ "fewshot_config": {
765
+ "sampler": "id_sampler",
766
+ "sampler_config": {
767
+ "id_list": [
768
+ "2010-01_1",
769
+ "2010-01_11",
770
+ "2010-01_13",
771
+ "2010-01_23",
772
+ "2010-01_26",
773
+ "2010-01_28",
774
+ "2010-01_38",
775
+ "2010-01_48",
776
+ "2010-01_58",
777
+ "2010-01_68",
778
+ "2010-01_76",
779
+ "2010-01_83",
780
+ "2010-01_85",
781
+ "2010-01_91",
782
+ "2010-01_99"
783
+ ],
784
+ "id_column": "id",
785
+ "exclude_from_task": true
786
+ }
787
+ },
788
+ "num_fewshot": 3,
789
+ "metric_list": [
790
+ {
791
+ "metric": "acc",
792
+ "aggregation": "acc",
793
+ "higher_is_better": true
794
+ }
795
+ ],
796
+ "output_type": "generate_until",
797
+ "generation_kwargs": {
798
+ "max_gen_toks": 32,
799
+ "do_sample": false,
800
+ "temperature": 0.0,
801
+ "top_k": null,
802
+ "top_p": null,
803
+ "until": [
804
+ "\n\n"
805
+ ]
806
+ },
807
+ "repeats": 1,
808
+ "filter_list": [
809
+ {
810
+ "name": "all",
811
+ "filter": [
812
+ {
813
+ "function": "normalize_spaces"
814
+ },
815
+ {
816
+ "function": "remove_accents"
817
+ },
818
+ {
819
+ "function": "find_choices",
820
+ "choices": [
821
+ "A",
822
+ "B",
823
+ "C",
824
+ "D"
825
+ ],
826
+ "regex_patterns": [
827
+ "(?:[Ll]etra|[Aa]lternativa|[Rr]esposta|[Rr]esposta [Cc]orreta|[Rr]esposta [Cc]orreta e|[Oo]pcao):? ([ABCD])\\b",
828
+ "\\b([ABCD])\\.",
829
+ "\\b([ABCD]) ?[.):-]",
830
+ "\\b([ABCD])$",
831
+ "\\b([ABCD])\\b"
832
+ ]
833
+ },
834
+ {
835
+ "function": "take_first"
836
+ }
837
+ ],
838
+ "group_by": {
839
+ "column": "exam_id"
840
+ }
841
+ }
842
+ ],
843
+ "should_decontaminate": true,
844
+ "doc_to_decontamination_query": "<function doc_to_text at 0x7ff89009f4c0>",
845
+ "metadata": {
846
+ "version": 1.5
847
+ }
848
+ },
849
+ "portuguese_hate_speech": {
850
+ "task": "portuguese_hate_speech",
851
+ "task_alias": "portuguese_hate_speech_binary",
852
+ "group": [
853
+ "pt_benchmark"
854
+ ],
855
+ "dataset_path": "eduagarcia/portuguese_benchmark",
856
+ "dataset_name": "Portuguese_Hate_Speech_binary",
857
+ "test_split": "test",
858
+ "fewshot_split": "train",
859
+ "doc_to_text": "Texto: {{sentence}}\nPergunta: O texto contém discurso de ódio?\nResposta:",
860
+ "doc_to_target": "{{'Sim' if label == 1 else 'Não'}}",
861
+ "description": "Abaixo contém o texto de tweets de usuários do Twitter em português, sua tarefa é classificar se o texto contém discurso de ódio ou não. Responda apenas com \"Sim\" ou \"Não\".\n\n",
862
+ "target_delimiter": " ",
863
+ "fewshot_delimiter": "\n\n",
864
+ "fewshot_config": {
865
+ "sampler": "id_sampler",
866
+ "sampler_config": {
867
+ "id_list": [
868
+ 52,
869
+ 50,
870
+ 39,
871
+ 28,
872
+ 3,
873
+ 105,
874
+ 22,
875
+ 25,
876
+ 60,
877
+ 11,
878
+ 66,
879
+ 41,
880
+ 9,
881
+ 4,
882
+ 91,
883
+ 42,
884
+ 7,
885
+ 20,
886
+ 76,
887
+ 1,
888
+ 104,
889
+ 13,
890
+ 67,
891
+ 54,
892
+ 97,
893
+ 27,
894
+ 24,
895
+ 14,
896
+ 16,
897
+ 48,
898
+ 53,
899
+ 40,
900
+ 34,
901
+ 49,
902
+ 32,
903
+ 119,
904
+ 114,
905
+ 2,
906
+ 58,
907
+ 83,
908
+ 18,
909
+ 36,
910
+ 5,
911
+ 6,
912
+ 10,
913
+ 35,
914
+ 38,
915
+ 0,
916
+ 21,
917
+ 46
918
+ ],
919
+ "id_column": "idx"
920
+ }
921
+ },
922
+ "num_fewshot": 25,
923
+ "metric_list": [
924
+ {
925
+ "metric": "f1_macro",
926
+ "aggregation": "f1_macro",
927
+ "higher_is_better": true
928
+ },
929
+ {
930
+ "metric": "acc",
931
+ "aggregation": "acc",
932
+ "higher_is_better": true
933
+ }
934
+ ],
935
+ "output_type": "generate_until",
936
+ "generation_kwargs": {
937
+ "do_sample": false,
938
+ "temperature": 0.0,
939
+ "top_k": null,
940
+ "top_p": null,
941
+ "until": [
942
+ "\n\n"
943
+ ]
944
+ },
945
+ "repeats": 1,
946
+ "filter_list": [
947
+ {
948
+ "name": "all",
949
+ "filter": [
950
+ {
951
+ "function": "find_similar_label",
952
+ "labels": [
953
+ "Sim",
954
+ "Não"
955
+ ]
956
+ },
957
+ {
958
+ "function": "take_first"
959
+ }
960
+ ]
961
+ }
962
+ ],
963
+ "should_decontaminate": false,
964
+ "metadata": {
965
+ "version": 1.0
966
+ }
967
+ },
968
+ "tweetsentbr": {
969
+ "task": "tweetsentbr",
970
+ "group": [
971
+ "pt_benchmark"
972
+ ],
973
+ "dataset_path": "eduagarcia/tweetsentbr_fewshot",
974
+ "test_split": "test",
975
+ "fewshot_split": "train",
976
+ "doc_to_text": "Texto: {{sentence}}\nPergunta: O sentimento do texto é Positivo, Neutro ou Negativo?\nResposta:",
977
+ "doc_to_target": "{{'Positivo' if label == 'Positive' else ('Negativo' if label == 'Negative' else 'Neutro')}}",
978
+ "description": "Abaixo contém o texto de tweets de usuários do Twitter em português, sua tarefa é classificar se o sentimento do texto é Positivo, Neutro ou Negativo. Responda apenas com uma das opções.\n\n",
979
+ "target_delimiter": " ",
980
+ "fewshot_delimiter": "\n\n",
981
+ "fewshot_config": {
982
+ "sampler": "first_n"
983
+ },
984
+ "num_fewshot": 25,
985
+ "metric_list": [
986
+ {
987
+ "metric": "f1_macro",
988
+ "aggregation": "f1_macro",
989
+ "higher_is_better": true
990
+ },
991
+ {
992
+ "metric": "acc",
993
+ "aggregation": "acc",
994
+ "higher_is_better": true
995
+ }
996
+ ],
997
+ "output_type": "generate_until",
998
+ "generation_kwargs": {
999
+ "do_sample": false,
1000
+ "temperature": 0.0,
1001
+ "top_k": null,
1002
+ "top_p": null,
1003
+ "until": [
1004
+ "\n\n"
1005
+ ]
1006
+ },
1007
+ "repeats": 1,
1008
+ "filter_list": [
1009
+ {
1010
+ "name": "all",
1011
+ "filter": [
1012
+ {
1013
+ "function": "find_similar_label",
1014
+ "labels": [
1015
+ "Positivo",
1016
+ "Neutro",
1017
+ "Negativo"
1018
+ ]
1019
+ },
1020
+ {
1021
+ "function": "take_first"
1022
+ }
1023
+ ]
1024
+ }
1025
+ ],
1026
+ "should_decontaminate": false,
1027
+ "metadata": {
1028
+ "version": 1.0
1029
+ }
1030
+ }
1031
+ },
1032
+ "versions": {
1033
+ "assin2_rte": 1.1,
1034
+ "assin2_sts": 1.1,
1035
+ "bluex": 1.1,
1036
+ "enem_challenge": 1.1,
1037
+ "faquad_nli": 1.1,
1038
+ "hatebr_offensive": 1.0,
1039
+ "oab_exams": 1.5,
1040
+ "portuguese_hate_speech": 1.0,
1041
+ "tweetsentbr": 1.0
1042
+ },
1043
+ "n-shot": {
1044
+ "assin2_rte": 15,
1045
+ "assin2_sts": 15,
1046
+ "bluex": 3,
1047
+ "enem_challenge": 3,
1048
+ "faquad_nli": 15,
1049
+ "hatebr_offensive": 25,
1050
+ "oab_exams": 3,
1051
+ "portuguese_hate_speech": 25,
1052
+ "tweetsentbr": 25
1053
+ },
1054
+ "model_meta": {
1055
+ "truncated": 115,
1056
+ "non_truncated": 14035,
1057
+ "padded": 6559,
1058
+ "non_padded": 7591,
1059
+ "fewshots_truncated": 134,
1060
+ "has_chat_template": true,
1061
+ "chat_type": "system_user_assistant",
1062
+ "n_gpus": 1,
1063
+ "accelerate_num_process": null,
1064
+ "model_sha": "a42c86c61b98ca4fdf238d688fe6ea11cf414d29",
1065
+ "model_dtype": "torch.bfloat16",
1066
+ "model_memory_footprint": 17658814720,
1067
+ "model_num_parameters": 8829407232,
1068
+ "model_is_loaded_in_4bit": null,
1069
+ "model_is_loaded_in_8bit": null,
1070
+ "model_is_quantized": null,
1071
+ "model_device": "cuda:0",
1072
+ "batch_size": 32,
1073
+ "max_length": 2560,
1074
+ "max_ctx_length": 2528,
1075
+ "max_gen_toks": 32,
1076
+ "effective_batch_size": 31.941309255079005
1077
+ },
1078
+ "task_model_meta": {
1079
+ "assin2_rte": {
1080
+ "sample_size": 2448,
1081
+ "truncated": 0,
1082
+ "non_truncated": 2448,
1083
+ "padded": 216,
1084
+ "non_padded": 2232,
1085
+ "fewshots_truncated": 0,
1086
+ "mean_seq_length": 1619.5996732026144,
1087
+ "min_seq_length": 1595,
1088
+ "max_seq_length": 1693,
1089
+ "max_ctx_length": 2304,
1090
+ "max_gen_toks": 256,
1091
+ "mean_original_fewshots_size": 15.0,
1092
+ "mean_effective_fewshot_size": 15.0
1093
+ },
1094
+ "assin2_sts": {
1095
+ "sample_size": 2448,
1096
+ "truncated": 0,
1097
+ "non_truncated": 2448,
1098
+ "padded": 250,
1099
+ "non_padded": 2198,
1100
+ "fewshots_truncated": 0,
1101
+ "mean_seq_length": 1873.5996732026144,
1102
+ "min_seq_length": 1849,
1103
+ "max_seq_length": 1947,
1104
+ "max_ctx_length": 2304,
1105
+ "max_gen_toks": 256,
1106
+ "mean_original_fewshots_size": 15.0,
1107
+ "mean_effective_fewshot_size": 15.0
1108
+ },
1109
+ "bluex": {
1110
+ "sample_size": 719,
1111
+ "truncated": 65,
1112
+ "non_truncated": 654,
1113
+ "padded": 478,
1114
+ "non_padded": 241,
1115
+ "fewshots_truncated": 82,
1116
+ "mean_seq_length": 1987.8052851182197,
1117
+ "min_seq_length": 1559,
1118
+ "max_seq_length": 2891,
1119
+ "max_ctx_length": 2304,
1120
+ "max_gen_toks": 256,
1121
+ "mean_original_fewshots_size": 3.0,
1122
+ "mean_effective_fewshot_size": 2.885952712100139
1123
+ },
1124
+ "enem_challenge": {
1125
+ "sample_size": 1429,
1126
+ "truncated": 50,
1127
+ "non_truncated": 1379,
1128
+ "padded": 940,
1129
+ "non_padded": 489,
1130
+ "fewshots_truncated": 52,
1131
+ "mean_seq_length": 1919.8929321203639,
1132
+ "min_seq_length": 1608,
1133
+ "max_seq_length": 2922,
1134
+ "max_ctx_length": 2304,
1135
+ "max_gen_toks": 256,
1136
+ "mean_original_fewshots_size": 3.0,
1137
+ "mean_effective_fewshot_size": 2.9636109167249827
1138
+ },
1139
+ "faquad_nli": {
1140
+ "sample_size": 650,
1141
+ "truncated": 0,
1142
+ "non_truncated": 650,
1143
+ "padded": 488,
1144
+ "non_padded": 162,
1145
+ "fewshots_truncated": 0,
1146
+ "mean_seq_length": 1939.183076923077,
1147
+ "min_seq_length": 1877,
1148
+ "max_seq_length": 2071,
1149
+ "max_ctx_length": 2304,
1150
+ "max_gen_toks": 256,
1151
+ "mean_original_fewshots_size": 15.0,
1152
+ "mean_effective_fewshot_size": 15.0
1153
+ },
1154
+ "hatebr_offensive": {
1155
+ "sample_size": 1400,
1156
+ "truncated": 0,
1157
+ "non_truncated": 1400,
1158
+ "padded": 810,
1159
+ "non_padded": 590,
1160
+ "fewshots_truncated": 0,
1161
+ "mean_seq_length": 1640.8,
1162
+ "min_seq_length": 1612,
1163
+ "max_seq_length": 1927,
1164
+ "max_ctx_length": 2304,
1165
+ "max_gen_toks": 256,
1166
+ "mean_original_fewshots_size": 25.0,
1167
+ "mean_effective_fewshot_size": 25.0
1168
+ },
1169
+ "oab_exams": {
1170
+ "sample_size": 2195,
1171
+ "truncated": 0,
1172
+ "non_truncated": 2195,
1173
+ "padded": 1987,
1174
+ "non_padded": 208,
1175
+ "fewshots_truncated": 0,
1176
+ "mean_seq_length": 1631.1088838268793,
1177
+ "min_seq_length": 1318,
1178
+ "max_seq_length": 2189,
1179
+ "max_ctx_length": 2528,
1180
+ "max_gen_toks": 32,
1181
+ "mean_original_fewshots_size": 3.0,
1182
+ "mean_effective_fewshot_size": 3.0
1183
+ },
1184
+ "portuguese_hate_speech": {
1185
+ "sample_size": 851,
1186
+ "truncated": 0,
1187
+ "non_truncated": 851,
1188
+ "padded": 633,
1189
+ "non_padded": 218,
1190
+ "fewshots_truncated": 0,
1191
+ "mean_seq_length": 2182.0340775558166,
1192
+ "min_seq_length": 2143,
1193
+ "max_seq_length": 2218,
1194
+ "max_ctx_length": 2304,
1195
+ "max_gen_toks": 256,
1196
+ "mean_original_fewshots_size": 25.0,
1197
+ "mean_effective_fewshot_size": 25.0
1198
+ },
1199
+ "tweetsentbr": {
1200
+ "sample_size": 2010,
1201
+ "truncated": 0,
1202
+ "non_truncated": 2010,
1203
+ "padded": 757,
1204
+ "non_padded": 1253,
1205
+ "fewshots_truncated": 0,
1206
+ "mean_seq_length": 1881.9114427860698,
1207
+ "min_seq_length": 1859,
1208
+ "max_seq_length": 1997,
1209
+ "max_ctx_length": 2304,
1210
+ "max_gen_toks": 256,
1211
+ "mean_original_fewshots_size": 25.0,
1212
+ "mean_effective_fewshot_size": 25.0
1213
+ }
1214
+ },
1215
+ "config": {
1216
+ "model": "huggingface",
1217
+ "model_args": "pretrained=BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B,dtype=bfloat16,device=cuda:0,revision=main,trust_remote_code=True,starting_max_length=2560",
1218
+ "batch_size": "auto",
1219
+ "batch_sizes": [],
1220
+ "device": null,
1221
+ "use_cache": null,
1222
+ "limit": [
1223
+ null,
1224
+ null,
1225
+ null,
1226
+ null,
1227
+ null,
1228
+ null,
1229
+ null,
1230
+ null,
1231
+ null
1232
+ ],
1233
+ "bootstrap_iters": 0,
1234
+ "gen_kwargs": null
1235
+ },
1236
+ "git_hash": null
1237
+ }
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B/results_2025-05-04T10-35-32.472663.json ADDED
@@ -0,0 +1,326 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config_general": {
3
+ "start_date": "2025-05-04T10-35-32.472663",
4
+ "start_time": 1746354934.248489,
5
+ "end_time": 1746358687.1596806,
6
+ "total_evaluation_time_seconds": 3752.911191701889,
7
+ "has_chat_template": true,
8
+ "chat_type": "system_user_assistant",
9
+ "n_gpus": 1,
10
+ "accelerate_num_process": null,
11
+ "model_sha": "a42c86c61b98ca4fdf238d688fe6ea11cf414d29",
12
+ "model_dtype": "bfloat16",
13
+ "model_memory_footprint": 17658814720,
14
+ "model_num_parameters": 8829407232,
15
+ "model_is_loaded_in_4bit": null,
16
+ "model_is_loaded_in_8bit": null,
17
+ "model_is_quantized": null,
18
+ "model_device": "cuda:0",
19
+ "batch_size": 32,
20
+ "max_length": 2560,
21
+ "max_ctx_length": 2528,
22
+ "max_gen_toks": 32,
23
+ "effective_batch_size": 31.941309255079005,
24
+ "model_name": "BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B",
25
+ "job_id": 1577,
26
+ "model_id": "BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B_eval_request_False_bfloat16_Original",
27
+ "model_base_model": "",
28
+ "model_weight_type": "Original",
29
+ "model_revision": "main",
30
+ "model_private": false,
31
+ "model_type": "💬 : chat (RLHF, DPO, IFT, ...)",
32
+ "model_architectures": "LlamaForCausalLM",
33
+ "submitted_time": "2024-07-18T22:23:09Z",
34
+ "lm_eval_model_type": "huggingface",
35
+ "eval_version": "1.1.0"
36
+ },
37
+ "results": {
38
+ "all_grouped_average": 0.7065741404156267,
39
+ "all_grouped_npm": 0.5637577452168115,
40
+ "all_grouped": {
41
+ "enem_challenge": 0.6913925822253324,
42
+ "bluex": 0.5841446453407511,
43
+ "oab_exams": 0.475626423690205,
44
+ "assin2_rte": 0.8902591771846999,
45
+ "assin2_sts": 0.7417032662621263,
46
+ "faquad_nli": 0.7691782381060779,
47
+ "hatebr_offensive": 0.8633798389731095,
48
+ "portuguese_hate_speech": 0.6466753105050977,
49
+ "tweetsentbr": 0.6968077814532404
50
+ },
51
+ "all": {
52
+ "harness|enem_challenge|enem_challenge|None|3": 0.6913925822253324,
53
+ "harness|bluex|bluex|None|3": 0.5841446453407511,
54
+ "harness|oab_exams|oab_exams|None|3": 0.475626423690205,
55
+ "harness|assin2_rte|assin2_rte|None|15": 0.8902591771846999,
56
+ "harness|assin2_sts|assin2_sts|None|15": 0.7417032662621263,
57
+ "harness|faquad_nli|faquad_nli|None|15": 0.7691782381060779,
58
+ "harness|hatebr_offensive|hatebr_offensive|None|25": 0.8633798389731095,
59
+ "harness|portuguese_hate_speech|portuguese_hate_speech|None|25": 0.6466753105050977,
60
+ "harness|tweetsentbr|tweetsentbr|None|25": 0.6968077814532404
61
+ },
62
+ "harness|enem_challenge|enem_challenge|None|3": {
63
+ "acc,all": 0.6913925822253324,
64
+ "acc,exam_id__2011": 0.7521367521367521,
65
+ "acc,exam_id__2012": 0.6896551724137931,
66
+ "acc,exam_id__2014": 0.7064220183486238,
67
+ "acc,exam_id__2023": 0.6518518518518519,
68
+ "acc,exam_id__2010": 0.7094017094017094,
69
+ "acc,exam_id__2009": 0.6956521739130435,
70
+ "acc,exam_id__2013": 0.6851851851851852,
71
+ "acc,exam_id__2017": 0.6810344827586207,
72
+ "acc,exam_id__2016_2": 0.6422764227642277,
73
+ "acc,exam_id__2015": 0.680672268907563,
74
+ "acc,exam_id__2022": 0.6917293233082706,
75
+ "acc,exam_id__2016": 0.71900826446281,
76
+ "main_score": 0.6913925822253324
77
+ },
78
+ "harness|bluex|bluex|None|3": {
79
+ "acc,all": 0.5841446453407511,
80
+ "acc,exam_id__USP_2021": 0.5384615384615384,
81
+ "acc,exam_id__USP_2020": 0.5535714285714286,
82
+ "acc,exam_id__USP_2019": 0.55,
83
+ "acc,exam_id__UNICAMP_2019": 0.6,
84
+ "acc,exam_id__UNICAMP_2021_2": 0.5686274509803921,
85
+ "acc,exam_id__USP_2023": 0.75,
86
+ "acc,exam_id__UNICAMP_2024": 0.4666666666666667,
87
+ "acc,exam_id__UNICAMP_2018": 0.5,
88
+ "acc,exam_id__USP_2024": 0.7560975609756098,
89
+ "acc,exam_id__USP_2018": 0.48148148148148145,
90
+ "acc,exam_id__USP_2022": 0.5714285714285714,
91
+ "acc,exam_id__UNICAMP_2022": 0.5641025641025641,
92
+ "acc,exam_id__UNICAMP_2023": 0.7209302325581395,
93
+ "acc,exam_id__UNICAMP_2020": 0.5636363636363636,
94
+ "acc,exam_id__UNICAMP_2021_1": 0.6521739130434783,
95
+ "main_score": 0.5841446453407511
96
+ },
97
+ "harness|oab_exams|oab_exams|None|3": {
98
+ "acc,all": 0.475626423690205,
99
+ "acc,exam_id__2017-23": 0.4875,
100
+ "acc,exam_id__2017-22": 0.525,
101
+ "acc,exam_id__2014-14": 0.55,
102
+ "acc,exam_id__2012-06": 0.45,
103
+ "acc,exam_id__2011-05": 0.5,
104
+ "acc,exam_id__2012-09": 0.4025974025974026,
105
+ "acc,exam_id__2011-03": 0.41414141414141414,
106
+ "acc,exam_id__2012-08": 0.525,
107
+ "acc,exam_id__2010-02": 0.52,
108
+ "acc,exam_id__2012-06a": 0.4875,
109
+ "acc,exam_id__2014-13": 0.3875,
110
+ "acc,exam_id__2014-15": 0.5641025641025641,
111
+ "acc,exam_id__2015-16": 0.4375,
112
+ "acc,exam_id__2016-20a": 0.4125,
113
+ "acc,exam_id__2013-10": 0.4625,
114
+ "acc,exam_id__2010-01": 0.4588235294117647,
115
+ "acc,exam_id__2015-17": 0.6025641025641025,
116
+ "acc,exam_id__2013-12": 0.5625,
117
+ "acc,exam_id__2011-04": 0.5,
118
+ "acc,exam_id__2018-25": 0.475,
119
+ "acc,exam_id__2015-18": 0.3375,
120
+ "acc,exam_id__2016-21": 0.45,
121
+ "acc,exam_id__2012-07": 0.375,
122
+ "acc,exam_id__2017-24": 0.5,
123
+ "acc,exam_id__2016-19": 0.5256410256410257,
124
+ "acc,exam_id__2013-11": 0.5,
125
+ "acc,exam_id__2016-20": 0.4375,
126
+ "main_score": 0.475626423690205
127
+ },
128
+ "harness|assin2_rte|assin2_rte|None|15": {
129
+ "f1_macro,all": 0.8902591771846999,
130
+ "acc,all": 0.8905228758169934,
131
+ "main_score": 0.8902591771846999
132
+ },
133
+ "harness|assin2_sts|assin2_sts|None|15": {
134
+ "pearson,all": 0.7417032662621263,
135
+ "mse,all": 0.4840727124183006,
136
+ "main_score": 0.7417032662621263
137
+ },
138
+ "harness|faquad_nli|faquad_nli|None|15": {
139
+ "f1_macro,all": 0.7691782381060779,
140
+ "acc,all": 0.82,
141
+ "main_score": 0.7691782381060779
142
+ },
143
+ "harness|hatebr_offensive|hatebr_offensive|None|25": {
144
+ "f1_macro,all": 0.8633798389731095,
145
+ "acc,all": 0.8642857142857143,
146
+ "main_score": 0.8633798389731095
147
+ },
148
+ "harness|portuguese_hate_speech|portuguese_hate_speech|None|25": {
149
+ "f1_macro,all": 0.6466753105050977,
150
+ "acc,all": 0.6556991774383079,
151
+ "main_score": 0.6466753105050977
152
+ },
153
+ "harness|tweetsentbr|tweetsentbr|None|25": {
154
+ "f1_macro,all": 0.6968077814532404,
155
+ "acc,all": 0.7179104477611941,
156
+ "main_score": 0.6968077814532404
157
+ }
158
+ },
159
+ "config_tasks": {
160
+ "harness|enem_challenge|enem_challenge": "LM Harness task",
161
+ "harness|bluex|bluex": "LM Harness task",
162
+ "harness|oab_exams|oab_exams": "LM Harness task",
163
+ "harness|assin2_rte|assin2_rte": "LM Harness task",
164
+ "harness|assin2_sts|assin2_sts": "LM Harness task",
165
+ "harness|faquad_nli|faquad_nli": "LM Harness task",
166
+ "harness|hatebr_offensive|hatebr_offensive": "LM Harness task",
167
+ "harness|portuguese_hate_speech|portuguese_hate_speech": "LM Harness task",
168
+ "harness|tweetsentbr|tweetsentbr": "LM Harness task"
169
+ },
170
+ "versions": {
171
+ "all": 0,
172
+ "harness|enem_challenge|enem_challenge": 1.1,
173
+ "harness|bluex|bluex": 1.1,
174
+ "harness|oab_exams|oab_exams": 1.5,
175
+ "harness|assin2_rte|assin2_rte": 1.1,
176
+ "harness|assin2_sts|assin2_sts": 1.1,
177
+ "harness|faquad_nli|faquad_nli": 1.1,
178
+ "harness|hatebr_offensive|hatebr_offensive": 1.0,
179
+ "harness|portuguese_hate_speech|portuguese_hate_speech": 1.0,
180
+ "harness|tweetsentbr|tweetsentbr": 1.0
181
+ },
182
+ "summary_tasks": {
183
+ "harness|enem_challenge|enem_challenge|None|3": {
184
+ "sample_size": 1429,
185
+ "truncated": 50,
186
+ "non_truncated": 1379,
187
+ "padded": 940,
188
+ "non_padded": 489,
189
+ "fewshots_truncated": 52,
190
+ "mean_seq_length": 1919.8929321203639,
191
+ "min_seq_length": 1608,
192
+ "max_seq_length": 2922,
193
+ "max_ctx_length": 2304,
194
+ "max_gen_toks": 256,
195
+ "mean_original_fewshots_size": 3.0,
196
+ "mean_effective_fewshot_size": 2.9636109167249827
197
+ },
198
+ "harness|bluex|bluex|None|3": {
199
+ "sample_size": 719,
200
+ "truncated": 65,
201
+ "non_truncated": 654,
202
+ "padded": 478,
203
+ "non_padded": 241,
204
+ "fewshots_truncated": 82,
205
+ "mean_seq_length": 1987.8052851182197,
206
+ "min_seq_length": 1559,
207
+ "max_seq_length": 2891,
208
+ "max_ctx_length": 2304,
209
+ "max_gen_toks": 256,
210
+ "mean_original_fewshots_size": 3.0,
211
+ "mean_effective_fewshot_size": 2.885952712100139
212
+ },
213
+ "harness|oab_exams|oab_exams|None|3": {
214
+ "sample_size": 2195,
215
+ "truncated": 0,
216
+ "non_truncated": 2195,
217
+ "padded": 1987,
218
+ "non_padded": 208,
219
+ "fewshots_truncated": 0,
220
+ "mean_seq_length": 1631.1088838268793,
221
+ "min_seq_length": 1318,
222
+ "max_seq_length": 2189,
223
+ "max_ctx_length": 2528,
224
+ "max_gen_toks": 32,
225
+ "mean_original_fewshots_size": 3.0,
226
+ "mean_effective_fewshot_size": 3.0
227
+ },
228
+ "harness|assin2_rte|assin2_rte|None|15": {
229
+ "sample_size": 2448,
230
+ "truncated": 0,
231
+ "non_truncated": 2448,
232
+ "padded": 216,
233
+ "non_padded": 2232,
234
+ "fewshots_truncated": 0,
235
+ "mean_seq_length": 1619.5996732026144,
236
+ "min_seq_length": 1595,
237
+ "max_seq_length": 1693,
238
+ "max_ctx_length": 2304,
239
+ "max_gen_toks": 256,
240
+ "mean_original_fewshots_size": 15.0,
241
+ "mean_effective_fewshot_size": 15.0
242
+ },
243
+ "harness|assin2_sts|assin2_sts|None|15": {
244
+ "sample_size": 2448,
245
+ "truncated": 0,
246
+ "non_truncated": 2448,
247
+ "padded": 250,
248
+ "non_padded": 2198,
249
+ "fewshots_truncated": 0,
250
+ "mean_seq_length": 1873.5996732026144,
251
+ "min_seq_length": 1849,
252
+ "max_seq_length": 1947,
253
+ "max_ctx_length": 2304,
254
+ "max_gen_toks": 256,
255
+ "mean_original_fewshots_size": 15.0,
256
+ "mean_effective_fewshot_size": 15.0
257
+ },
258
+ "harness|faquad_nli|faquad_nli|None|15": {
259
+ "sample_size": 650,
260
+ "truncated": 0,
261
+ "non_truncated": 650,
262
+ "padded": 488,
263
+ "non_padded": 162,
264
+ "fewshots_truncated": 0,
265
+ "mean_seq_length": 1939.183076923077,
266
+ "min_seq_length": 1877,
267
+ "max_seq_length": 2071,
268
+ "max_ctx_length": 2304,
269
+ "max_gen_toks": 256,
270
+ "mean_original_fewshots_size": 15.0,
271
+ "mean_effective_fewshot_size": 15.0
272
+ },
273
+ "harness|hatebr_offensive|hatebr_offensive|None|25": {
274
+ "sample_size": 1400,
275
+ "truncated": 0,
276
+ "non_truncated": 1400,
277
+ "padded": 810,
278
+ "non_padded": 590,
279
+ "fewshots_truncated": 0,
280
+ "mean_seq_length": 1640.8,
281
+ "min_seq_length": 1612,
282
+ "max_seq_length": 1927,
283
+ "max_ctx_length": 2304,
284
+ "max_gen_toks": 256,
285
+ "mean_original_fewshots_size": 25.0,
286
+ "mean_effective_fewshot_size": 25.0
287
+ },
288
+ "harness|portuguese_hate_speech|portuguese_hate_speech|None|25": {
289
+ "sample_size": 851,
290
+ "truncated": 0,
291
+ "non_truncated": 851,
292
+ "padded": 633,
293
+ "non_padded": 218,
294
+ "fewshots_truncated": 0,
295
+ "mean_seq_length": 2182.0340775558166,
296
+ "min_seq_length": 2143,
297
+ "max_seq_length": 2218,
298
+ "max_ctx_length": 2304,
299
+ "max_gen_toks": 256,
300
+ "mean_original_fewshots_size": 25.0,
301
+ "mean_effective_fewshot_size": 25.0
302
+ },
303
+ "harness|tweetsentbr|tweetsentbr|None|25": {
304
+ "sample_size": 2010,
305
+ "truncated": 0,
306
+ "non_truncated": 2010,
307
+ "padded": 757,
308
+ "non_padded": 1253,
309
+ "fewshots_truncated": 0,
310
+ "mean_seq_length": 1881.9114427860698,
311
+ "min_seq_length": 1859,
312
+ "max_seq_length": 1997,
313
+ "max_ctx_length": 2304,
314
+ "max_gen_toks": 256,
315
+ "mean_original_fewshots_size": 25.0,
316
+ "mean_effective_fewshot_size": 25.0
317
+ }
318
+ },
319
+ "summary_general": {
320
+ "truncated": 115,
321
+ "non_truncated": 14035,
322
+ "padded": 6559,
323
+ "non_padded": 7591,
324
+ "fewshots_truncated": 134
325
+ }
326
+ }