picocreator commited on
Commit
893fbd3
·
1 Parent(s): cd5b155

result updates

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +18 -18
  2. lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  3. lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +10 -10
  4. lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  5. lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +14 -14
  6. lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  7. lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +10 -10
  8. lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  9. lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +3 -3
  10. lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  11. lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +125 -125
  12. lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  13. lm-eval-output/allenai/OLMo-7B/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +62 -0
  14. lm-eval-output/allenai/OLMo-7B/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  15. lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +5 -5
  16. lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  17. lm-eval-output/allenai/OLMo-7B/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +2590 -0
  18. lm-eval-output/allenai/OLMo-7B/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  19. lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +220 -220
  20. lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  21. lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +2 -2
  22. lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  23. lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +3 -3
  24. lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  25. lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +87 -87
  26. lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  27. lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +7 -7
  28. lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  29. lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +28 -28
  30. lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  31. lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +3 -3
  32. lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  33. lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +5 -5
  34. lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  35. lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +160 -160
  36. lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  37. lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +27 -27
  38. lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  39. lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +18 -18
  40. lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  41. lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +17 -17
  42. lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  43. lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +3 -3
  44. lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  45. lm-eval-output/allenai/OLMo-7B/logiqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +66 -0
  46. lm-eval-output/allenai/OLMo-7B/logiqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  47. lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +6 -6
  48. lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
  49. lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +6 -6
  50. lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +2 -2
lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,33 +1,33 @@
1
  {
2
  "results": {
3
  "ai2_arc": {
4
- "acc,none": 0.6141488162344984,
5
- "acc_stderr,none": 0.11602553865195812,
6
- "acc_norm,none": 0.5944193912063134,
7
- "acc_norm_stderr,none": 0.09008281087077372,
8
  "alias": "ai2_arc"
9
  },
10
  "arc_challenge": {
11
- "acc,none": 0.36860068259385664,
12
- "acc_stderr,none": 0.014097810678042187,
13
- "acc_norm,none": 0.4044368600682594,
14
- "acc_norm_stderr,none": 0.014342036483436174,
15
  "alias": " - arc_challenge"
16
  },
17
  "arc_easy": {
18
- "acc,none": 0.7352693602693603,
19
- "acc_stderr,none": 0.009053021086173977,
20
- "acc_norm,none": 0.6881313131313131,
21
- "acc_norm_stderr,none": 0.00950582334581765,
22
  "alias": " - arc_easy"
23
  }
24
  },
25
  "groups": {
26
  "ai2_arc": {
27
- "acc,none": 0.6141488162344984,
28
- "acc_stderr,none": 0.11602553865195812,
29
- "acc_norm,none": 0.5944193912063134,
30
- "acc_norm_stderr,none": 0.09008281087077372,
31
  "alias": "ai2_arc"
32
  }
33
  },
@@ -120,7 +120,7 @@
120
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
121
  "batch_size": "auto",
122
  "batch_sizes": [
123
- 32
124
  ],
125
  "device": null,
126
  "use_cache": null,
@@ -128,5 +128,5 @@
128
  "bootstrap_iters": 100000,
129
  "gen_kwargs": null
130
  },
131
- "git_hash": "2e3ceb0"
132
  }
 
1
  {
2
  "results": {
3
  "ai2_arc": {
4
+ "acc,none": 0.6149943630214205,
5
+ "acc_stderr,none": 0.11562268036188031,
6
+ "acc_norm,none": 0.5938556933483653,
7
+ "acc_norm_stderr,none": 0.09141109524640427,
8
  "alias": "ai2_arc"
9
  },
10
  "arc_challenge": {
11
+ "acc,none": 0.3703071672354949,
12
+ "acc_stderr,none": 0.01411129875167495,
13
+ "acc_norm,none": 0.40102389078498296,
14
+ "acc_norm_stderr,none": 0.014322255790719867,
15
  "alias": " - arc_challenge"
16
  },
17
  "arc_easy": {
18
+ "acc,none": 0.7356902356902357,
19
+ "acc_stderr,none": 0.009048410451863016,
20
+ "acc_norm,none": 0.688973063973064,
21
+ "acc_norm_stderr,none": 0.009498790639757611,
22
  "alias": " - arc_easy"
23
  }
24
  },
25
  "groups": {
26
  "ai2_arc": {
27
+ "acc,none": 0.6149943630214205,
28
+ "acc_stderr,none": 0.11562268036188031,
29
+ "acc_norm,none": 0.5938556933483653,
30
+ "acc_norm_stderr,none": 0.09141109524640427,
31
  "alias": "ai2_arc"
32
  }
33
  },
 
120
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
121
  "batch_size": "auto",
122
  "batch_sizes": [
123
+ 64
124
  ],
125
  "device": null,
126
  "use_cache": null,
 
128
  "bootstrap_iters": 100000,
129
  "gen_kwargs": null
130
  },
131
+ "git_hash": "4d19ea9"
132
  }
lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:feda1a9a8d31385965b7cd26c524ddba7445b789e2a118919a275c549d0cea1e
3
- size 16263
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c97297d1125d6bc354e3013661d26de77d0f9150d0fcfdeac9fffde89d17ad3e
3
+ size 17347
lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "results": {
3
  "anli": {
4
- "acc,none": 0.3465625,
5
- "acc_stderr,none": 0.016792339011968412,
6
  "alias": "anli"
7
  },
8
  "anli_r1": {
9
- "acc,none": 0.325,
10
- "acc_stderr,none": 0.014818724459095526,
11
  "alias": " - anli_r1"
12
  },
13
  "anli_r2": {
@@ -16,15 +16,15 @@
16
  "alias": " - anli_r2"
17
  },
18
  "anli_r3": {
19
- "acc,none": 0.3566666666666667,
20
- "acc_stderr,none": 0.013833742805050717,
21
  "alias": " - anli_r3"
22
  }
23
  },
24
  "groups": {
25
  "anli": {
26
- "acc,none": 0.3465625,
27
- "acc_stderr,none": 0.016792339011968412,
28
  "alias": "anli"
29
  }
30
  },
@@ -149,7 +149,7 @@
149
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
150
  "batch_size": "auto",
151
  "batch_sizes": [
152
- 32
153
  ],
154
  "device": null,
155
  "use_cache": null,
@@ -157,5 +157,5 @@
157
  "bootstrap_iters": 100000,
158
  "gen_kwargs": null
159
  },
160
- "git_hash": "2e3ceb0"
161
  }
 
1
  {
2
  "results": {
3
  "anli": {
4
+ "acc,none": 0.3478125,
5
+ "acc_stderr,none": 0.017229187207023548,
6
  "alias": "anli"
7
  },
8
  "anli_r1": {
9
+ "acc,none": 0.328,
10
+ "acc_stderr,none": 0.014853842487270334,
11
  "alias": " - anli_r1"
12
  },
13
  "anli_r2": {
 
16
  "alias": " - anli_r2"
17
  },
18
  "anli_r3": {
19
+ "acc,none": 0.3575,
20
+ "acc_stderr,none": 0.0138409212452578,
21
  "alias": " - anli_r3"
22
  }
23
  },
24
  "groups": {
25
  "anli": {
26
+ "acc,none": 0.3478125,
27
+ "acc_stderr,none": 0.017229187207023548,
28
  "alias": "anli"
29
  }
30
  },
 
149
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
150
  "batch_size": "auto",
151
  "batch_sizes": [
152
+ 64
153
  ],
154
  "device": null,
155
  "use_cache": null,
 
157
  "bootstrap_iters": 100000,
158
  "gen_kwargs": null
159
  },
160
+ "git_hash": "4d19ea9"
161
  }
lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df4971e214a122144a0bdc530721b0e62f5d9807f2626826442b773b85ceb849
3
- size 14676
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8c0b720223175926be53cf7cab9e909d4caa03012d509c9fdfa54c566c5f98c
3
+ size 218218
lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,43 +1,43 @@
1
  {
2
  "results": {
3
  "arithmetic": {
4
- "acc,none": 0.007,
5
- "acc_stderr,none": 0.006724488098523242,
6
  "alias": "arithmetic"
7
  },
8
  "arithmetic_1dc": {
9
- "acc,none": 0.0065,
10
- "acc_stderr,none": 0.0017973564602277768,
11
  "alias": " - arithmetic_1dc"
12
  },
13
  "arithmetic_2da": {
14
- "acc,none": 0.0155,
15
- "acc_stderr,none": 0.0027629136515503164,
16
  "alias": " - arithmetic_2da"
17
  },
18
  "arithmetic_2dm": {
19
  "acc,none": 0.029,
20
- "acc_stderr,none": 0.0037532044004605246,
21
  "alias": " - arithmetic_2dm"
22
  },
23
  "arithmetic_2ds": {
24
  "acc,none": 0.0155,
25
- "acc_stderr,none": 0.002762913651550328,
26
  "alias": " - arithmetic_2ds"
27
  },
28
  "arithmetic_3da": {
29
  "acc,none": 0.0015,
30
- "acc_stderr,none": 0.0008655920660521528,
31
  "alias": " - arithmetic_3da"
32
  },
33
  "arithmetic_3ds": {
34
  "acc,none": 0.0015,
35
- "acc_stderr,none": 0.0008655920660521539,
36
  "alias": " - arithmetic_3ds"
37
  },
38
  "arithmetic_4da": {
39
  "acc,none": 0.0005,
40
- "acc_stderr,none": 0.0005000000000000151,
41
  "alias": " - arithmetic_4da"
42
  },
43
  "arithmetic_4ds": {
@@ -58,8 +58,8 @@
58
  },
59
  "groups": {
60
  "arithmetic": {
61
- "acc,none": 0.007,
62
- "acc_stderr,none": 0.006724488098523242,
63
  "alias": "arithmetic"
64
  }
65
  },
@@ -374,5 +374,5 @@
374
  "bootstrap_iters": 100000,
375
  "gen_kwargs": null
376
  },
377
- "git_hash": "2e3ceb0"
378
  }
 
1
  {
2
  "results": {
3
  "arithmetic": {
4
+ "acc,none": 0.00685,
5
+ "acc_stderr,none": 0.008348131833090362,
6
  "alias": "arithmetic"
7
  },
8
  "arithmetic_1dc": {
9
+ "acc,none": 0.0055,
10
+ "acc_stderr,none": 0.0016541593398342208,
11
  "alias": " - arithmetic_1dc"
12
  },
13
  "arithmetic_2da": {
14
+ "acc,none": 0.015,
15
+ "acc_stderr,none": 0.0027186753387999584,
16
  "alias": " - arithmetic_2da"
17
  },
18
  "arithmetic_2dm": {
19
  "acc,none": 0.029,
20
+ "acc_stderr,none": 0.003753204400460514,
21
  "alias": " - arithmetic_2dm"
22
  },
23
  "arithmetic_2ds": {
24
  "acc,none": 0.0155,
25
+ "acc_stderr,none": 0.002762913651550316,
26
  "alias": " - arithmetic_2ds"
27
  },
28
  "arithmetic_3da": {
29
  "acc,none": 0.0015,
30
+ "acc_stderr,none": 0.0008655920660521572,
31
  "alias": " - arithmetic_3da"
32
  },
33
  "arithmetic_3ds": {
34
  "acc,none": 0.0015,
35
+ "acc_stderr,none": 0.0008655920660521436,
36
  "alias": " - arithmetic_3ds"
37
  },
38
  "arithmetic_4da": {
39
  "acc,none": 0.0005,
40
+ "acc_stderr,none": 0.0005000000000000152,
41
  "alias": " - arithmetic_4da"
42
  },
43
  "arithmetic_4ds": {
 
58
  },
59
  "groups": {
60
  "arithmetic": {
61
+ "acc,none": 0.00685,
62
+ "acc_stderr,none": 0.008348131833090362,
63
  "alias": "arithmetic"
64
  }
65
  },
 
374
  "bootstrap_iters": 100000,
375
  "gen_kwargs": null
376
  },
377
+ "git_hash": "4d19ea9"
378
  }
lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:336d5042f526e0b5a3e6045ddb2334d66ffc6ba7c4f85d38364256c924609891
3
- size 25619
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c42b15aac6b807123fa0afb129e8c7c6a2716fd36275569cae8addaffa8156a
3
+ size 24315
lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -17,37 +17,37 @@
17
  },
18
  "arithmetic_4da": {
19
  "acc,none": 0.0005,
20
- "acc_stderr,none": 0.0005000000000000151,
21
  "alias": "arithmetic_4da"
22
  },
23
  "arithmetic_3ds": {
24
  "acc,none": 0.0015,
25
- "acc_stderr,none": 0.0008655920660521539,
26
  "alias": "arithmetic_3ds"
27
  },
28
  "arithmetic_3da": {
29
  "acc,none": 0.0015,
30
- "acc_stderr,none": 0.0008655920660521528,
31
  "alias": "arithmetic_3da"
32
  },
33
  "arithmetic_2ds": {
34
  "acc,none": 0.0155,
35
- "acc_stderr,none": 0.002762913651550328,
36
  "alias": "arithmetic_2ds"
37
  },
38
  "arithmetic_2dm": {
39
  "acc,none": 0.029,
40
- "acc_stderr,none": 0.0037532044004605246,
41
  "alias": "arithmetic_2dm"
42
  },
43
  "arithmetic_2da": {
44
- "acc,none": 0.0155,
45
- "acc_stderr,none": 0.0027629136515503164,
46
  "alias": "arithmetic_2da"
47
  },
48
  "arithmetic_1dc": {
49
- "acc,none": 0.0065,
50
- "acc_stderr,none": 0.0017973564602277768,
51
  "alias": "arithmetic_1dc"
52
  }
53
  },
@@ -360,5 +360,5 @@
360
  "bootstrap_iters": 100000,
361
  "gen_kwargs": null
362
  },
363
- "git_hash": "2e3ceb0"
364
  }
 
17
  },
18
  "arithmetic_4da": {
19
  "acc,none": 0.0005,
20
+ "acc_stderr,none": 0.0005000000000000152,
21
  "alias": "arithmetic_4da"
22
  },
23
  "arithmetic_3ds": {
24
  "acc,none": 0.0015,
25
+ "acc_stderr,none": 0.0008655920660521436,
26
  "alias": "arithmetic_3ds"
27
  },
28
  "arithmetic_3da": {
29
  "acc,none": 0.0015,
30
+ "acc_stderr,none": 0.0008655920660521572,
31
  "alias": "arithmetic_3da"
32
  },
33
  "arithmetic_2ds": {
34
  "acc,none": 0.0155,
35
+ "acc_stderr,none": 0.002762913651550316,
36
  "alias": "arithmetic_2ds"
37
  },
38
  "arithmetic_2dm": {
39
  "acc,none": 0.029,
40
+ "acc_stderr,none": 0.003753204400460514,
41
  "alias": "arithmetic_2dm"
42
  },
43
  "arithmetic_2da": {
44
+ "acc,none": 0.015,
45
+ "acc_stderr,none": 0.0027186753387999584,
46
  "alias": "arithmetic_2da"
47
  },
48
  "arithmetic_1dc": {
49
+ "acc,none": 0.0055,
50
+ "acc_stderr,none": 0.0016541593398342208,
51
  "alias": "arithmetic_1dc"
52
  }
53
  },
 
360
  "bootstrap_iters": 100000,
361
  "gen_kwargs": null
362
  },
363
+ "git_hash": "4d19ea9"
364
  }
lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c280fce1e1826f134c773e5a830152245f71f8505f03d4c5cfcb4af0ff0f9df8
3
- size 21272
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32fa05c3bc87cf418bde346540377951c16ce44c4ad8c62b9906df9e3a6a9fa8
3
+ size 24196
lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "results": {
3
  "asdiv": {
4
- "acc,none": 0.015618221258134491,
5
- "acc_stderr,none": 0.002583189883690767,
6
  "alias": "asdiv"
7
  }
8
  },
@@ -51,5 +51,5 @@
51
  "bootstrap_iters": 100000,
52
  "gen_kwargs": null
53
  },
54
- "git_hash": "2e3ceb0"
55
  }
 
1
  {
2
  "results": {
3
  "asdiv": {
4
+ "acc,none": 0.016052060737527116,
5
+ "acc_stderr,none": 0.002618244621382576,
6
  "alias": "asdiv"
7
  }
8
  },
 
51
  "bootstrap_iters": 100000,
52
  "gen_kwargs": null
53
  },
54
+ "git_hash": "4d19ea9"
55
  }
lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc981690a7e19989d96bb07b4c91f750b1d9921eca7d50386cdc5233a531770c
3
- size 16390
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce27e7da42d70ee894ebf832619cd226c6537795aa655df00e3003a3e19e7a29
3
+ size 5085
lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,163 +1,163 @@
1
  {
2
  "results": {
3
  "blimp": {
4
- "acc,none": 0.8316865671641791,
5
- "acc_stderr,none": 0.1603584899107365,
6
  "alias": "blimp"
7
  },
8
  "blimp_adjunct_island": {
9
- "acc,none": 0.904,
10
- "acc_stderr,none": 0.009320454434783215,
11
  "alias": " - blimp_adjunct_island"
12
  },
13
  "blimp_anaphor_gender_agreement": {
14
  "acc,none": 0.995,
15
- "acc_stderr,none": 0.002231586874844882,
16
  "alias": " - blimp_anaphor_gender_agreement"
17
  },
18
  "blimp_anaphor_number_agreement": {
19
- "acc,none": 0.994,
20
- "acc_stderr,none": 0.0024433521993298428,
21
  "alias": " - blimp_anaphor_number_agreement"
22
  },
23
  "blimp_animate_subject_passive": {
24
- "acc,none": 0.807,
25
- "acc_stderr,none": 0.012486268734370145,
26
  "alias": " - blimp_animate_subject_passive"
27
  },
28
  "blimp_animate_subject_trans": {
29
- "acc,none": 0.913,
30
- "acc_stderr,none": 0.008916866630745918,
31
  "alias": " - blimp_animate_subject_trans"
32
  },
33
  "blimp_causative": {
34
- "acc,none": 0.728,
35
- "acc_stderr,none": 0.014078856992462623,
36
  "alias": " - blimp_causative"
37
  },
38
  "blimp_complex_NP_island": {
39
- "acc,none": 0.596,
40
- "acc_stderr,none": 0.015524980677122581,
41
  "alias": " - blimp_complex_NP_island"
42
  },
43
  "blimp_coordinate_structure_constraint_complex_left_branch": {
44
- "acc,none": 0.82,
45
- "acc_stderr,none": 0.012155153135511949,
46
  "alias": " - blimp_coordinate_structure_constraint_complex_left_branch"
47
  },
48
  "blimp_coordinate_structure_constraint_object_extraction": {
49
  "acc,none": 0.891,
50
- "acc_stderr,none": 0.009859828407037188,
51
  "alias": " - blimp_coordinate_structure_constraint_object_extraction"
52
  },
53
  "blimp_determiner_noun_agreement_1": {
54
  "acc,none": 0.986,
55
- "acc_stderr,none": 0.0037172325482565877,
56
  "alias": " - blimp_determiner_noun_agreement_1"
57
  },
58
  "blimp_determiner_noun_agreement_2": {
59
- "acc,none": 0.973,
60
- "acc_stderr,none": 0.005128089049275288,
61
  "alias": " - blimp_determiner_noun_agreement_2"
62
  },
63
  "blimp_determiner_noun_agreement_irregular_1": {
64
- "acc,none": 0.933,
65
- "acc_stderr,none": 0.00791034598317755,
66
  "alias": " - blimp_determiner_noun_agreement_irregular_1"
67
  },
68
  "blimp_determiner_noun_agreement_irregular_2": {
69
- "acc,none": 0.954,
70
- "acc_stderr,none": 0.006627814717380719,
71
  "alias": " - blimp_determiner_noun_agreement_irregular_2"
72
  },
73
  "blimp_determiner_noun_agreement_with_adj_2": {
74
- "acc,none": 0.95,
75
- "acc_stderr,none": 0.0068954729748979,
76
  "alias": " - blimp_determiner_noun_agreement_with_adj_2"
77
  },
78
  "blimp_determiner_noun_agreement_with_adj_irregular_1": {
79
- "acc,none": 0.879,
80
- "acc_stderr,none": 0.010318210380946088,
81
  "alias": " - blimp_determiner_noun_agreement_with_adj_irregular_1"
82
  },
83
  "blimp_determiner_noun_agreement_with_adj_irregular_2": {
84
- "acc,none": 0.931,
85
- "acc_stderr,none": 0.00801893405031516,
86
  "alias": " - blimp_determiner_noun_agreement_with_adj_irregular_2"
87
  },
88
  "blimp_determiner_noun_agreement_with_adjective_1": {
89
- "acc,none": 0.973,
90
- "acc_stderr,none": 0.00512808904927529,
91
  "alias": " - blimp_determiner_noun_agreement_with_adjective_1"
92
  },
93
  "blimp_distractor_agreement_relational_noun": {
94
- "acc,none": 0.923,
95
- "acc_stderr,none": 0.008434580140240648,
96
  "alias": " - blimp_distractor_agreement_relational_noun"
97
  },
98
  "blimp_distractor_agreement_relative_clause": {
99
- "acc,none": 0.717,
100
- "acc_stderr,none": 0.014251810906481728,
101
  "alias": " - blimp_distractor_agreement_relative_clause"
102
  },
103
  "blimp_drop_argument": {
104
- "acc,none": 0.747,
105
- "acc_stderr,none": 0.01375427861358708,
106
  "alias": " - blimp_drop_argument"
107
  },
108
  "blimp_ellipsis_n_bar_1": {
109
  "acc,none": 0.812,
110
- "acc_stderr,none": 0.012361586015103744,
111
  "alias": " - blimp_ellipsis_n_bar_1"
112
  },
113
  "blimp_ellipsis_n_bar_2": {
114
- "acc,none": 0.949,
115
- "acc_stderr,none": 0.006960420062571421,
116
  "alias": " - blimp_ellipsis_n_bar_2"
117
  },
118
  "blimp_existential_there_object_raising": {
119
- "acc,none": 0.864,
120
- "acc_stderr,none": 0.010845350230472988,
121
  "alias": " - blimp_existential_there_object_raising"
122
  },
123
  "blimp_existential_there_quantifiers_1": {
124
  "acc,none": 0.985,
125
- "acc_stderr,none": 0.0038457495745030067,
126
  "alias": " - blimp_existential_there_quantifiers_1"
127
  },
128
  "blimp_existential_there_quantifiers_2": {
129
- "acc,none": 0.377,
130
- "acc_stderr,none": 0.015333170125779847,
131
  "alias": " - blimp_existential_there_quantifiers_2"
132
  },
133
  "blimp_existential_there_subject_raising": {
134
  "acc,none": 0.911,
135
- "acc_stderr,none": 0.009008893392651523,
136
  "alias": " - blimp_existential_there_subject_raising"
137
  },
138
  "blimp_expletive_it_object_raising": {
139
- "acc,none": 0.826,
140
- "acc_stderr,none": 0.01199449323097343,
141
  "alias": " - blimp_expletive_it_object_raising"
142
  },
143
  "blimp_inchoative": {
144
- "acc,none": 0.68,
145
- "acc_stderr,none": 0.014758652303574874,
146
  "alias": " - blimp_inchoative"
147
  },
148
  "blimp_intransitive": {
149
- "acc,none": 0.791,
150
- "acc_stderr,none": 0.012864077288499321,
151
  "alias": " - blimp_intransitive"
152
  },
153
  "blimp_irregular_past_participle_adjectives": {
154
  "acc,none": 0.979,
155
- "acc_stderr,none": 0.00453647215130652,
156
  "alias": " - blimp_irregular_past_participle_adjectives"
157
  },
158
  "blimp_irregular_past_participle_verbs": {
159
- "acc,none": 0.906,
160
- "acc_stderr,none": 0.009233052000787736,
161
  "alias": " - blimp_irregular_past_participle_verbs"
162
  },
163
  "blimp_irregular_plural_subject_verb_agreement_1": {
@@ -166,58 +166,58 @@
166
  "alias": " - blimp_irregular_plural_subject_verb_agreement_1"
167
  },
168
  "blimp_irregular_plural_subject_verb_agreement_2": {
169
- "acc,none": 0.932,
170
- "acc_stderr,none": 0.007964887911291603,
171
  "alias": " - blimp_irregular_plural_subject_verb_agreement_2"
172
  },
173
  "blimp_left_branch_island_echo_question": {
174
- "acc,none": 0.648,
175
- "acc_stderr,none": 0.015110404505648661,
176
  "alias": " - blimp_left_branch_island_echo_question"
177
  },
178
  "blimp_left_branch_island_simple_question": {
179
- "acc,none": 0.911,
180
- "acc_stderr,none": 0.009008893392651526,
181
  "alias": " - blimp_left_branch_island_simple_question"
182
  },
183
  "blimp_matrix_question_npi_licensor_present": {
184
- "acc,none": 0.607,
185
- "acc_stderr,none": 0.015452824654081496,
186
  "alias": " - blimp_matrix_question_npi_licensor_present"
187
  },
188
  "blimp_npi_present_1": {
189
- "acc,none": 0.674,
190
- "acc_stderr,none": 0.014830507204541038,
191
  "alias": " - blimp_npi_present_1"
192
  },
193
  "blimp_npi_present_2": {
194
- "acc,none": 0.73,
195
- "acc_stderr,none": 0.014046255632633915,
196
  "alias": " - blimp_npi_present_2"
197
  },
198
  "blimp_only_npi_licensor_present": {
199
- "acc,none": 0.974,
200
- "acc_stderr,none": 0.005034813735318216,
201
  "alias": " - blimp_only_npi_licensor_present"
202
  },
203
  "blimp_only_npi_scope": {
204
- "acc,none": 0.706,
205
- "acc_stderr,none": 0.01441429054000822,
206
  "alias": " - blimp_only_npi_scope"
207
  },
208
  "blimp_passive_1": {
209
- "acc,none": 0.895,
210
- "acc_stderr,none": 0.009698921026024971,
211
  "alias": " - blimp_passive_1"
212
  },
213
  "blimp_passive_2": {
214
- "acc,none": 0.906,
215
- "acc_stderr,none": 0.009233052000787728,
216
  "alias": " - blimp_passive_2"
217
  },
218
  "blimp_principle_A_c_command": {
219
- "acc,none": 0.741,
220
- "acc_stderr,none": 0.01386041525752791,
221
  "alias": " - blimp_principle_A_c_command"
222
  },
223
  "blimp_principle_A_case_1": {
@@ -226,125 +226,125 @@
226
  "alias": " - blimp_principle_A_case_1"
227
  },
228
  "blimp_principle_A_case_2": {
229
- "acc,none": 0.963,
230
- "acc_stderr,none": 0.005972157622389646,
231
  "alias": " - blimp_principle_A_case_2"
232
  },
233
  "blimp_principle_A_domain_1": {
234
- "acc,none": 0.999,
235
- "acc_stderr,none": 0.0010000000000000124,
236
  "alias": " - blimp_principle_A_domain_1"
237
  },
238
  "blimp_principle_A_domain_2": {
239
- "acc,none": 0.835,
240
- "acc_stderr,none": 0.011743632866916164,
241
  "alias": " - blimp_principle_A_domain_2"
242
  },
243
  "blimp_principle_A_domain_3": {
244
- "acc,none": 0.739,
245
- "acc_stderr,none": 0.013895037677965136,
246
  "alias": " - blimp_principle_A_domain_3"
247
  },
248
  "blimp_principle_A_reconstruction": {
249
- "acc,none": 0.376,
250
- "acc_stderr,none": 0.01532510550889813,
251
  "alias": " - blimp_principle_A_reconstruction"
252
  },
253
  "blimp_regular_plural_subject_verb_agreement_1": {
254
- "acc,none": 0.965,
255
- "acc_stderr,none": 0.005814534272734963,
256
  "alias": " - blimp_regular_plural_subject_verb_agreement_1"
257
  },
258
  "blimp_regular_plural_subject_verb_agreement_2": {
259
- "acc,none": 0.931,
260
- "acc_stderr,none": 0.008018934050315146,
261
  "alias": " - blimp_regular_plural_subject_verb_agreement_2"
262
  },
263
  "blimp_sentential_negation_npi_licensor_present": {
264
  "acc,none": 0.996,
265
- "acc_stderr,none": 0.00199699473909873,
266
  "alias": " - blimp_sentential_negation_npi_licensor_present"
267
  },
268
  "blimp_sentential_negation_npi_scope": {
269
- "acc,none": 0.759,
270
- "acc_stderr,none": 0.013531522534515419,
271
  "alias": " - blimp_sentential_negation_npi_scope"
272
  },
273
  "blimp_sentential_subject_island": {
274
- "acc,none": 0.559,
275
- "acc_stderr,none": 0.01570877989424268,
276
  "alias": " - blimp_sentential_subject_island"
277
  },
278
  "blimp_superlative_quantifiers_1": {
279
- "acc,none": 0.892,
280
- "acc_stderr,none": 0.009820001651345714,
281
  "alias": " - blimp_superlative_quantifiers_1"
282
  },
283
  "blimp_superlative_quantifiers_2": {
284
- "acc,none": 0.877,
285
- "acc_stderr,none": 0.010391293421849879,
286
  "alias": " - blimp_superlative_quantifiers_2"
287
  },
288
  "blimp_tough_vs_raising_1": {
289
- "acc,none": 0.663,
290
- "acc_stderr,none": 0.014955087918653603,
291
  "alias": " - blimp_tough_vs_raising_1"
292
  },
293
  "blimp_tough_vs_raising_2": {
294
- "acc,none": 0.853,
295
- "acc_stderr,none": 0.011203415395160335,
296
  "alias": " - blimp_tough_vs_raising_2"
297
  },
298
  "blimp_transitive": {
299
- "acc,none": 0.867,
300
- "acc_stderr,none": 0.010743669132397346,
301
  "alias": " - blimp_transitive"
302
  },
303
  "blimp_wh_island": {
304
  "acc,none": 0.877,
305
- "acc_stderr,none": 0.010391293421849877,
306
  "alias": " - blimp_wh_island"
307
  },
308
  "blimp_wh_questions_object_gap": {
309
- "acc,none": 0.841,
310
- "acc_stderr,none": 0.011569479368271296,
311
  "alias": " - blimp_wh_questions_object_gap"
312
  },
313
  "blimp_wh_questions_subject_gap": {
314
  "acc,none": 0.933,
315
- "acc_stderr,none": 0.007910345983177547,
316
  "alias": " - blimp_wh_questions_subject_gap"
317
  },
318
  "blimp_wh_questions_subject_gap_long_distance": {
319
- "acc,none": 0.922,
320
- "acc_stderr,none": 0.008484573530118588,
321
  "alias": " - blimp_wh_questions_subject_gap_long_distance"
322
  },
323
  "blimp_wh_vs_that_no_gap": {
324
- "acc,none": 0.974,
325
- "acc_stderr,none": 0.0050348137353182325,
326
  "alias": " - blimp_wh_vs_that_no_gap"
327
  },
328
  "blimp_wh_vs_that_no_gap_long_distance": {
329
- "acc,none": 0.962,
330
- "acc_stderr,none": 0.006049181150584946,
331
  "alias": " - blimp_wh_vs_that_no_gap_long_distance"
332
  },
333
  "blimp_wh_vs_that_with_gap": {
334
- "acc,none": 0.341,
335
- "acc_stderr,none": 0.0149981313484027,
336
  "alias": " - blimp_wh_vs_that_with_gap"
337
  },
338
  "blimp_wh_vs_that_with_gap_long_distance": {
339
- "acc,none": 0.253,
340
- "acc_stderr,none": 0.01375427861358708,
341
  "alias": " - blimp_wh_vs_that_with_gap_long_distance"
342
  }
343
  },
344
  "groups": {
345
  "blimp": {
346
- "acc,none": 0.8316865671641791,
347
- "acc_stderr,none": 0.1603584899107365,
348
  "alias": "blimp"
349
  }
350
  },
@@ -2245,5 +2245,5 @@
2245
  "bootstrap_iters": 100000,
2246
  "gen_kwargs": null
2247
  },
2248
- "git_hash": "2e3ceb0"
2249
  }
 
1
  {
2
  "results": {
3
  "blimp": {
4
+ "acc,none": 0.8317313432835821,
5
+ "acc_stderr,none": 0.1508121857988114,
6
  "alias": "blimp"
7
  },
8
  "blimp_adjunct_island": {
9
+ "acc,none": 0.903,
10
+ "acc_stderr,none": 0.009363689373248121,
11
  "alias": " - blimp_adjunct_island"
12
  },
13
  "blimp_anaphor_gender_agreement": {
14
  "acc,none": 0.995,
15
+ "acc_stderr,none": 0.0022315868748448786,
16
  "alias": " - blimp_anaphor_gender_agreement"
17
  },
18
  "blimp_anaphor_number_agreement": {
19
+ "acc,none": 0.992,
20
+ "acc_stderr,none": 0.002818500300504507,
21
  "alias": " - blimp_anaphor_number_agreement"
22
  },
23
  "blimp_animate_subject_passive": {
24
+ "acc,none": 0.811,
25
+ "acc_stderr,none": 0.012386784588117709,
26
  "alias": " - blimp_animate_subject_passive"
27
  },
28
  "blimp_animate_subject_trans": {
29
+ "acc,none": 0.911,
30
+ "acc_stderr,none": 0.00900889339265154,
31
  "alias": " - blimp_animate_subject_trans"
32
  },
33
  "blimp_causative": {
34
+ "acc,none": 0.736,
35
+ "acc_stderr,none": 0.013946271849440469,
36
  "alias": " - blimp_causative"
37
  },
38
  "blimp_complex_NP_island": {
39
+ "acc,none": 0.59,
40
+ "acc_stderr,none": 0.015560917136921672,
41
  "alias": " - blimp_complex_NP_island"
42
  },
43
  "blimp_coordinate_structure_constraint_complex_left_branch": {
44
+ "acc,none": 0.823,
45
+ "acc_stderr,none": 0.012075463420375061,
46
  "alias": " - blimp_coordinate_structure_constraint_complex_left_branch"
47
  },
48
  "blimp_coordinate_structure_constraint_object_extraction": {
49
  "acc,none": 0.891,
50
+ "acc_stderr,none": 0.009859828407037183,
51
  "alias": " - blimp_coordinate_structure_constraint_object_extraction"
52
  },
53
  "blimp_determiner_noun_agreement_1": {
54
  "acc,none": 0.986,
55
+ "acc_stderr,none": 0.003717232548256596,
56
  "alias": " - blimp_determiner_noun_agreement_1"
57
  },
58
  "blimp_determiner_noun_agreement_2": {
59
+ "acc,none": 0.975,
60
+ "acc_stderr,none": 0.004939574819698454,
61
  "alias": " - blimp_determiner_noun_agreement_2"
62
  },
63
  "blimp_determiner_noun_agreement_irregular_1": {
64
+ "acc,none": 0.937,
65
+ "acc_stderr,none": 0.007687007876286421,
66
  "alias": " - blimp_determiner_noun_agreement_irregular_1"
67
  },
68
  "blimp_determiner_noun_agreement_irregular_2": {
69
+ "acc,none": 0.951,
70
+ "acc_stderr,none": 0.006829761756140908,
71
  "alias": " - blimp_determiner_noun_agreement_irregular_2"
72
  },
73
  "blimp_determiner_noun_agreement_with_adj_2": {
74
+ "acc,none": 0.947,
75
+ "acc_stderr,none": 0.007088105617246446,
76
  "alias": " - blimp_determiner_noun_agreement_with_adj_2"
77
  },
78
  "blimp_determiner_noun_agreement_with_adj_irregular_1": {
79
+ "acc,none": 0.882,
80
+ "acc_stderr,none": 0.010206869264381796,
81
  "alias": " - blimp_determiner_noun_agreement_with_adj_irregular_1"
82
  },
83
  "blimp_determiner_noun_agreement_with_adj_irregular_2": {
84
+ "acc,none": 0.932,
85
+ "acc_stderr,none": 0.007964887911291603,
86
  "alias": " - blimp_determiner_noun_agreement_with_adj_irregular_2"
87
  },
88
  "blimp_determiner_noun_agreement_with_adjective_1": {
89
+ "acc,none": 0.971,
90
+ "acc_stderr,none": 0.005309160685756994,
91
  "alias": " - blimp_determiner_noun_agreement_with_adjective_1"
92
  },
93
  "blimp_distractor_agreement_relational_noun": {
94
+ "acc,none": 0.925,
95
+ "acc_stderr,none": 0.008333333333333345,
96
  "alias": " - blimp_distractor_agreement_relational_noun"
97
  },
98
  "blimp_distractor_agreement_relative_clause": {
99
+ "acc,none": 0.708,
100
+ "acc_stderr,none": 0.014385511563477341,
101
  "alias": " - blimp_distractor_agreement_relative_clause"
102
  },
103
  "blimp_drop_argument": {
104
+ "acc,none": 0.749,
105
+ "acc_stderr,none": 0.01371813351688892,
106
  "alias": " - blimp_drop_argument"
107
  },
108
  "blimp_ellipsis_n_bar_1": {
109
  "acc,none": 0.812,
110
+ "acc_stderr,none": 0.012361586015103761,
111
  "alias": " - blimp_ellipsis_n_bar_1"
112
  },
113
  "blimp_ellipsis_n_bar_2": {
114
+ "acc,none": 0.948,
115
+ "acc_stderr,none": 0.007024624213817138,
116
  "alias": " - blimp_ellipsis_n_bar_2"
117
  },
118
  "blimp_existential_there_object_raising": {
119
+ "acc,none": 0.863,
120
+ "acc_stderr,none": 0.010878848714333322,
121
  "alias": " - blimp_existential_there_object_raising"
122
  },
123
  "blimp_existential_there_quantifiers_1": {
124
  "acc,none": 0.985,
125
+ "acc_stderr,none": 0.003845749574503001,
126
  "alias": " - blimp_existential_there_quantifiers_1"
127
  },
128
  "blimp_existential_there_quantifiers_2": {
129
+ "acc,none": 0.361,
130
+ "acc_stderr,none": 0.015195720118175113,
131
  "alias": " - blimp_existential_there_quantifiers_2"
132
  },
133
  "blimp_existential_there_subject_raising": {
134
  "acc,none": 0.911,
135
+ "acc_stderr,none": 0.009008893392651526,
136
  "alias": " - blimp_existential_there_subject_raising"
137
  },
138
  "blimp_expletive_it_object_raising": {
139
+ "acc,none": 0.823,
140
+ "acc_stderr,none": 0.012075463420375061,
141
  "alias": " - blimp_expletive_it_object_raising"
142
  },
143
  "blimp_inchoative": {
144
+ "acc,none": 0.674,
145
+ "acc_stderr,none": 0.014830507204541035,
146
  "alias": " - blimp_inchoative"
147
  },
148
  "blimp_intransitive": {
149
+ "acc,none": 0.8,
150
+ "acc_stderr,none": 0.012655439943366662,
151
  "alias": " - blimp_intransitive"
152
  },
153
  "blimp_irregular_past_participle_adjectives": {
154
  "acc,none": 0.979,
155
+ "acc_stderr,none": 0.004536472151306486,
156
  "alias": " - blimp_irregular_past_participle_adjectives"
157
  },
158
  "blimp_irregular_past_participle_verbs": {
159
+ "acc,none": 0.914,
160
+ "acc_stderr,none": 0.008870325962594766,
161
  "alias": " - blimp_irregular_past_participle_verbs"
162
  },
163
  "blimp_irregular_plural_subject_verb_agreement_1": {
 
166
  "alias": " - blimp_irregular_plural_subject_verb_agreement_1"
167
  },
168
  "blimp_irregular_plural_subject_verb_agreement_2": {
169
+ "acc,none": 0.936,
170
+ "acc_stderr,none": 0.007743640226919306,
171
  "alias": " - blimp_irregular_plural_subject_verb_agreement_2"
172
  },
173
  "blimp_left_branch_island_echo_question": {
174
+ "acc,none": 0.636,
175
+ "acc_stderr,none": 0.015222868840522024,
176
  "alias": " - blimp_left_branch_island_echo_question"
177
  },
178
  "blimp_left_branch_island_simple_question": {
179
+ "acc,none": 0.91,
180
+ "acc_stderr,none": 0.009054390204866435,
181
  "alias": " - blimp_left_branch_island_simple_question"
182
  },
183
  "blimp_matrix_question_npi_licensor_present": {
184
+ "acc,none": 0.61,
185
+ "acc_stderr,none": 0.015431725053866608,
186
  "alias": " - blimp_matrix_question_npi_licensor_present"
187
  },
188
  "blimp_npi_present_1": {
189
+ "acc,none": 0.671,
190
+ "acc_stderr,none": 0.014865395385928362,
191
  "alias": " - blimp_npi_present_1"
192
  },
193
  "blimp_npi_present_2": {
194
+ "acc,none": 0.735,
195
+ "acc_stderr,none": 0.013963164754809953,
196
  "alias": " - blimp_npi_present_2"
197
  },
198
  "blimp_only_npi_licensor_present": {
199
+ "acc,none": 0.97,
200
+ "acc_stderr,none": 0.005397140829099204,
201
  "alias": " - blimp_only_npi_licensor_present"
202
  },
203
  "blimp_only_npi_scope": {
204
+ "acc,none": 0.711,
205
+ "acc_stderr,none": 0.014341711358296188,
206
  "alias": " - blimp_only_npi_scope"
207
  },
208
  "blimp_passive_1": {
209
+ "acc,none": 0.894,
210
+ "acc_stderr,none": 0.009739551265785127,
211
  "alias": " - blimp_passive_1"
212
  },
213
  "blimp_passive_2": {
214
+ "acc,none": 0.908,
215
+ "acc_stderr,none": 0.009144376393151125,
216
  "alias": " - blimp_passive_2"
217
  },
218
  "blimp_principle_A_c_command": {
219
+ "acc,none": 0.74,
220
+ "acc_stderr,none": 0.013877773329774164,
221
  "alias": " - blimp_principle_A_c_command"
222
  },
223
  "blimp_principle_A_case_1": {
 
226
  "alias": " - blimp_principle_A_case_1"
227
  },
228
  "blimp_principle_A_case_2": {
229
+ "acc,none": 0.967,
230
+ "acc_stderr,none": 0.005651808820452373,
231
  "alias": " - blimp_principle_A_case_2"
232
  },
233
  "blimp_principle_A_domain_1": {
234
+ "acc,none": 0.998,
235
+ "acc_stderr,none": 0.0014135055705578243,
236
  "alias": " - blimp_principle_A_domain_1"
237
  },
238
  "blimp_principle_A_domain_2": {
239
+ "acc,none": 0.834,
240
+ "acc_stderr,none": 0.011772110370812185,
241
  "alias": " - blimp_principle_A_domain_2"
242
  },
243
  "blimp_principle_A_domain_3": {
244
+ "acc,none": 0.74,
245
+ "acc_stderr,none": 0.013877773329774164,
246
  "alias": " - blimp_principle_A_domain_3"
247
  },
248
  "blimp_principle_A_reconstruction": {
249
+ "acc,none": 0.375,
250
+ "acc_stderr,none": 0.015316971293620996,
251
  "alias": " - blimp_principle_A_reconstruction"
252
  },
253
  "blimp_regular_plural_subject_verb_agreement_1": {
254
+ "acc,none": 0.967,
255
+ "acc_stderr,none": 0.005651808820452373,
256
  "alias": " - blimp_regular_plural_subject_verb_agreement_1"
257
  },
258
  "blimp_regular_plural_subject_verb_agreement_2": {
259
+ "acc,none": 0.936,
260
+ "acc_stderr,none": 0.0077436402269193145,
261
  "alias": " - blimp_regular_plural_subject_verb_agreement_2"
262
  },
263
  "blimp_sentential_negation_npi_licensor_present": {
264
  "acc,none": 0.996,
265
+ "acc_stderr,none": 0.001996994739098729,
266
  "alias": " - blimp_sentential_negation_npi_licensor_present"
267
  },
268
  "blimp_sentential_negation_npi_scope": {
269
+ "acc,none": 0.755,
270
+ "acc_stderr,none": 0.013607356839598118,
271
  "alias": " - blimp_sentential_negation_npi_scope"
272
  },
273
  "blimp_sentential_subject_island": {
274
+ "acc,none": 0.57,
275
+ "acc_stderr,none": 0.015663503610155283,
276
  "alias": " - blimp_sentential_subject_island"
277
  },
278
  "blimp_superlative_quantifiers_1": {
279
+ "acc,none": 0.894,
280
+ "acc_stderr,none": 0.009739551265785141,
281
  "alias": " - blimp_superlative_quantifiers_1"
282
  },
283
  "blimp_superlative_quantifiers_2": {
284
+ "acc,none": 0.875,
285
+ "acc_stderr,none": 0.010463483381956722,
286
  "alias": " - blimp_superlative_quantifiers_2"
287
  },
288
  "blimp_tough_vs_raising_1": {
289
+ "acc,none": 0.667,
290
+ "acc_stderr,none": 0.01491084616422986,
291
  "alias": " - blimp_tough_vs_raising_1"
292
  },
293
  "blimp_tough_vs_raising_2": {
294
+ "acc,none": 0.852,
295
+ "acc_stderr,none": 0.011234866364235244,
296
  "alias": " - blimp_tough_vs_raising_2"
297
  },
298
  "blimp_transitive": {
299
+ "acc,none": 0.856,
300
+ "acc_stderr,none": 0.011107987548939149,
301
  "alias": " - blimp_transitive"
302
  },
303
  "blimp_wh_island": {
304
  "acc,none": 0.877,
305
+ "acc_stderr,none": 0.010391293421849879,
306
  "alias": " - blimp_wh_island"
307
  },
308
  "blimp_wh_questions_object_gap": {
309
+ "acc,none": 0.84,
310
+ "acc_stderr,none": 0.011598902298689012,
311
  "alias": " - blimp_wh_questions_object_gap"
312
  },
313
  "blimp_wh_questions_subject_gap": {
314
  "acc,none": 0.933,
315
+ "acc_stderr,none": 0.007910345983177546,
316
  "alias": " - blimp_wh_questions_subject_gap"
317
  },
318
  "blimp_wh_questions_subject_gap_long_distance": {
319
+ "acc,none": 0.92,
320
+ "acc_stderr,none": 0.008583336977753655,
321
  "alias": " - blimp_wh_questions_subject_gap_long_distance"
322
  },
323
  "blimp_wh_vs_that_no_gap": {
324
+ "acc,none": 0.976,
325
+ "acc_stderr,none": 0.004842256441727058,
326
  "alias": " - blimp_wh_vs_that_no_gap"
327
  },
328
  "blimp_wh_vs_that_no_gap_long_distance": {
329
+ "acc,none": 0.963,
330
+ "acc_stderr,none": 0.005972157622389631,
331
  "alias": " - blimp_wh_vs_that_no_gap_long_distance"
332
  },
333
  "blimp_wh_vs_that_with_gap": {
334
+ "acc,none": 0.345,
335
+ "acc_stderr,none": 0.015039986742055237,
336
  "alias": " - blimp_wh_vs_that_with_gap"
337
  },
338
  "blimp_wh_vs_that_with_gap_long_distance": {
339
+ "acc,none": 0.256,
340
+ "acc_stderr,none": 0.013807775152234188,
341
  "alias": " - blimp_wh_vs_that_with_gap_long_distance"
342
  }
343
  },
344
  "groups": {
345
  "blimp": {
346
+ "acc,none": 0.8317313432835821,
347
+ "acc_stderr,none": 0.1508121857988114,
348
  "alias": "blimp"
349
  }
350
  },
 
2245
  "bootstrap_iters": 100000,
2246
  "gen_kwargs": null
2247
  },
2248
+ "git_hash": "4d19ea9"
2249
  }
lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a3b311b45000c90305578cb19f065bec59f430e131dab0963128cb73e9786b4
3
- size 294489
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed06ad1b40158e39bb8a97d96ca9ea78c4553bada4d6771c7f732c00d13ec072
3
+ size 150812
lm-eval-output/allenai/OLMo-7B/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "boolq": {
4
+ "acc,none": 0.7262996941896025,
5
+ "acc_stderr,none": 0.0077980876386284275,
6
+ "alias": "boolq"
7
+ }
8
+ },
9
+ "configs": {
10
+ "boolq": {
11
+ "task": "boolq",
12
+ "group": [
13
+ "super-glue-lm-eval-v1"
14
+ ],
15
+ "dataset_path": "super_glue",
16
+ "dataset_name": "boolq",
17
+ "training_split": "train",
18
+ "validation_split": "validation",
19
+ "doc_to_text": "{{passage}}\nQuestion: {{question}}?\nAnswer:",
20
+ "doc_to_target": "label",
21
+ "doc_to_choice": [
22
+ "no",
23
+ "yes"
24
+ ],
25
+ "description": "",
26
+ "target_delimiter": " ",
27
+ "fewshot_delimiter": "\n\n",
28
+ "metric_list": [
29
+ {
30
+ "metric": "acc"
31
+ }
32
+ ],
33
+ "output_type": "multiple_choice",
34
+ "repeats": 1,
35
+ "should_decontaminate": true,
36
+ "doc_to_decontamination_query": "passage",
37
+ "metadata": {
38
+ "version": 2.0
39
+ }
40
+ }
41
+ },
42
+ "versions": {
43
+ "boolq": 2.0
44
+ },
45
+ "n-shot": {
46
+ "boolq": 0
47
+ },
48
+ "config": {
49
+ "model": "hf",
50
+ "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
51
+ "batch_size": "auto",
52
+ "batch_sizes": [
53
+ 16
54
+ ],
55
+ "device": null,
56
+ "use_cache": null,
57
+ "limit": null,
58
+ "bootstrap_iters": 100000,
59
+ "gen_kwargs": null
60
+ },
61
+ "git_hash": "4d19ea9"
62
+ }
lm-eval-output/allenai/OLMo-7B/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:444fe6f0cbaf5443ee1dfa05e3d4f1806c4556054ceabcfea74b2c4eb6ee803a
3
- size 21711
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:febb881e0dbe0198174c3f778342100789ca36e759ae461989f8e7af2ac253e3
3
+ size 28126
lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "results": {
3
  "cb": {
4
- "acc,none": 0.16071428571428573,
5
- "acc_stderr,none": 0.049522300593062986,
6
- "f1,none": 0.14181286549707603,
7
  "f1_stderr,none": "N/A",
8
  "alias": "cb"
9
  }
@@ -56,7 +56,7 @@
56
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
57
  "batch_size": "auto",
58
  "batch_sizes": [
59
- 16
60
  ],
61
  "device": null,
62
  "use_cache": null,
@@ -64,5 +64,5 @@
64
  "bootstrap_iters": 100000,
65
  "gen_kwargs": null
66
  },
67
- "git_hash": "2e3ceb0"
68
  }
 
1
  {
2
  "results": {
3
  "cb": {
4
+ "acc,none": 0.125,
5
+ "acc_stderr,none": 0.04459412925079224,
6
+ "f1,none": 0.11129975476325221,
7
  "f1_stderr,none": "N/A",
8
  "alias": "cb"
9
  }
 
56
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
57
  "batch_size": "auto",
58
  "batch_sizes": [
59
+ 64
60
  ],
61
  "device": null,
62
  "use_cache": null,
 
64
  "bootstrap_iters": 100000,
65
  "gen_kwargs": null
66
  },
67
+ "git_hash": "4d19ea9"
68
  }
lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:239d0b8aec1eddb6f22bc82f54f5fa42e3a45de06aff4fc8e0aef720286f58fe
3
- size 14061
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18b9cbd95de5c6daf9044a1521bf7698566fb7e3bdd55b4737f9e8dd8f4aca57
3
+ size 2948
lm-eval-output/allenai/OLMo-7B/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,2590 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "ceval-valid": {
4
+ "acc,none": 0.2526002971768202,
5
+ "acc_stderr,none": 0.11248875724999531,
6
+ "acc_norm,none": 0.2526002971768202,
7
+ "acc_norm_stderr,none": 0.11248875724999531,
8
+ "alias": "ceval-valid"
9
+ },
10
+ "ceval-valid_accountant": {
11
+ "acc,none": 0.24489795918367346,
12
+ "acc_stderr,none": 0.062069005411206336,
13
+ "acc_norm,none": 0.24489795918367346,
14
+ "acc_norm_stderr,none": 0.062069005411206336,
15
+ "alias": " - ceval-valid_accountant"
16
+ },
17
+ "ceval-valid_advanced_mathematics": {
18
+ "acc,none": 0.10526315789473684,
19
+ "acc_stderr,none": 0.0723351864143449,
20
+ "acc_norm,none": 0.10526315789473684,
21
+ "acc_norm_stderr,none": 0.0723351864143449,
22
+ "alias": " - ceval-valid_advanced_mathematics"
23
+ },
24
+ "ceval-valid_art_studies": {
25
+ "acc,none": 0.2727272727272727,
26
+ "acc_stderr,none": 0.07872958216222173,
27
+ "acc_norm,none": 0.2727272727272727,
28
+ "acc_norm_stderr,none": 0.07872958216222173,
29
+ "alias": " - ceval-valid_art_studies"
30
+ },
31
+ "ceval-valid_basic_medicine": {
32
+ "acc,none": 0.15789473684210525,
33
+ "acc_stderr,none": 0.08594700851870798,
34
+ "acc_norm,none": 0.15789473684210525,
35
+ "acc_norm_stderr,none": 0.08594700851870798,
36
+ "alias": " - ceval-valid_basic_medicine"
37
+ },
38
+ "ceval-valid_business_administration": {
39
+ "acc,none": 0.30303030303030304,
40
+ "acc_stderr,none": 0.08124094920275463,
41
+ "acc_norm,none": 0.30303030303030304,
42
+ "acc_norm_stderr,none": 0.08124094920275463,
43
+ "alias": " - ceval-valid_business_administration"
44
+ },
45
+ "ceval-valid_chinese_language_and_literature": {
46
+ "acc,none": 0.2608695652173913,
47
+ "acc_stderr,none": 0.09361833424764437,
48
+ "acc_norm,none": 0.2608695652173913,
49
+ "acc_norm_stderr,none": 0.09361833424764437,
50
+ "alias": " - ceval-valid_chinese_language_and_literature"
51
+ },
52
+ "ceval-valid_civil_servant": {
53
+ "acc,none": 0.19148936170212766,
54
+ "acc_stderr,none": 0.05801446334976932,
55
+ "acc_norm,none": 0.19148936170212766,
56
+ "acc_norm_stderr,none": 0.05801446334976932,
57
+ "alias": " - ceval-valid_civil_servant"
58
+ },
59
+ "ceval-valid_clinical_medicine": {
60
+ "acc,none": 0.18181818181818182,
61
+ "acc_stderr,none": 0.08416546361568647,
62
+ "acc_norm,none": 0.18181818181818182,
63
+ "acc_norm_stderr,none": 0.08416546361568647,
64
+ "alias": " - ceval-valid_clinical_medicine"
65
+ },
66
+ "ceval-valid_college_chemistry": {
67
+ "acc,none": 0.25,
68
+ "acc_stderr,none": 0.09028938981432691,
69
+ "acc_norm,none": 0.25,
70
+ "acc_norm_stderr,none": 0.09028938981432691,
71
+ "alias": " - ceval-valid_college_chemistry"
72
+ },
73
+ "ceval-valid_college_economics": {
74
+ "acc,none": 0.34545454545454546,
75
+ "acc_stderr,none": 0.06470956516382613,
76
+ "acc_norm,none": 0.34545454545454546,
77
+ "acc_norm_stderr,none": 0.06470956516382613,
78
+ "alias": " - ceval-valid_college_economics"
79
+ },
80
+ "ceval-valid_college_physics": {
81
+ "acc,none": 0.15789473684210525,
82
+ "acc_stderr,none": 0.08594700851870798,
83
+ "acc_norm,none": 0.15789473684210525,
84
+ "acc_norm_stderr,none": 0.08594700851870798,
85
+ "alias": " - ceval-valid_college_physics"
86
+ },
87
+ "ceval-valid_college_programming": {
88
+ "acc,none": 0.21621621621621623,
89
+ "acc_stderr,none": 0.0686105685212965,
90
+ "acc_norm,none": 0.21621621621621623,
91
+ "acc_norm_stderr,none": 0.0686105685212965,
92
+ "alias": " - ceval-valid_college_programming"
93
+ },
94
+ "ceval-valid_computer_architecture": {
95
+ "acc,none": 0.3333333333333333,
96
+ "acc_stderr,none": 0.10540925533894599,
97
+ "acc_norm,none": 0.3333333333333333,
98
+ "acc_norm_stderr,none": 0.10540925533894599,
99
+ "alias": " - ceval-valid_computer_architecture"
100
+ },
101
+ "ceval-valid_computer_network": {
102
+ "acc,none": 0.3684210526315789,
103
+ "acc_stderr,none": 0.11369720523522557,
104
+ "acc_norm,none": 0.3684210526315789,
105
+ "acc_norm_stderr,none": 0.11369720523522557,
106
+ "alias": " - ceval-valid_computer_network"
107
+ },
108
+ "ceval-valid_discrete_mathematics": {
109
+ "acc,none": 0.1875,
110
+ "acc_stderr,none": 0.10077822185373188,
111
+ "acc_norm,none": 0.1875,
112
+ "acc_norm_stderr,none": 0.10077822185373188,
113
+ "alias": " - ceval-valid_discrete_mathematics"
114
+ },
115
+ "ceval-valid_education_science": {
116
+ "acc,none": 0.27586206896551724,
117
+ "acc_stderr,none": 0.08446516354424752,
118
+ "acc_norm,none": 0.27586206896551724,
119
+ "acc_norm_stderr,none": 0.08446516354424752,
120
+ "alias": " - ceval-valid_education_science"
121
+ },
122
+ "ceval-valid_electrical_engineer": {
123
+ "acc,none": 0.2702702702702703,
124
+ "acc_stderr,none": 0.07401656182502246,
125
+ "acc_norm,none": 0.2702702702702703,
126
+ "acc_norm_stderr,none": 0.07401656182502246,
127
+ "alias": " - ceval-valid_electrical_engineer"
128
+ },
129
+ "ceval-valid_environmental_impact_assessment_engineer": {
130
+ "acc,none": 0.25806451612903225,
131
+ "acc_stderr,none": 0.0798889274021794,
132
+ "acc_norm,none": 0.25806451612903225,
133
+ "acc_norm_stderr,none": 0.0798889274021794,
134
+ "alias": " - ceval-valid_environmental_impact_assessment_engineer"
135
+ },
136
+ "ceval-valid_fire_engineer": {
137
+ "acc,none": 0.3225806451612903,
138
+ "acc_stderr,none": 0.08534681648595453,
139
+ "acc_norm,none": 0.3225806451612903,
140
+ "acc_norm_stderr,none": 0.08534681648595453,
141
+ "alias": " - ceval-valid_fire_engineer"
142
+ },
143
+ "ceval-valid_high_school_biology": {
144
+ "acc,none": 0.21052631578947367,
145
+ "acc_stderr,none": 0.0960916767552923,
146
+ "acc_norm,none": 0.21052631578947367,
147
+ "acc_norm_stderr,none": 0.0960916767552923,
148
+ "alias": " - ceval-valid_high_school_biology"
149
+ },
150
+ "ceval-valid_high_school_chemistry": {
151
+ "acc,none": 0.3684210526315789,
152
+ "acc_stderr,none": 0.11369720523522561,
153
+ "acc_norm,none": 0.3684210526315789,
154
+ "acc_norm_stderr,none": 0.11369720523522561,
155
+ "alias": " - ceval-valid_high_school_chemistry"
156
+ },
157
+ "ceval-valid_high_school_chinese": {
158
+ "acc,none": 0.10526315789473684,
159
+ "acc_stderr,none": 0.0723351864143449,
160
+ "acc_norm,none": 0.10526315789473684,
161
+ "acc_norm_stderr,none": 0.0723351864143449,
162
+ "alias": " - ceval-valid_high_school_chinese"
163
+ },
164
+ "ceval-valid_high_school_geography": {
165
+ "acc,none": 0.3157894736842105,
166
+ "acc_stderr,none": 0.10956136839295434,
167
+ "acc_norm,none": 0.3157894736842105,
168
+ "acc_norm_stderr,none": 0.10956136839295434,
169
+ "alias": " - ceval-valid_high_school_geography"
170
+ },
171
+ "ceval-valid_high_school_history": {
172
+ "acc,none": 0.25,
173
+ "acc_stderr,none": 0.09933992677987828,
174
+ "acc_norm,none": 0.25,
175
+ "acc_norm_stderr,none": 0.09933992677987828,
176
+ "alias": " - ceval-valid_high_school_history"
177
+ },
178
+ "ceval-valid_high_school_mathematics": {
179
+ "acc,none": 0.2777777777777778,
180
+ "acc_stderr,none": 0.1086324845659782,
181
+ "acc_norm,none": 0.2777777777777778,
182
+ "acc_norm_stderr,none": 0.1086324845659782,
183
+ "alias": " - ceval-valid_high_school_mathematics"
184
+ },
185
+ "ceval-valid_high_school_physics": {
186
+ "acc,none": 0.3157894736842105,
187
+ "acc_stderr,none": 0.10956136839295433,
188
+ "acc_norm,none": 0.3157894736842105,
189
+ "acc_norm_stderr,none": 0.10956136839295433,
190
+ "alias": " - ceval-valid_high_school_physics"
191
+ },
192
+ "ceval-valid_high_school_politics": {
193
+ "acc,none": 0.10526315789473684,
194
+ "acc_stderr,none": 0.07233518641434492,
195
+ "acc_norm,none": 0.10526315789473684,
196
+ "acc_norm_stderr,none": 0.07233518641434492,
197
+ "alias": " - ceval-valid_high_school_politics"
198
+ },
199
+ "ceval-valid_ideological_and_moral_cultivation": {
200
+ "acc,none": 0.15789473684210525,
201
+ "acc_stderr,none": 0.08594700851870798,
202
+ "acc_norm,none": 0.15789473684210525,
203
+ "acc_norm_stderr,none": 0.08594700851870798,
204
+ "alias": " - ceval-valid_ideological_and_moral_cultivation"
205
+ },
206
+ "ceval-valid_law": {
207
+ "acc,none": 0.3333333333333333,
208
+ "acc_stderr,none": 0.09829463743659808,
209
+ "acc_norm,none": 0.3333333333333333,
210
+ "acc_norm_stderr,none": 0.09829463743659808,
211
+ "alias": " - ceval-valid_law"
212
+ },
213
+ "ceval-valid_legal_professional": {
214
+ "acc,none": 0.17391304347826086,
215
+ "acc_stderr,none": 0.0808104675899639,
216
+ "acc_norm,none": 0.17391304347826086,
217
+ "acc_norm_stderr,none": 0.0808104675899639,
218
+ "alias": " - ceval-valid_legal_professional"
219
+ },
220
+ "ceval-valid_logic": {
221
+ "acc,none": 0.22727272727272727,
222
+ "acc_stderr,none": 0.09144861547306321,
223
+ "acc_norm,none": 0.22727272727272727,
224
+ "acc_norm_stderr,none": 0.09144861547306321,
225
+ "alias": " - ceval-valid_logic"
226
+ },
227
+ "ceval-valid_mao_zedong_thought": {
228
+ "acc,none": 0.2916666666666667,
229
+ "acc_stderr,none": 0.09477598811252415,
230
+ "acc_norm,none": 0.2916666666666667,
231
+ "acc_norm_stderr,none": 0.09477598811252415,
232
+ "alias": " - ceval-valid_mao_zedong_thought"
233
+ },
234
+ "ceval-valid_marxism": {
235
+ "acc,none": 0.2631578947368421,
236
+ "acc_stderr,none": 0.10379087338771256,
237
+ "acc_norm,none": 0.2631578947368421,
238
+ "acc_norm_stderr,none": 0.10379087338771256,
239
+ "alias": " - ceval-valid_marxism"
240
+ },
241
+ "ceval-valid_metrology_engineer": {
242
+ "acc,none": 0.25,
243
+ "acc_stderr,none": 0.09028938981432691,
244
+ "acc_norm,none": 0.25,
245
+ "acc_norm_stderr,none": 0.09028938981432691,
246
+ "alias": " - ceval-valid_metrology_engineer"
247
+ },
248
+ "ceval-valid_middle_school_biology": {
249
+ "acc,none": 0.2857142857142857,
250
+ "acc_stderr,none": 0.10101525445522108,
251
+ "acc_norm,none": 0.2857142857142857,
252
+ "acc_norm_stderr,none": 0.10101525445522108,
253
+ "alias": " - ceval-valid_middle_school_biology"
254
+ },
255
+ "ceval-valid_middle_school_chemistry": {
256
+ "acc,none": 0.3,
257
+ "acc_stderr,none": 0.10513149660756935,
258
+ "acc_norm,none": 0.3,
259
+ "acc_norm_stderr,none": 0.10513149660756935,
260
+ "alias": " - ceval-valid_middle_school_chemistry"
261
+ },
262
+ "ceval-valid_middle_school_geography": {
263
+ "acc,none": 0.08333333333333333,
264
+ "acc_stderr,none": 0.08333333333333331,
265
+ "acc_norm,none": 0.08333333333333333,
266
+ "acc_norm_stderr,none": 0.08333333333333331,
267
+ "alias": " - ceval-valid_middle_school_geography"
268
+ },
269
+ "ceval-valid_middle_school_history": {
270
+ "acc,none": 0.4090909090909091,
271
+ "acc_stderr,none": 0.10729033533674223,
272
+ "acc_norm,none": 0.4090909090909091,
273
+ "acc_norm_stderr,none": 0.10729033533674223,
274
+ "alias": " - ceval-valid_middle_school_history"
275
+ },
276
+ "ceval-valid_middle_school_mathematics": {
277
+ "acc,none": 0.3684210526315789,
278
+ "acc_stderr,none": 0.11369720523522561,
279
+ "acc_norm,none": 0.3684210526315789,
280
+ "acc_norm_stderr,none": 0.11369720523522561,
281
+ "alias": " - ceval-valid_middle_school_mathematics"
282
+ },
283
+ "ceval-valid_middle_school_physics": {
284
+ "acc,none": 0.47368421052631576,
285
+ "acc_stderr,none": 0.1176877882894626,
286
+ "acc_norm,none": 0.47368421052631576,
287
+ "acc_norm_stderr,none": 0.1176877882894626,
288
+ "alias": " - ceval-valid_middle_school_physics"
289
+ },
290
+ "ceval-valid_middle_school_politics": {
291
+ "acc,none": 0.3333333333333333,
292
+ "acc_stderr,none": 0.10540925533894598,
293
+ "acc_norm,none": 0.3333333333333333,
294
+ "acc_norm_stderr,none": 0.10540925533894598,
295
+ "alias": " - ceval-valid_middle_school_politics"
296
+ },
297
+ "ceval-valid_modern_chinese_history": {
298
+ "acc,none": 0.2608695652173913,
299
+ "acc_stderr,none": 0.09361833424764437,
300
+ "acc_norm,none": 0.2608695652173913,
301
+ "acc_norm_stderr,none": 0.09361833424764437,
302
+ "alias": " - ceval-valid_modern_chinese_history"
303
+ },
304
+ "ceval-valid_operating_system": {
305
+ "acc,none": 0.05263157894736842,
306
+ "acc_stderr,none": 0.05263157894736841,
307
+ "acc_norm,none": 0.05263157894736842,
308
+ "acc_norm_stderr,none": 0.05263157894736841,
309
+ "alias": " - ceval-valid_operating_system"
310
+ },
311
+ "ceval-valid_physician": {
312
+ "acc,none": 0.20408163265306123,
313
+ "acc_stderr,none": 0.05817221556628251,
314
+ "acc_norm,none": 0.20408163265306123,
315
+ "acc_norm_stderr,none": 0.05817221556628251,
316
+ "alias": " - ceval-valid_physician"
317
+ },
318
+ "ceval-valid_plant_protection": {
319
+ "acc,none": 0.2727272727272727,
320
+ "acc_stderr,none": 0.0971859061499725,
321
+ "acc_norm,none": 0.2727272727272727,
322
+ "acc_norm_stderr,none": 0.0971859061499725,
323
+ "alias": " - ceval-valid_plant_protection"
324
+ },
325
+ "ceval-valid_probability_and_statistics": {
326
+ "acc,none": 0.2222222222222222,
327
+ "acc_stderr,none": 0.10083169033033673,
328
+ "acc_norm,none": 0.2222222222222222,
329
+ "acc_norm_stderr,none": 0.10083169033033673,
330
+ "alias": " - ceval-valid_probability_and_statistics"
331
+ },
332
+ "ceval-valid_professional_tour_guide": {
333
+ "acc,none": 0.41379310344827586,
334
+ "acc_stderr,none": 0.0930760769837004,
335
+ "acc_norm,none": 0.41379310344827586,
336
+ "acc_norm_stderr,none": 0.0930760769837004,
337
+ "alias": " - ceval-valid_professional_tour_guide"
338
+ },
339
+ "ceval-valid_sports_science": {
340
+ "acc,none": 0.15789473684210525,
341
+ "acc_stderr,none": 0.08594700851870798,
342
+ "acc_norm,none": 0.15789473684210525,
343
+ "acc_norm_stderr,none": 0.08594700851870798,
344
+ "alias": " - ceval-valid_sports_science"
345
+ },
346
+ "ceval-valid_tax_accountant": {
347
+ "acc,none": 0.22448979591836735,
348
+ "acc_stderr,none": 0.06022425581505364,
349
+ "acc_norm,none": 0.22448979591836735,
350
+ "acc_norm_stderr,none": 0.06022425581505364,
351
+ "alias": " - ceval-valid_tax_accountant"
352
+ },
353
+ "ceval-valid_teacher_qualification": {
354
+ "acc,none": 0.22727272727272727,
355
+ "acc_stderr,none": 0.06390760676613884,
356
+ "acc_norm,none": 0.22727272727272727,
357
+ "acc_norm_stderr,none": 0.06390760676613884,
358
+ "alias": " - ceval-valid_teacher_qualification"
359
+ },
360
+ "ceval-valid_urban_and_rural_planner": {
361
+ "acc,none": 0.21739130434782608,
362
+ "acc_stderr,none": 0.061487546190134544,
363
+ "acc_norm,none": 0.21739130434782608,
364
+ "acc_norm_stderr,none": 0.061487546190134544,
365
+ "alias": " - ceval-valid_urban_and_rural_planner"
366
+ },
367
+ "ceval-valid_veterinary_medicine": {
368
+ "acc,none": 0.13043478260869565,
369
+ "acc_stderr,none": 0.07180198468215396,
370
+ "acc_norm,none": 0.13043478260869565,
371
+ "acc_norm_stderr,none": 0.07180198468215396,
372
+ "alias": " - ceval-valid_veterinary_medicine"
373
+ }
374
+ },
375
+ "groups": {
376
+ "ceval-valid": {
377
+ "acc,none": 0.2526002971768202,
378
+ "acc_stderr,none": 0.11248875724999531,
379
+ "acc_norm,none": 0.2526002971768202,
380
+ "acc_norm_stderr,none": 0.11248875724999531,
381
+ "alias": "ceval-valid"
382
+ }
383
+ },
384
+ "configs": {
385
+ "ceval-valid_accountant": {
386
+ "task": "ceval-valid_accountant",
387
+ "group": "ceval-valid",
388
+ "dataset_path": "ceval/ceval-exam",
389
+ "dataset_name": "accountant",
390
+ "validation_split": "val",
391
+ "fewshot_split": "dev",
392
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
393
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
394
+ "doc_to_choice": [
395
+ "A",
396
+ "B",
397
+ "C",
398
+ "D"
399
+ ],
400
+ "description": "以下是中国关于注册会计师的单项选择题,请选出其中的正确答案。\n\n",
401
+ "target_delimiter": " ",
402
+ "fewshot_delimiter": "\n\n",
403
+ "fewshot_config": {
404
+ "sampler": "first_n"
405
+ },
406
+ "metric_list": [
407
+ {
408
+ "metric": "acc",
409
+ "aggregation": "mean",
410
+ "higher_is_better": true
411
+ },
412
+ {
413
+ "metric": "acc_norm",
414
+ "aggregation": "mean",
415
+ "higher_is_better": true
416
+ }
417
+ ],
418
+ "output_type": "multiple_choice",
419
+ "repeats": 1,
420
+ "should_decontaminate": false,
421
+ "metadata": {
422
+ "version": 1.0
423
+ }
424
+ },
425
+ "ceval-valid_advanced_mathematics": {
426
+ "task": "ceval-valid_advanced_mathematics",
427
+ "group": "ceval-valid",
428
+ "dataset_path": "ceval/ceval-exam",
429
+ "dataset_name": "advanced_mathematics",
430
+ "validation_split": "val",
431
+ "fewshot_split": "dev",
432
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
433
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
434
+ "doc_to_choice": [
435
+ "A",
436
+ "B",
437
+ "C",
438
+ "D"
439
+ ],
440
+ "description": "以下是中国关于高等数学的单项选择题,请选出其中的正确答案。\n\n",
441
+ "target_delimiter": " ",
442
+ "fewshot_delimiter": "\n\n",
443
+ "fewshot_config": {
444
+ "sampler": "first_n"
445
+ },
446
+ "metric_list": [
447
+ {
448
+ "metric": "acc",
449
+ "aggregation": "mean",
450
+ "higher_is_better": true
451
+ },
452
+ {
453
+ "metric": "acc_norm",
454
+ "aggregation": "mean",
455
+ "higher_is_better": true
456
+ }
457
+ ],
458
+ "output_type": "multiple_choice",
459
+ "repeats": 1,
460
+ "should_decontaminate": false,
461
+ "metadata": {
462
+ "version": 1.0
463
+ }
464
+ },
465
+ "ceval-valid_art_studies": {
466
+ "task": "ceval-valid_art_studies",
467
+ "group": "ceval-valid",
468
+ "dataset_path": "ceval/ceval-exam",
469
+ "dataset_name": "art_studies",
470
+ "validation_split": "val",
471
+ "fewshot_split": "dev",
472
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
473
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
474
+ "doc_to_choice": [
475
+ "A",
476
+ "B",
477
+ "C",
478
+ "D"
479
+ ],
480
+ "description": "以下是中国关于艺术学的单项选择题,请选出其中的正确答案。\n\n",
481
+ "target_delimiter": " ",
482
+ "fewshot_delimiter": "\n\n",
483
+ "fewshot_config": {
484
+ "sampler": "first_n"
485
+ },
486
+ "metric_list": [
487
+ {
488
+ "metric": "acc",
489
+ "aggregation": "mean",
490
+ "higher_is_better": true
491
+ },
492
+ {
493
+ "metric": "acc_norm",
494
+ "aggregation": "mean",
495
+ "higher_is_better": true
496
+ }
497
+ ],
498
+ "output_type": "multiple_choice",
499
+ "repeats": 1,
500
+ "should_decontaminate": false,
501
+ "metadata": {
502
+ "version": 1.0
503
+ }
504
+ },
505
+ "ceval-valid_basic_medicine": {
506
+ "task": "ceval-valid_basic_medicine",
507
+ "group": "ceval-valid",
508
+ "dataset_path": "ceval/ceval-exam",
509
+ "dataset_name": "basic_medicine",
510
+ "validation_split": "val",
511
+ "fewshot_split": "dev",
512
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
513
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
514
+ "doc_to_choice": [
515
+ "A",
516
+ "B",
517
+ "C",
518
+ "D"
519
+ ],
520
+ "description": "以下是中国关于基础医学的单项选择题,请选出其中的正确答案。\n\n",
521
+ "target_delimiter": " ",
522
+ "fewshot_delimiter": "\n\n",
523
+ "fewshot_config": {
524
+ "sampler": "first_n"
525
+ },
526
+ "metric_list": [
527
+ {
528
+ "metric": "acc",
529
+ "aggregation": "mean",
530
+ "higher_is_better": true
531
+ },
532
+ {
533
+ "metric": "acc_norm",
534
+ "aggregation": "mean",
535
+ "higher_is_better": true
536
+ }
537
+ ],
538
+ "output_type": "multiple_choice",
539
+ "repeats": 1,
540
+ "should_decontaminate": false,
541
+ "metadata": {
542
+ "version": 1.0
543
+ }
544
+ },
545
+ "ceval-valid_business_administration": {
546
+ "task": "ceval-valid_business_administration",
547
+ "group": "ceval-valid",
548
+ "dataset_path": "ceval/ceval-exam",
549
+ "dataset_name": "business_administration",
550
+ "validation_split": "val",
551
+ "fewshot_split": "dev",
552
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
553
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
554
+ "doc_to_choice": [
555
+ "A",
556
+ "B",
557
+ "C",
558
+ "D"
559
+ ],
560
+ "description": "以下是中国关于工商管理的单项选择题,请选出其中的正确答案。\n\n",
561
+ "target_delimiter": " ",
562
+ "fewshot_delimiter": "\n\n",
563
+ "fewshot_config": {
564
+ "sampler": "first_n"
565
+ },
566
+ "metric_list": [
567
+ {
568
+ "metric": "acc",
569
+ "aggregation": "mean",
570
+ "higher_is_better": true
571
+ },
572
+ {
573
+ "metric": "acc_norm",
574
+ "aggregation": "mean",
575
+ "higher_is_better": true
576
+ }
577
+ ],
578
+ "output_type": "multiple_choice",
579
+ "repeats": 1,
580
+ "should_decontaminate": false,
581
+ "metadata": {
582
+ "version": 1.0
583
+ }
584
+ },
585
+ "ceval-valid_chinese_language_and_literature": {
586
+ "task": "ceval-valid_chinese_language_and_literature",
587
+ "group": "ceval-valid",
588
+ "dataset_path": "ceval/ceval-exam",
589
+ "dataset_name": "chinese_language_and_literature",
590
+ "validation_split": "val",
591
+ "fewshot_split": "dev",
592
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
593
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
594
+ "doc_to_choice": [
595
+ "A",
596
+ "B",
597
+ "C",
598
+ "D"
599
+ ],
600
+ "description": "以下是中国关于中国语言文学的单项选择题,请选出其中的正确答案。\n\n",
601
+ "target_delimiter": " ",
602
+ "fewshot_delimiter": "\n\n",
603
+ "fewshot_config": {
604
+ "sampler": "first_n"
605
+ },
606
+ "metric_list": [
607
+ {
608
+ "metric": "acc",
609
+ "aggregation": "mean",
610
+ "higher_is_better": true
611
+ },
612
+ {
613
+ "metric": "acc_norm",
614
+ "aggregation": "mean",
615
+ "higher_is_better": true
616
+ }
617
+ ],
618
+ "output_type": "multiple_choice",
619
+ "repeats": 1,
620
+ "should_decontaminate": false,
621
+ "metadata": {
622
+ "version": 1.0
623
+ }
624
+ },
625
+ "ceval-valid_civil_servant": {
626
+ "task": "ceval-valid_civil_servant",
627
+ "group": "ceval-valid",
628
+ "dataset_path": "ceval/ceval-exam",
629
+ "dataset_name": "civil_servant",
630
+ "validation_split": "val",
631
+ "fewshot_split": "dev",
632
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
633
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
634
+ "doc_to_choice": [
635
+ "A",
636
+ "B",
637
+ "C",
638
+ "D"
639
+ ],
640
+ "description": "以下是中国关于公务员的单项选择题,请选出其中的正确答案。\n\n",
641
+ "target_delimiter": " ",
642
+ "fewshot_delimiter": "\n\n",
643
+ "fewshot_config": {
644
+ "sampler": "first_n"
645
+ },
646
+ "metric_list": [
647
+ {
648
+ "metric": "acc",
649
+ "aggregation": "mean",
650
+ "higher_is_better": true
651
+ },
652
+ {
653
+ "metric": "acc_norm",
654
+ "aggregation": "mean",
655
+ "higher_is_better": true
656
+ }
657
+ ],
658
+ "output_type": "multiple_choice",
659
+ "repeats": 1,
660
+ "should_decontaminate": false,
661
+ "metadata": {
662
+ "version": 1.0
663
+ }
664
+ },
665
+ "ceval-valid_clinical_medicine": {
666
+ "task": "ceval-valid_clinical_medicine",
667
+ "group": "ceval-valid",
668
+ "dataset_path": "ceval/ceval-exam",
669
+ "dataset_name": "clinical_medicine",
670
+ "validation_split": "val",
671
+ "fewshot_split": "dev",
672
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
673
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
674
+ "doc_to_choice": [
675
+ "A",
676
+ "B",
677
+ "C",
678
+ "D"
679
+ ],
680
+ "description": "以下是中国关于临床医学的单项选择题,请选出其中的正确答案。\n\n",
681
+ "target_delimiter": " ",
682
+ "fewshot_delimiter": "\n\n",
683
+ "fewshot_config": {
684
+ "sampler": "first_n"
685
+ },
686
+ "metric_list": [
687
+ {
688
+ "metric": "acc",
689
+ "aggregation": "mean",
690
+ "higher_is_better": true
691
+ },
692
+ {
693
+ "metric": "acc_norm",
694
+ "aggregation": "mean",
695
+ "higher_is_better": true
696
+ }
697
+ ],
698
+ "output_type": "multiple_choice",
699
+ "repeats": 1,
700
+ "should_decontaminate": false,
701
+ "metadata": {
702
+ "version": 1.0
703
+ }
704
+ },
705
+ "ceval-valid_college_chemistry": {
706
+ "task": "ceval-valid_college_chemistry",
707
+ "group": "ceval-valid",
708
+ "dataset_path": "ceval/ceval-exam",
709
+ "dataset_name": "college_chemistry",
710
+ "validation_split": "val",
711
+ "fewshot_split": "dev",
712
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
713
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
714
+ "doc_to_choice": [
715
+ "A",
716
+ "B",
717
+ "C",
718
+ "D"
719
+ ],
720
+ "description": "以下是中国关于大学化���的单项选择题,请选出其中的正确答案。\n\n",
721
+ "target_delimiter": " ",
722
+ "fewshot_delimiter": "\n\n",
723
+ "fewshot_config": {
724
+ "sampler": "first_n"
725
+ },
726
+ "metric_list": [
727
+ {
728
+ "metric": "acc",
729
+ "aggregation": "mean",
730
+ "higher_is_better": true
731
+ },
732
+ {
733
+ "metric": "acc_norm",
734
+ "aggregation": "mean",
735
+ "higher_is_better": true
736
+ }
737
+ ],
738
+ "output_type": "multiple_choice",
739
+ "repeats": 1,
740
+ "should_decontaminate": false,
741
+ "metadata": {
742
+ "version": 1.0
743
+ }
744
+ },
745
+ "ceval-valid_college_economics": {
746
+ "task": "ceval-valid_college_economics",
747
+ "group": "ceval-valid",
748
+ "dataset_path": "ceval/ceval-exam",
749
+ "dataset_name": "college_economics",
750
+ "validation_split": "val",
751
+ "fewshot_split": "dev",
752
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
753
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
754
+ "doc_to_choice": [
755
+ "A",
756
+ "B",
757
+ "C",
758
+ "D"
759
+ ],
760
+ "description": "以下是中国关于大学经济学的单项选择题,请选出其中的正确答案。\n\n",
761
+ "target_delimiter": " ",
762
+ "fewshot_delimiter": "\n\n",
763
+ "fewshot_config": {
764
+ "sampler": "first_n"
765
+ },
766
+ "metric_list": [
767
+ {
768
+ "metric": "acc",
769
+ "aggregation": "mean",
770
+ "higher_is_better": true
771
+ },
772
+ {
773
+ "metric": "acc_norm",
774
+ "aggregation": "mean",
775
+ "higher_is_better": true
776
+ }
777
+ ],
778
+ "output_type": "multiple_choice",
779
+ "repeats": 1,
780
+ "should_decontaminate": false,
781
+ "metadata": {
782
+ "version": 1.0
783
+ }
784
+ },
785
+ "ceval-valid_college_physics": {
786
+ "task": "ceval-valid_college_physics",
787
+ "group": "ceval-valid",
788
+ "dataset_path": "ceval/ceval-exam",
789
+ "dataset_name": "college_physics",
790
+ "validation_split": "val",
791
+ "fewshot_split": "dev",
792
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
793
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
794
+ "doc_to_choice": [
795
+ "A",
796
+ "B",
797
+ "C",
798
+ "D"
799
+ ],
800
+ "description": "以下是中国关于大学物理的单项选择题,请选出其中的正确答案。\n\n",
801
+ "target_delimiter": " ",
802
+ "fewshot_delimiter": "\n\n",
803
+ "fewshot_config": {
804
+ "sampler": "first_n"
805
+ },
806
+ "metric_list": [
807
+ {
808
+ "metric": "acc",
809
+ "aggregation": "mean",
810
+ "higher_is_better": true
811
+ },
812
+ {
813
+ "metric": "acc_norm",
814
+ "aggregation": "mean",
815
+ "higher_is_better": true
816
+ }
817
+ ],
818
+ "output_type": "multiple_choice",
819
+ "repeats": 1,
820
+ "should_decontaminate": false,
821
+ "metadata": {
822
+ "version": 1.0
823
+ }
824
+ },
825
+ "ceval-valid_college_programming": {
826
+ "task": "ceval-valid_college_programming",
827
+ "group": "ceval-valid",
828
+ "dataset_path": "ceval/ceval-exam",
829
+ "dataset_name": "college_programming",
830
+ "validation_split": "val",
831
+ "fewshot_split": "dev",
832
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
833
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
834
+ "doc_to_choice": [
835
+ "A",
836
+ "B",
837
+ "C",
838
+ "D"
839
+ ],
840
+ "description": "以下是中国关于大学编程的单项选择题,请选出其中的正确答案。\n\n",
841
+ "target_delimiter": " ",
842
+ "fewshot_delimiter": "\n\n",
843
+ "fewshot_config": {
844
+ "sampler": "first_n"
845
+ },
846
+ "metric_list": [
847
+ {
848
+ "metric": "acc",
849
+ "aggregation": "mean",
850
+ "higher_is_better": true
851
+ },
852
+ {
853
+ "metric": "acc_norm",
854
+ "aggregation": "mean",
855
+ "higher_is_better": true
856
+ }
857
+ ],
858
+ "output_type": "multiple_choice",
859
+ "repeats": 1,
860
+ "should_decontaminate": false,
861
+ "metadata": {
862
+ "version": 1.0
863
+ }
864
+ },
865
+ "ceval-valid_computer_architecture": {
866
+ "task": "ceval-valid_computer_architecture",
867
+ "group": "ceval-valid",
868
+ "dataset_path": "ceval/ceval-exam",
869
+ "dataset_name": "computer_architecture",
870
+ "validation_split": "val",
871
+ "fewshot_split": "dev",
872
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
873
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
874
+ "doc_to_choice": [
875
+ "A",
876
+ "B",
877
+ "C",
878
+ "D"
879
+ ],
880
+ "description": "以下是中国关于计算机组成的单项选择题,请选出其中的正确答案。\n\n",
881
+ "target_delimiter": " ",
882
+ "fewshot_delimiter": "\n\n",
883
+ "fewshot_config": {
884
+ "sampler": "first_n"
885
+ },
886
+ "metric_list": [
887
+ {
888
+ "metric": "acc",
889
+ "aggregation": "mean",
890
+ "higher_is_better": true
891
+ },
892
+ {
893
+ "metric": "acc_norm",
894
+ "aggregation": "mean",
895
+ "higher_is_better": true
896
+ }
897
+ ],
898
+ "output_type": "multiple_choice",
899
+ "repeats": 1,
900
+ "should_decontaminate": false,
901
+ "metadata": {
902
+ "version": 1.0
903
+ }
904
+ },
905
+ "ceval-valid_computer_network": {
906
+ "task": "ceval-valid_computer_network",
907
+ "group": "ceval-valid",
908
+ "dataset_path": "ceval/ceval-exam",
909
+ "dataset_name": "computer_network",
910
+ "validation_split": "val",
911
+ "fewshot_split": "dev",
912
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
913
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
914
+ "doc_to_choice": [
915
+ "A",
916
+ "B",
917
+ "C",
918
+ "D"
919
+ ],
920
+ "description": "以下是中国关于计算机网络的单项选择题,请选出其中的正确答案。\n\n",
921
+ "target_delimiter": " ",
922
+ "fewshot_delimiter": "\n\n",
923
+ "fewshot_config": {
924
+ "sampler": "first_n"
925
+ },
926
+ "metric_list": [
927
+ {
928
+ "metric": "acc",
929
+ "aggregation": "mean",
930
+ "higher_is_better": true
931
+ },
932
+ {
933
+ "metric": "acc_norm",
934
+ "aggregation": "mean",
935
+ "higher_is_better": true
936
+ }
937
+ ],
938
+ "output_type": "multiple_choice",
939
+ "repeats": 1,
940
+ "should_decontaminate": false,
941
+ "metadata": {
942
+ "version": 1.0
943
+ }
944
+ },
945
+ "ceval-valid_discrete_mathematics": {
946
+ "task": "ceval-valid_discrete_mathematics",
947
+ "group": "ceval-valid",
948
+ "dataset_path": "ceval/ceval-exam",
949
+ "dataset_name": "discrete_mathematics",
950
+ "validation_split": "val",
951
+ "fewshot_split": "dev",
952
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
953
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
954
+ "doc_to_choice": [
955
+ "A",
956
+ "B",
957
+ "C",
958
+ "D"
959
+ ],
960
+ "description": "以下是中国关于离散数学的单项选择题,请选出其中的正确答案。\n\n",
961
+ "target_delimiter": " ",
962
+ "fewshot_delimiter": "\n\n",
963
+ "fewshot_config": {
964
+ "sampler": "first_n"
965
+ },
966
+ "metric_list": [
967
+ {
968
+ "metric": "acc",
969
+ "aggregation": "mean",
970
+ "higher_is_better": true
971
+ },
972
+ {
973
+ "metric": "acc_norm",
974
+ "aggregation": "mean",
975
+ "higher_is_better": true
976
+ }
977
+ ],
978
+ "output_type": "multiple_choice",
979
+ "repeats": 1,
980
+ "should_decontaminate": false,
981
+ "metadata": {
982
+ "version": 1.0
983
+ }
984
+ },
985
+ "ceval-valid_education_science": {
986
+ "task": "ceval-valid_education_science",
987
+ "group": "ceval-valid",
988
+ "dataset_path": "ceval/ceval-exam",
989
+ "dataset_name": "education_science",
990
+ "validation_split": "val",
991
+ "fewshot_split": "dev",
992
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
993
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
994
+ "doc_to_choice": [
995
+ "A",
996
+ "B",
997
+ "C",
998
+ "D"
999
+ ],
1000
+ "description": "以下是中国关于教育学的单项选择题,请选出其中的正确答案。\n\n",
1001
+ "target_delimiter": " ",
1002
+ "fewshot_delimiter": "\n\n",
1003
+ "fewshot_config": {
1004
+ "sampler": "first_n"
1005
+ },
1006
+ "metric_list": [
1007
+ {
1008
+ "metric": "acc",
1009
+ "aggregation": "mean",
1010
+ "higher_is_better": true
1011
+ },
1012
+ {
1013
+ "metric": "acc_norm",
1014
+ "aggregation": "mean",
1015
+ "higher_is_better": true
1016
+ }
1017
+ ],
1018
+ "output_type": "multiple_choice",
1019
+ "repeats": 1,
1020
+ "should_decontaminate": false,
1021
+ "metadata": {
1022
+ "version": 1.0
1023
+ }
1024
+ },
1025
+ "ceval-valid_electrical_engineer": {
1026
+ "task": "ceval-valid_electrical_engineer",
1027
+ "group": "ceval-valid",
1028
+ "dataset_path": "ceval/ceval-exam",
1029
+ "dataset_name": "electrical_engineer",
1030
+ "validation_split": "val",
1031
+ "fewshot_split": "dev",
1032
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1033
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1034
+ "doc_to_choice": [
1035
+ "A",
1036
+ "B",
1037
+ "C",
1038
+ "D"
1039
+ ],
1040
+ "description": "以下是中国关于注册电气工程师的单项选择题,请选出其中的正确答案。\n\n",
1041
+ "target_delimiter": " ",
1042
+ "fewshot_delimiter": "\n\n",
1043
+ "fewshot_config": {
1044
+ "sampler": "first_n"
1045
+ },
1046
+ "metric_list": [
1047
+ {
1048
+ "metric": "acc",
1049
+ "aggregation": "mean",
1050
+ "higher_is_better": true
1051
+ },
1052
+ {
1053
+ "metric": "acc_norm",
1054
+ "aggregation": "mean",
1055
+ "higher_is_better": true
1056
+ }
1057
+ ],
1058
+ "output_type": "multiple_choice",
1059
+ "repeats": 1,
1060
+ "should_decontaminate": false,
1061
+ "metadata": {
1062
+ "version": 1.0
1063
+ }
1064
+ },
1065
+ "ceval-valid_environmental_impact_assessment_engineer": {
1066
+ "task": "ceval-valid_environmental_impact_assessment_engineer",
1067
+ "group": "ceval-valid",
1068
+ "dataset_path": "ceval/ceval-exam",
1069
+ "dataset_name": "environmental_impact_assessment_engineer",
1070
+ "validation_split": "val",
1071
+ "fewshot_split": "dev",
1072
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1073
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1074
+ "doc_to_choice": [
1075
+ "A",
1076
+ "B",
1077
+ "C",
1078
+ "D"
1079
+ ],
1080
+ "description": "以下是中国关于环境影响评价工程师的单项选择题,请选出其中的正确答案。\n\n",
1081
+ "target_delimiter": " ",
1082
+ "fewshot_delimiter": "\n\n",
1083
+ "fewshot_config": {
1084
+ "sampler": "first_n"
1085
+ },
1086
+ "metric_list": [
1087
+ {
1088
+ "metric": "acc",
1089
+ "aggregation": "mean",
1090
+ "higher_is_better": true
1091
+ },
1092
+ {
1093
+ "metric": "acc_norm",
1094
+ "aggregation": "mean",
1095
+ "higher_is_better": true
1096
+ }
1097
+ ],
1098
+ "output_type": "multiple_choice",
1099
+ "repeats": 1,
1100
+ "should_decontaminate": false,
1101
+ "metadata": {
1102
+ "version": 1.0
1103
+ }
1104
+ },
1105
+ "ceval-valid_fire_engineer": {
1106
+ "task": "ceval-valid_fire_engineer",
1107
+ "group": "ceval-valid",
1108
+ "dataset_path": "ceval/ceval-exam",
1109
+ "dataset_name": "fire_engineer",
1110
+ "validation_split": "val",
1111
+ "fewshot_split": "dev",
1112
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1113
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1114
+ "doc_to_choice": [
1115
+ "A",
1116
+ "B",
1117
+ "C",
1118
+ "D"
1119
+ ],
1120
+ "description": "以下是中国关于注册消防工程师的单项选择题,请选出其中的正确答案。\n\n",
1121
+ "target_delimiter": " ",
1122
+ "fewshot_delimiter": "\n\n",
1123
+ "fewshot_config": {
1124
+ "sampler": "first_n"
1125
+ },
1126
+ "metric_list": [
1127
+ {
1128
+ "metric": "acc",
1129
+ "aggregation": "mean",
1130
+ "higher_is_better": true
1131
+ },
1132
+ {
1133
+ "metric": "acc_norm",
1134
+ "aggregation": "mean",
1135
+ "higher_is_better": true
1136
+ }
1137
+ ],
1138
+ "output_type": "multiple_choice",
1139
+ "repeats": 1,
1140
+ "should_decontaminate": false,
1141
+ "metadata": {
1142
+ "version": 1.0
1143
+ }
1144
+ },
1145
+ "ceval-valid_high_school_biology": {
1146
+ "task": "ceval-valid_high_school_biology",
1147
+ "group": "ceval-valid",
1148
+ "dataset_path": "ceval/ceval-exam",
1149
+ "dataset_name": "high_school_biology",
1150
+ "validation_split": "val",
1151
+ "fewshot_split": "dev",
1152
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1153
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1154
+ "doc_to_choice": [
1155
+ "A",
1156
+ "B",
1157
+ "C",
1158
+ "D"
1159
+ ],
1160
+ "description": "以下是中国关于高中生物的单项选择题,请选出其中的正确答案。\n\n",
1161
+ "target_delimiter": " ",
1162
+ "fewshot_delimiter": "\n\n",
1163
+ "fewshot_config": {
1164
+ "sampler": "first_n"
1165
+ },
1166
+ "metric_list": [
1167
+ {
1168
+ "metric": "acc",
1169
+ "aggregation": "mean",
1170
+ "higher_is_better": true
1171
+ },
1172
+ {
1173
+ "metric": "acc_norm",
1174
+ "aggregation": "mean",
1175
+ "higher_is_better": true
1176
+ }
1177
+ ],
1178
+ "output_type": "multiple_choice",
1179
+ "repeats": 1,
1180
+ "should_decontaminate": false,
1181
+ "metadata": {
1182
+ "version": 1.0
1183
+ }
1184
+ },
1185
+ "ceval-valid_high_school_chemistry": {
1186
+ "task": "ceval-valid_high_school_chemistry",
1187
+ "group": "ceval-valid",
1188
+ "dataset_path": "ceval/ceval-exam",
1189
+ "dataset_name": "high_school_chemistry",
1190
+ "validation_split": "val",
1191
+ "fewshot_split": "dev",
1192
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1193
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1194
+ "doc_to_choice": [
1195
+ "A",
1196
+ "B",
1197
+ "C",
1198
+ "D"
1199
+ ],
1200
+ "description": "以下是中国关于高中化学的单项选择题,请选出其中的正确答案。\n\n",
1201
+ "target_delimiter": " ",
1202
+ "fewshot_delimiter": "\n\n",
1203
+ "fewshot_config": {
1204
+ "sampler": "first_n"
1205
+ },
1206
+ "metric_list": [
1207
+ {
1208
+ "metric": "acc",
1209
+ "aggregation": "mean",
1210
+ "higher_is_better": true
1211
+ },
1212
+ {
1213
+ "metric": "acc_norm",
1214
+ "aggregation": "mean",
1215
+ "higher_is_better": true
1216
+ }
1217
+ ],
1218
+ "output_type": "multiple_choice",
1219
+ "repeats": 1,
1220
+ "should_decontaminate": false,
1221
+ "metadata": {
1222
+ "version": 1.0
1223
+ }
1224
+ },
1225
+ "ceval-valid_high_school_chinese": {
1226
+ "task": "ceval-valid_high_school_chinese",
1227
+ "group": "ceval-valid",
1228
+ "dataset_path": "ceval/ceval-exam",
1229
+ "dataset_name": "high_school_chinese",
1230
+ "validation_split": "val",
1231
+ "fewshot_split": "dev",
1232
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1233
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1234
+ "doc_to_choice": [
1235
+ "A",
1236
+ "B",
1237
+ "C",
1238
+ "D"
1239
+ ],
1240
+ "description": "以下是中国关于高中语文的单项选择题,请选出其中的正确答案。\n\n",
1241
+ "target_delimiter": " ",
1242
+ "fewshot_delimiter": "\n\n",
1243
+ "fewshot_config": {
1244
+ "sampler": "first_n"
1245
+ },
1246
+ "metric_list": [
1247
+ {
1248
+ "metric": "acc",
1249
+ "aggregation": "mean",
1250
+ "higher_is_better": true
1251
+ },
1252
+ {
1253
+ "metric": "acc_norm",
1254
+ "aggregation": "mean",
1255
+ "higher_is_better": true
1256
+ }
1257
+ ],
1258
+ "output_type": "multiple_choice",
1259
+ "repeats": 1,
1260
+ "should_decontaminate": false,
1261
+ "metadata": {
1262
+ "version": 1.0
1263
+ }
1264
+ },
1265
+ "ceval-valid_high_school_geography": {
1266
+ "task": "ceval-valid_high_school_geography",
1267
+ "group": "ceval-valid",
1268
+ "dataset_path": "ceval/ceval-exam",
1269
+ "dataset_name": "high_school_geography",
1270
+ "validation_split": "val",
1271
+ "fewshot_split": "dev",
1272
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1273
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1274
+ "doc_to_choice": [
1275
+ "A",
1276
+ "B",
1277
+ "C",
1278
+ "D"
1279
+ ],
1280
+ "description": "以下是中国关于高中地理的单项选择题,请选出其中的正确答案。\n\n",
1281
+ "target_delimiter": " ",
1282
+ "fewshot_delimiter": "\n\n",
1283
+ "fewshot_config": {
1284
+ "sampler": "first_n"
1285
+ },
1286
+ "metric_list": [
1287
+ {
1288
+ "metric": "acc",
1289
+ "aggregation": "mean",
1290
+ "higher_is_better": true
1291
+ },
1292
+ {
1293
+ "metric": "acc_norm",
1294
+ "aggregation": "mean",
1295
+ "higher_is_better": true
1296
+ }
1297
+ ],
1298
+ "output_type": "multiple_choice",
1299
+ "repeats": 1,
1300
+ "should_decontaminate": false,
1301
+ "metadata": {
1302
+ "version": 1.0
1303
+ }
1304
+ },
1305
+ "ceval-valid_high_school_history": {
1306
+ "task": "ceval-valid_high_school_history",
1307
+ "group": "ceval-valid",
1308
+ "dataset_path": "ceval/ceval-exam",
1309
+ "dataset_name": "high_school_history",
1310
+ "validation_split": "val",
1311
+ "fewshot_split": "dev",
1312
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1313
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1314
+ "doc_to_choice": [
1315
+ "A",
1316
+ "B",
1317
+ "C",
1318
+ "D"
1319
+ ],
1320
+ "description": "以下是中国关于高中历史的单项选择题,请选出其中的正确答案。\n\n",
1321
+ "target_delimiter": " ",
1322
+ "fewshot_delimiter": "\n\n",
1323
+ "fewshot_config": {
1324
+ "sampler": "first_n"
1325
+ },
1326
+ "metric_list": [
1327
+ {
1328
+ "metric": "acc",
1329
+ "aggregation": "mean",
1330
+ "higher_is_better": true
1331
+ },
1332
+ {
1333
+ "metric": "acc_norm",
1334
+ "aggregation": "mean",
1335
+ "higher_is_better": true
1336
+ }
1337
+ ],
1338
+ "output_type": "multiple_choice",
1339
+ "repeats": 1,
1340
+ "should_decontaminate": false,
1341
+ "metadata": {
1342
+ "version": 1.0
1343
+ }
1344
+ },
1345
+ "ceval-valid_high_school_mathematics": {
1346
+ "task": "ceval-valid_high_school_mathematics",
1347
+ "group": "ceval-valid",
1348
+ "dataset_path": "ceval/ceval-exam",
1349
+ "dataset_name": "high_school_mathematics",
1350
+ "validation_split": "val",
1351
+ "fewshot_split": "dev",
1352
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1353
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1354
+ "doc_to_choice": [
1355
+ "A",
1356
+ "B",
1357
+ "C",
1358
+ "D"
1359
+ ],
1360
+ "description": "以下是中国关于高中数学的单项选择题,请选出其中的正确答案。\n\n",
1361
+ "target_delimiter": " ",
1362
+ "fewshot_delimiter": "\n\n",
1363
+ "fewshot_config": {
1364
+ "sampler": "first_n"
1365
+ },
1366
+ "metric_list": [
1367
+ {
1368
+ "metric": "acc",
1369
+ "aggregation": "mean",
1370
+ "higher_is_better": true
1371
+ },
1372
+ {
1373
+ "metric": "acc_norm",
1374
+ "aggregation": "mean",
1375
+ "higher_is_better": true
1376
+ }
1377
+ ],
1378
+ "output_type": "multiple_choice",
1379
+ "repeats": 1,
1380
+ "should_decontaminate": false,
1381
+ "metadata": {
1382
+ "version": 1.0
1383
+ }
1384
+ },
1385
+ "ceval-valid_high_school_physics": {
1386
+ "task": "ceval-valid_high_school_physics",
1387
+ "group": "ceval-valid",
1388
+ "dataset_path": "ceval/ceval-exam",
1389
+ "dataset_name": "high_school_physics",
1390
+ "validation_split": "val",
1391
+ "fewshot_split": "dev",
1392
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1393
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1394
+ "doc_to_choice": [
1395
+ "A",
1396
+ "B",
1397
+ "C",
1398
+ "D"
1399
+ ],
1400
+ "description": "以下是中国关于高中物理的单项选择题,请选出其中的正确答案。\n\n",
1401
+ "target_delimiter": " ",
1402
+ "fewshot_delimiter": "\n\n",
1403
+ "fewshot_config": {
1404
+ "sampler": "first_n"
1405
+ },
1406
+ "metric_list": [
1407
+ {
1408
+ "metric": "acc",
1409
+ "aggregation": "mean",
1410
+ "higher_is_better": true
1411
+ },
1412
+ {
1413
+ "metric": "acc_norm",
1414
+ "aggregation": "mean",
1415
+ "higher_is_better": true
1416
+ }
1417
+ ],
1418
+ "output_type": "multiple_choice",
1419
+ "repeats": 1,
1420
+ "should_decontaminate": false,
1421
+ "metadata": {
1422
+ "version": 1.0
1423
+ }
1424
+ },
1425
+ "ceval-valid_high_school_politics": {
1426
+ "task": "ceval-valid_high_school_politics",
1427
+ "group": "ceval-valid",
1428
+ "dataset_path": "ceval/ceval-exam",
1429
+ "dataset_name": "high_school_politics",
1430
+ "validation_split": "val",
1431
+ "fewshot_split": "dev",
1432
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1433
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1434
+ "doc_to_choice": [
1435
+ "A",
1436
+ "B",
1437
+ "C",
1438
+ "D"
1439
+ ],
1440
+ "description": "以下是中国关于高中政治的单项选择题,请选出其中的正确答案。\n\n",
1441
+ "target_delimiter": " ",
1442
+ "fewshot_delimiter": "\n\n",
1443
+ "fewshot_config": {
1444
+ "sampler": "first_n"
1445
+ },
1446
+ "metric_list": [
1447
+ {
1448
+ "metric": "acc",
1449
+ "aggregation": "mean",
1450
+ "higher_is_better": true
1451
+ },
1452
+ {
1453
+ "metric": "acc_norm",
1454
+ "aggregation": "mean",
1455
+ "higher_is_better": true
1456
+ }
1457
+ ],
1458
+ "output_type": "multiple_choice",
1459
+ "repeats": 1,
1460
+ "should_decontaminate": false,
1461
+ "metadata": {
1462
+ "version": 1.0
1463
+ }
1464
+ },
1465
+ "ceval-valid_ideological_and_moral_cultivation": {
1466
+ "task": "ceval-valid_ideological_and_moral_cultivation",
1467
+ "group": "ceval-valid",
1468
+ "dataset_path": "ceval/ceval-exam",
1469
+ "dataset_name": "ideological_and_moral_cultivation",
1470
+ "validation_split": "val",
1471
+ "fewshot_split": "dev",
1472
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1473
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1474
+ "doc_to_choice": [
1475
+ "A",
1476
+ "B",
1477
+ "C",
1478
+ "D"
1479
+ ],
1480
+ "description": "以下是中国关于思想道德修养与法律基础的单项选择题,请选出其中的正确答案。\n\n",
1481
+ "target_delimiter": " ",
1482
+ "fewshot_delimiter": "\n\n",
1483
+ "fewshot_config": {
1484
+ "sampler": "first_n"
1485
+ },
1486
+ "metric_list": [
1487
+ {
1488
+ "metric": "acc",
1489
+ "aggregation": "mean",
1490
+ "higher_is_better": true
1491
+ },
1492
+ {
1493
+ "metric": "acc_norm",
1494
+ "aggregation": "mean",
1495
+ "higher_is_better": true
1496
+ }
1497
+ ],
1498
+ "output_type": "multiple_choice",
1499
+ "repeats": 1,
1500
+ "should_decontaminate": false,
1501
+ "metadata": {
1502
+ "version": 1.0
1503
+ }
1504
+ },
1505
+ "ceval-valid_law": {
1506
+ "task": "ceval-valid_law",
1507
+ "group": "ceval-valid",
1508
+ "dataset_path": "ceval/ceval-exam",
1509
+ "dataset_name": "law",
1510
+ "validation_split": "val",
1511
+ "fewshot_split": "dev",
1512
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1513
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1514
+ "doc_to_choice": [
1515
+ "A",
1516
+ "B",
1517
+ "C",
1518
+ "D"
1519
+ ],
1520
+ "description": "以下是中国关于法学的单项选择题,请选出其中的正确答案。\n\n",
1521
+ "target_delimiter": " ",
1522
+ "fewshot_delimiter": "\n\n",
1523
+ "fewshot_config": {
1524
+ "sampler": "first_n"
1525
+ },
1526
+ "metric_list": [
1527
+ {
1528
+ "metric": "acc",
1529
+ "aggregation": "mean",
1530
+ "higher_is_better": true
1531
+ },
1532
+ {
1533
+ "metric": "acc_norm",
1534
+ "aggregation": "mean",
1535
+ "higher_is_better": true
1536
+ }
1537
+ ],
1538
+ "output_type": "multiple_choice",
1539
+ "repeats": 1,
1540
+ "should_decontaminate": false,
1541
+ "metadata": {
1542
+ "version": 1.0
1543
+ }
1544
+ },
1545
+ "ceval-valid_legal_professional": {
1546
+ "task": "ceval-valid_legal_professional",
1547
+ "group": "ceval-valid",
1548
+ "dataset_path": "ceval/ceval-exam",
1549
+ "dataset_name": "legal_professional",
1550
+ "validation_split": "val",
1551
+ "fewshot_split": "dev",
1552
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答��:",
1553
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1554
+ "doc_to_choice": [
1555
+ "A",
1556
+ "B",
1557
+ "C",
1558
+ "D"
1559
+ ],
1560
+ "description": "以下是中国关于法律职业资格的单项选择题,请选出其中的正确答案。\n\n",
1561
+ "target_delimiter": " ",
1562
+ "fewshot_delimiter": "\n\n",
1563
+ "fewshot_config": {
1564
+ "sampler": "first_n"
1565
+ },
1566
+ "metric_list": [
1567
+ {
1568
+ "metric": "acc",
1569
+ "aggregation": "mean",
1570
+ "higher_is_better": true
1571
+ },
1572
+ {
1573
+ "metric": "acc_norm",
1574
+ "aggregation": "mean",
1575
+ "higher_is_better": true
1576
+ }
1577
+ ],
1578
+ "output_type": "multiple_choice",
1579
+ "repeats": 1,
1580
+ "should_decontaminate": false,
1581
+ "metadata": {
1582
+ "version": 1.0
1583
+ }
1584
+ },
1585
+ "ceval-valid_logic": {
1586
+ "task": "ceval-valid_logic",
1587
+ "group": "ceval-valid",
1588
+ "dataset_path": "ceval/ceval-exam",
1589
+ "dataset_name": "logic",
1590
+ "validation_split": "val",
1591
+ "fewshot_split": "dev",
1592
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1593
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1594
+ "doc_to_choice": [
1595
+ "A",
1596
+ "B",
1597
+ "C",
1598
+ "D"
1599
+ ],
1600
+ "description": "以下是中国关于逻辑学的单项选择题,请选出其中的正确答案。\n\n",
1601
+ "target_delimiter": " ",
1602
+ "fewshot_delimiter": "\n\n",
1603
+ "fewshot_config": {
1604
+ "sampler": "first_n"
1605
+ },
1606
+ "metric_list": [
1607
+ {
1608
+ "metric": "acc",
1609
+ "aggregation": "mean",
1610
+ "higher_is_better": true
1611
+ },
1612
+ {
1613
+ "metric": "acc_norm",
1614
+ "aggregation": "mean",
1615
+ "higher_is_better": true
1616
+ }
1617
+ ],
1618
+ "output_type": "multiple_choice",
1619
+ "repeats": 1,
1620
+ "should_decontaminate": false,
1621
+ "metadata": {
1622
+ "version": 1.0
1623
+ }
1624
+ },
1625
+ "ceval-valid_mao_zedong_thought": {
1626
+ "task": "ceval-valid_mao_zedong_thought",
1627
+ "group": "ceval-valid",
1628
+ "dataset_path": "ceval/ceval-exam",
1629
+ "dataset_name": "mao_zedong_thought",
1630
+ "validation_split": "val",
1631
+ "fewshot_split": "dev",
1632
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1633
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1634
+ "doc_to_choice": [
1635
+ "A",
1636
+ "B",
1637
+ "C",
1638
+ "D"
1639
+ ],
1640
+ "description": "以下是中国关于毛泽东思想和中国特色社会主义理论体系概论的单项选择题,请选出其中的正确答案。\n\n",
1641
+ "target_delimiter": " ",
1642
+ "fewshot_delimiter": "\n\n",
1643
+ "fewshot_config": {
1644
+ "sampler": "first_n"
1645
+ },
1646
+ "metric_list": [
1647
+ {
1648
+ "metric": "acc",
1649
+ "aggregation": "mean",
1650
+ "higher_is_better": true
1651
+ },
1652
+ {
1653
+ "metric": "acc_norm",
1654
+ "aggregation": "mean",
1655
+ "higher_is_better": true
1656
+ }
1657
+ ],
1658
+ "output_type": "multiple_choice",
1659
+ "repeats": 1,
1660
+ "should_decontaminate": false,
1661
+ "metadata": {
1662
+ "version": 1.0
1663
+ }
1664
+ },
1665
+ "ceval-valid_marxism": {
1666
+ "task": "ceval-valid_marxism",
1667
+ "group": "ceval-valid",
1668
+ "dataset_path": "ceval/ceval-exam",
1669
+ "dataset_name": "marxism",
1670
+ "validation_split": "val",
1671
+ "fewshot_split": "dev",
1672
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1673
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1674
+ "doc_to_choice": [
1675
+ "A",
1676
+ "B",
1677
+ "C",
1678
+ "D"
1679
+ ],
1680
+ "description": "以下是中国关于马克思主义基本原理的单项选择题,请选出其中的正确答案。\n\n",
1681
+ "target_delimiter": " ",
1682
+ "fewshot_delimiter": "\n\n",
1683
+ "fewshot_config": {
1684
+ "sampler": "first_n"
1685
+ },
1686
+ "metric_list": [
1687
+ {
1688
+ "metric": "acc",
1689
+ "aggregation": "mean",
1690
+ "higher_is_better": true
1691
+ },
1692
+ {
1693
+ "metric": "acc_norm",
1694
+ "aggregation": "mean",
1695
+ "higher_is_better": true
1696
+ }
1697
+ ],
1698
+ "output_type": "multiple_choice",
1699
+ "repeats": 1,
1700
+ "should_decontaminate": false,
1701
+ "metadata": {
1702
+ "version": 1.0
1703
+ }
1704
+ },
1705
+ "ceval-valid_metrology_engineer": {
1706
+ "task": "ceval-valid_metrology_engineer",
1707
+ "group": "ceval-valid",
1708
+ "dataset_path": "ceval/ceval-exam",
1709
+ "dataset_name": "metrology_engineer",
1710
+ "validation_split": "val",
1711
+ "fewshot_split": "dev",
1712
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1713
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1714
+ "doc_to_choice": [
1715
+ "A",
1716
+ "B",
1717
+ "C",
1718
+ "D"
1719
+ ],
1720
+ "description": "以下是中国关于注册计量师的单项选择题,请选出其中的正确答案。\n\n",
1721
+ "target_delimiter": " ",
1722
+ "fewshot_delimiter": "\n\n",
1723
+ "fewshot_config": {
1724
+ "sampler": "first_n"
1725
+ },
1726
+ "metric_list": [
1727
+ {
1728
+ "metric": "acc",
1729
+ "aggregation": "mean",
1730
+ "higher_is_better": true
1731
+ },
1732
+ {
1733
+ "metric": "acc_norm",
1734
+ "aggregation": "mean",
1735
+ "higher_is_better": true
1736
+ }
1737
+ ],
1738
+ "output_type": "multiple_choice",
1739
+ "repeats": 1,
1740
+ "should_decontaminate": false,
1741
+ "metadata": {
1742
+ "version": 1.0
1743
+ }
1744
+ },
1745
+ "ceval-valid_middle_school_biology": {
1746
+ "task": "ceval-valid_middle_school_biology",
1747
+ "group": "ceval-valid",
1748
+ "dataset_path": "ceval/ceval-exam",
1749
+ "dataset_name": "middle_school_biology",
1750
+ "validation_split": "val",
1751
+ "fewshot_split": "dev",
1752
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1753
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1754
+ "doc_to_choice": [
1755
+ "A",
1756
+ "B",
1757
+ "C",
1758
+ "D"
1759
+ ],
1760
+ "description": "以下是中国关于初中生物的单项选择题,请选出其中的正确答案。\n\n",
1761
+ "target_delimiter": " ",
1762
+ "fewshot_delimiter": "\n\n",
1763
+ "fewshot_config": {
1764
+ "sampler": "first_n"
1765
+ },
1766
+ "metric_list": [
1767
+ {
1768
+ "metric": "acc",
1769
+ "aggregation": "mean",
1770
+ "higher_is_better": true
1771
+ },
1772
+ {
1773
+ "metric": "acc_norm",
1774
+ "aggregation": "mean",
1775
+ "higher_is_better": true
1776
+ }
1777
+ ],
1778
+ "output_type": "multiple_choice",
1779
+ "repeats": 1,
1780
+ "should_decontaminate": false,
1781
+ "metadata": {
1782
+ "version": 1.0
1783
+ }
1784
+ },
1785
+ "ceval-valid_middle_school_chemistry": {
1786
+ "task": "ceval-valid_middle_school_chemistry",
1787
+ "group": "ceval-valid",
1788
+ "dataset_path": "ceval/ceval-exam",
1789
+ "dataset_name": "middle_school_chemistry",
1790
+ "validation_split": "val",
1791
+ "fewshot_split": "dev",
1792
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1793
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1794
+ "doc_to_choice": [
1795
+ "A",
1796
+ "B",
1797
+ "C",
1798
+ "D"
1799
+ ],
1800
+ "description": "以下是中国关于初中化学的单项选择题,请选出其中的正确答案。\n\n",
1801
+ "target_delimiter": " ",
1802
+ "fewshot_delimiter": "\n\n",
1803
+ "fewshot_config": {
1804
+ "sampler": "first_n"
1805
+ },
1806
+ "metric_list": [
1807
+ {
1808
+ "metric": "acc",
1809
+ "aggregation": "mean",
1810
+ "higher_is_better": true
1811
+ },
1812
+ {
1813
+ "metric": "acc_norm",
1814
+ "aggregation": "mean",
1815
+ "higher_is_better": true
1816
+ }
1817
+ ],
1818
+ "output_type": "multiple_choice",
1819
+ "repeats": 1,
1820
+ "should_decontaminate": false,
1821
+ "metadata": {
1822
+ "version": 1.0
1823
+ }
1824
+ },
1825
+ "ceval-valid_middle_school_geography": {
1826
+ "task": "ceval-valid_middle_school_geography",
1827
+ "group": "ceval-valid",
1828
+ "dataset_path": "ceval/ceval-exam",
1829
+ "dataset_name": "middle_school_geography",
1830
+ "validation_split": "val",
1831
+ "fewshot_split": "dev",
1832
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1833
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1834
+ "doc_to_choice": [
1835
+ "A",
1836
+ "B",
1837
+ "C",
1838
+ "D"
1839
+ ],
1840
+ "description": "以下是中国关于初中地理的单项选择题,请选出其中的正确答案。\n\n",
1841
+ "target_delimiter": " ",
1842
+ "fewshot_delimiter": "\n\n",
1843
+ "fewshot_config": {
1844
+ "sampler": "first_n"
1845
+ },
1846
+ "metric_list": [
1847
+ {
1848
+ "metric": "acc",
1849
+ "aggregation": "mean",
1850
+ "higher_is_better": true
1851
+ },
1852
+ {
1853
+ "metric": "acc_norm",
1854
+ "aggregation": "mean",
1855
+ "higher_is_better": true
1856
+ }
1857
+ ],
1858
+ "output_type": "multiple_choice",
1859
+ "repeats": 1,
1860
+ "should_decontaminate": false,
1861
+ "metadata": {
1862
+ "version": 1.0
1863
+ }
1864
+ },
1865
+ "ceval-valid_middle_school_history": {
1866
+ "task": "ceval-valid_middle_school_history",
1867
+ "group": "ceval-valid",
1868
+ "dataset_path": "ceval/ceval-exam",
1869
+ "dataset_name": "middle_school_history",
1870
+ "validation_split": "val",
1871
+ "fewshot_split": "dev",
1872
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1873
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1874
+ "doc_to_choice": [
1875
+ "A",
1876
+ "B",
1877
+ "C",
1878
+ "D"
1879
+ ],
1880
+ "description": "以下是中国关于初中历史的单项选择题,请选出其中的正确答案。\n\n",
1881
+ "target_delimiter": " ",
1882
+ "fewshot_delimiter": "\n\n",
1883
+ "fewshot_config": {
1884
+ "sampler": "first_n"
1885
+ },
1886
+ "metric_list": [
1887
+ {
1888
+ "metric": "acc",
1889
+ "aggregation": "mean",
1890
+ "higher_is_better": true
1891
+ },
1892
+ {
1893
+ "metric": "acc_norm",
1894
+ "aggregation": "mean",
1895
+ "higher_is_better": true
1896
+ }
1897
+ ],
1898
+ "output_type": "multiple_choice",
1899
+ "repeats": 1,
1900
+ "should_decontaminate": false,
1901
+ "metadata": {
1902
+ "version": 1.0
1903
+ }
1904
+ },
1905
+ "ceval-valid_middle_school_mathematics": {
1906
+ "task": "ceval-valid_middle_school_mathematics",
1907
+ "group": "ceval-valid",
1908
+ "dataset_path": "ceval/ceval-exam",
1909
+ "dataset_name": "middle_school_mathematics",
1910
+ "validation_split": "val",
1911
+ "fewshot_split": "dev",
1912
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1913
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1914
+ "doc_to_choice": [
1915
+ "A",
1916
+ "B",
1917
+ "C",
1918
+ "D"
1919
+ ],
1920
+ "description": "以下是中国关于初中数学的单项选择题,请选出其中的正确答案。\n\n",
1921
+ "target_delimiter": " ",
1922
+ "fewshot_delimiter": "\n\n",
1923
+ "fewshot_config": {
1924
+ "sampler": "first_n"
1925
+ },
1926
+ "metric_list": [
1927
+ {
1928
+ "metric": "acc",
1929
+ "aggregation": "mean",
1930
+ "higher_is_better": true
1931
+ },
1932
+ {
1933
+ "metric": "acc_norm",
1934
+ "aggregation": "mean",
1935
+ "higher_is_better": true
1936
+ }
1937
+ ],
1938
+ "output_type": "multiple_choice",
1939
+ "repeats": 1,
1940
+ "should_decontaminate": false,
1941
+ "metadata": {
1942
+ "version": 1.0
1943
+ }
1944
+ },
1945
+ "ceval-valid_middle_school_physics": {
1946
+ "task": "ceval-valid_middle_school_physics",
1947
+ "group": "ceval-valid",
1948
+ "dataset_path": "ceval/ceval-exam",
1949
+ "dataset_name": "middle_school_physics",
1950
+ "validation_split": "val",
1951
+ "fewshot_split": "dev",
1952
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1953
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1954
+ "doc_to_choice": [
1955
+ "A",
1956
+ "B",
1957
+ "C",
1958
+ "D"
1959
+ ],
1960
+ "description": "以下是中国关于初中物理的单项选择题,请选出其中的正确答案。\n\n",
1961
+ "target_delimiter": " ",
1962
+ "fewshot_delimiter": "\n\n",
1963
+ "fewshot_config": {
1964
+ "sampler": "first_n"
1965
+ },
1966
+ "metric_list": [
1967
+ {
1968
+ "metric": "acc",
1969
+ "aggregation": "mean",
1970
+ "higher_is_better": true
1971
+ },
1972
+ {
1973
+ "metric": "acc_norm",
1974
+ "aggregation": "mean",
1975
+ "higher_is_better": true
1976
+ }
1977
+ ],
1978
+ "output_type": "multiple_choice",
1979
+ "repeats": 1,
1980
+ "should_decontaminate": false,
1981
+ "metadata": {
1982
+ "version": 1.0
1983
+ }
1984
+ },
1985
+ "ceval-valid_middle_school_politics": {
1986
+ "task": "ceval-valid_middle_school_politics",
1987
+ "group": "ceval-valid",
1988
+ "dataset_path": "ceval/ceval-exam",
1989
+ "dataset_name": "middle_school_politics",
1990
+ "validation_split": "val",
1991
+ "fewshot_split": "dev",
1992
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
1993
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
1994
+ "doc_to_choice": [
1995
+ "A",
1996
+ "B",
1997
+ "C",
1998
+ "D"
1999
+ ],
2000
+ "description": "以下是中国关于初中政治的单项选择题,请选出其中的正确答案。\n\n",
2001
+ "target_delimiter": " ",
2002
+ "fewshot_delimiter": "\n\n",
2003
+ "fewshot_config": {
2004
+ "sampler": "first_n"
2005
+ },
2006
+ "metric_list": [
2007
+ {
2008
+ "metric": "acc",
2009
+ "aggregation": "mean",
2010
+ "higher_is_better": true
2011
+ },
2012
+ {
2013
+ "metric": "acc_norm",
2014
+ "aggregation": "mean",
2015
+ "higher_is_better": true
2016
+ }
2017
+ ],
2018
+ "output_type": "multiple_choice",
2019
+ "repeats": 1,
2020
+ "should_decontaminate": false,
2021
+ "metadata": {
2022
+ "version": 1.0
2023
+ }
2024
+ },
2025
+ "ceval-valid_modern_chinese_history": {
2026
+ "task": "ceval-valid_modern_chinese_history",
2027
+ "group": "ceval-valid",
2028
+ "dataset_path": "ceval/ceval-exam",
2029
+ "dataset_name": "modern_chinese_history",
2030
+ "validation_split": "val",
2031
+ "fewshot_split": "dev",
2032
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2033
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2034
+ "doc_to_choice": [
2035
+ "A",
2036
+ "B",
2037
+ "C",
2038
+ "D"
2039
+ ],
2040
+ "description": "以下是中国关于近代史纲要的单项选择题,请选出其中的正确答案。\n\n",
2041
+ "target_delimiter": " ",
2042
+ "fewshot_delimiter": "\n\n",
2043
+ "fewshot_config": {
2044
+ "sampler": "first_n"
2045
+ },
2046
+ "metric_list": [
2047
+ {
2048
+ "metric": "acc",
2049
+ "aggregation": "mean",
2050
+ "higher_is_better": true
2051
+ },
2052
+ {
2053
+ "metric": "acc_norm",
2054
+ "aggregation": "mean",
2055
+ "higher_is_better": true
2056
+ }
2057
+ ],
2058
+ "output_type": "multiple_choice",
2059
+ "repeats": 1,
2060
+ "should_decontaminate": false,
2061
+ "metadata": {
2062
+ "version": 1.0
2063
+ }
2064
+ },
2065
+ "ceval-valid_operating_system": {
2066
+ "task": "ceval-valid_operating_system",
2067
+ "group": "ceval-valid",
2068
+ "dataset_path": "ceval/ceval-exam",
2069
+ "dataset_name": "operating_system",
2070
+ "validation_split": "val",
2071
+ "fewshot_split": "dev",
2072
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2073
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2074
+ "doc_to_choice": [
2075
+ "A",
2076
+ "B",
2077
+ "C",
2078
+ "D"
2079
+ ],
2080
+ "description": "以下是中国关于操作系统的单项选择题,请选出其中的正确答案。\n\n",
2081
+ "target_delimiter": " ",
2082
+ "fewshot_delimiter": "\n\n",
2083
+ "fewshot_config": {
2084
+ "sampler": "first_n"
2085
+ },
2086
+ "metric_list": [
2087
+ {
2088
+ "metric": "acc",
2089
+ "aggregation": "mean",
2090
+ "higher_is_better": true
2091
+ },
2092
+ {
2093
+ "metric": "acc_norm",
2094
+ "aggregation": "mean",
2095
+ "higher_is_better": true
2096
+ }
2097
+ ],
2098
+ "output_type": "multiple_choice",
2099
+ "repeats": 1,
2100
+ "should_decontaminate": false,
2101
+ "metadata": {
2102
+ "version": 1.0
2103
+ }
2104
+ },
2105
+ "ceval-valid_physician": {
2106
+ "task": "ceval-valid_physician",
2107
+ "group": "ceval-valid",
2108
+ "dataset_path": "ceval/ceval-exam",
2109
+ "dataset_name": "physician",
2110
+ "validation_split": "val",
2111
+ "fewshot_split": "dev",
2112
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2113
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2114
+ "doc_to_choice": [
2115
+ "A",
2116
+ "B",
2117
+ "C",
2118
+ "D"
2119
+ ],
2120
+ "description": "以下是中国关于医师资格的单项选择题,请选出其中的正确答案。\n\n",
2121
+ "target_delimiter": " ",
2122
+ "fewshot_delimiter": "\n\n",
2123
+ "fewshot_config": {
2124
+ "sampler": "first_n"
2125
+ },
2126
+ "metric_list": [
2127
+ {
2128
+ "metric": "acc",
2129
+ "aggregation": "mean",
2130
+ "higher_is_better": true
2131
+ },
2132
+ {
2133
+ "metric": "acc_norm",
2134
+ "aggregation": "mean",
2135
+ "higher_is_better": true
2136
+ }
2137
+ ],
2138
+ "output_type": "multiple_choice",
2139
+ "repeats": 1,
2140
+ "should_decontaminate": false,
2141
+ "metadata": {
2142
+ "version": 1.0
2143
+ }
2144
+ },
2145
+ "ceval-valid_plant_protection": {
2146
+ "task": "ceval-valid_plant_protection",
2147
+ "group": "ceval-valid",
2148
+ "dataset_path": "ceval/ceval-exam",
2149
+ "dataset_name": "plant_protection",
2150
+ "validation_split": "val",
2151
+ "fewshot_split": "dev",
2152
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2153
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2154
+ "doc_to_choice": [
2155
+ "A",
2156
+ "B",
2157
+ "C",
2158
+ "D"
2159
+ ],
2160
+ "description": "以下是中国关于植物保护的单项选择题,请选出其中的正确答案。\n\n",
2161
+ "target_delimiter": " ",
2162
+ "fewshot_delimiter": "\n\n",
2163
+ "fewshot_config": {
2164
+ "sampler": "first_n"
2165
+ },
2166
+ "metric_list": [
2167
+ {
2168
+ "metric": "acc",
2169
+ "aggregation": "mean",
2170
+ "higher_is_better": true
2171
+ },
2172
+ {
2173
+ "metric": "acc_norm",
2174
+ "aggregation": "mean",
2175
+ "higher_is_better": true
2176
+ }
2177
+ ],
2178
+ "output_type": "multiple_choice",
2179
+ "repeats": 1,
2180
+ "should_decontaminate": false,
2181
+ "metadata": {
2182
+ "version": 1.0
2183
+ }
2184
+ },
2185
+ "ceval-valid_probability_and_statistics": {
2186
+ "task": "ceval-valid_probability_and_statistics",
2187
+ "group": "ceval-valid",
2188
+ "dataset_path": "ceval/ceval-exam",
2189
+ "dataset_name": "probability_and_statistics",
2190
+ "validation_split": "val",
2191
+ "fewshot_split": "dev",
2192
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2193
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2194
+ "doc_to_choice": [
2195
+ "A",
2196
+ "B",
2197
+ "C",
2198
+ "D"
2199
+ ],
2200
+ "description": "以下是中国关于概率统计的单项选择题,请选出其中的正确答案。\n\n",
2201
+ "target_delimiter": " ",
2202
+ "fewshot_delimiter": "\n\n",
2203
+ "fewshot_config": {
2204
+ "sampler": "first_n"
2205
+ },
2206
+ "metric_list": [
2207
+ {
2208
+ "metric": "acc",
2209
+ "aggregation": "mean",
2210
+ "higher_is_better": true
2211
+ },
2212
+ {
2213
+ "metric": "acc_norm",
2214
+ "aggregation": "mean",
2215
+ "higher_is_better": true
2216
+ }
2217
+ ],
2218
+ "output_type": "multiple_choice",
2219
+ "repeats": 1,
2220
+ "should_decontaminate": false,
2221
+ "metadata": {
2222
+ "version": 1.0
2223
+ }
2224
+ },
2225
+ "ceval-valid_professional_tour_guide": {
2226
+ "task": "ceval-valid_professional_tour_guide",
2227
+ "group": "ceval-valid",
2228
+ "dataset_path": "ceval/ceval-exam",
2229
+ "dataset_name": "professional_tour_guide",
2230
+ "validation_split": "val",
2231
+ "fewshot_split": "dev",
2232
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2233
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2234
+ "doc_to_choice": [
2235
+ "A",
2236
+ "B",
2237
+ "C",
2238
+ "D"
2239
+ ],
2240
+ "description": "以下是中国关于导游资格的单项选择题,请选出其中的正确答案。\n\n",
2241
+ "target_delimiter": " ",
2242
+ "fewshot_delimiter": "\n\n",
2243
+ "fewshot_config": {
2244
+ "sampler": "first_n"
2245
+ },
2246
+ "metric_list": [
2247
+ {
2248
+ "metric": "acc",
2249
+ "aggregation": "mean",
2250
+ "higher_is_better": true
2251
+ },
2252
+ {
2253
+ "metric": "acc_norm",
2254
+ "aggregation": "mean",
2255
+ "higher_is_better": true
2256
+ }
2257
+ ],
2258
+ "output_type": "multiple_choice",
2259
+ "repeats": 1,
2260
+ "should_decontaminate": false,
2261
+ "metadata": {
2262
+ "version": 1.0
2263
+ }
2264
+ },
2265
+ "ceval-valid_sports_science": {
2266
+ "task": "ceval-valid_sports_science",
2267
+ "group": "ceval-valid",
2268
+ "dataset_path": "ceval/ceval-exam",
2269
+ "dataset_name": "sports_science",
2270
+ "validation_split": "val",
2271
+ "fewshot_split": "dev",
2272
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2273
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2274
+ "doc_to_choice": [
2275
+ "A",
2276
+ "B",
2277
+ "C",
2278
+ "D"
2279
+ ],
2280
+ "description": "以下是中国关于体育学的单项选择题,请选出其中的正确答案。\n\n",
2281
+ "target_delimiter": " ",
2282
+ "fewshot_delimiter": "\n\n",
2283
+ "fewshot_config": {
2284
+ "sampler": "first_n"
2285
+ },
2286
+ "metric_list": [
2287
+ {
2288
+ "metric": "acc",
2289
+ "aggregation": "mean",
2290
+ "higher_is_better": true
2291
+ },
2292
+ {
2293
+ "metric": "acc_norm",
2294
+ "aggregation": "mean",
2295
+ "higher_is_better": true
2296
+ }
2297
+ ],
2298
+ "output_type": "multiple_choice",
2299
+ "repeats": 1,
2300
+ "should_decontaminate": false,
2301
+ "metadata": {
2302
+ "version": 1.0
2303
+ }
2304
+ },
2305
+ "ceval-valid_tax_accountant": {
2306
+ "task": "ceval-valid_tax_accountant",
2307
+ "group": "ceval-valid",
2308
+ "dataset_path": "ceval/ceval-exam",
2309
+ "dataset_name": "tax_accountant",
2310
+ "validation_split": "val",
2311
+ "fewshot_split": "dev",
2312
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2313
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2314
+ "doc_to_choice": [
2315
+ "A",
2316
+ "B",
2317
+ "C",
2318
+ "D"
2319
+ ],
2320
+ "description": "以下是中国关于税务师的单项选择题,请选出其中的正确答案。\n\n",
2321
+ "target_delimiter": " ",
2322
+ "fewshot_delimiter": "\n\n",
2323
+ "fewshot_config": {
2324
+ "sampler": "first_n"
2325
+ },
2326
+ "metric_list": [
2327
+ {
2328
+ "metric": "acc",
2329
+ "aggregation": "mean",
2330
+ "higher_is_better": true
2331
+ },
2332
+ {
2333
+ "metric": "acc_norm",
2334
+ "aggregation": "mean",
2335
+ "higher_is_better": true
2336
+ }
2337
+ ],
2338
+ "output_type": "multiple_choice",
2339
+ "repeats": 1,
2340
+ "should_decontaminate": false,
2341
+ "metadata": {
2342
+ "version": 1.0
2343
+ }
2344
+ },
2345
+ "ceval-valid_teacher_qualification": {
2346
+ "task": "ceval-valid_teacher_qualification",
2347
+ "group": "ceval-valid",
2348
+ "dataset_path": "ceval/ceval-exam",
2349
+ "dataset_name": "teacher_qualification",
2350
+ "validation_split": "val",
2351
+ "fewshot_split": "dev",
2352
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2353
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2354
+ "doc_to_choice": [
2355
+ "A",
2356
+ "B",
2357
+ "C",
2358
+ "D"
2359
+ ],
2360
+ "description": "以下是中国关于教师资格的单项选择题,请选出其中的正确答案。\n\n",
2361
+ "target_delimiter": " ",
2362
+ "fewshot_delimiter": "\n\n",
2363
+ "fewshot_config": {
2364
+ "sampler": "first_n"
2365
+ },
2366
+ "metric_list": [
2367
+ {
2368
+ "metric": "acc",
2369
+ "aggregation": "mean",
2370
+ "higher_is_better": true
2371
+ },
2372
+ {
2373
+ "metric": "acc_norm",
2374
+ "aggregation": "mean",
2375
+ "higher_is_better": true
2376
+ }
2377
+ ],
2378
+ "output_type": "multiple_choice",
2379
+ "repeats": 1,
2380
+ "should_decontaminate": false,
2381
+ "metadata": {
2382
+ "version": 1.0
2383
+ }
2384
+ },
2385
+ "ceval-valid_urban_and_rural_planner": {
2386
+ "task": "ceval-valid_urban_and_rural_planner",
2387
+ "group": "ceval-valid",
2388
+ "dataset_path": "ceval/ceval-exam",
2389
+ "dataset_name": "urban_and_rural_planner",
2390
+ "validation_split": "val",
2391
+ "fewshot_split": "dev",
2392
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2393
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2394
+ "doc_to_choice": [
2395
+ "A",
2396
+ "B",
2397
+ "C",
2398
+ "D"
2399
+ ],
2400
+ "description": "以下是中国关于注册城乡规划师的单项选择题,请选出其中的正确答案。\n\n",
2401
+ "target_delimiter": " ",
2402
+ "fewshot_delimiter": "\n\n",
2403
+ "fewshot_config": {
2404
+ "sampler": "first_n"
2405
+ },
2406
+ "metric_list": [
2407
+ {
2408
+ "metric": "acc",
2409
+ "aggregation": "mean",
2410
+ "higher_is_better": true
2411
+ },
2412
+ {
2413
+ "metric": "acc_norm",
2414
+ "aggregation": "mean",
2415
+ "higher_is_better": true
2416
+ }
2417
+ ],
2418
+ "output_type": "multiple_choice",
2419
+ "repeats": 1,
2420
+ "should_decontaminate": false,
2421
+ "metadata": {
2422
+ "version": 1.0
2423
+ }
2424
+ },
2425
+ "ceval-valid_veterinary_medicine": {
2426
+ "task": "ceval-valid_veterinary_medicine",
2427
+ "group": "ceval-valid",
2428
+ "dataset_path": "ceval/ceval-exam",
2429
+ "dataset_name": "veterinary_medicine",
2430
+ "validation_split": "val",
2431
+ "fewshot_split": "dev",
2432
+ "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:",
2433
+ "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}",
2434
+ "doc_to_choice": [
2435
+ "A",
2436
+ "B",
2437
+ "C",
2438
+ "D"
2439
+ ],
2440
+ "description": "以下是中国关于兽医学的单项选择题,请选出其中的正确答案。\n\n",
2441
+ "target_delimiter": " ",
2442
+ "fewshot_delimiter": "\n\n",
2443
+ "fewshot_config": {
2444
+ "sampler": "first_n"
2445
+ },
2446
+ "metric_list": [
2447
+ {
2448
+ "metric": "acc",
2449
+ "aggregation": "mean",
2450
+ "higher_is_better": true
2451
+ },
2452
+ {
2453
+ "metric": "acc_norm",
2454
+ "aggregation": "mean",
2455
+ "higher_is_better": true
2456
+ }
2457
+ ],
2458
+ "output_type": "multiple_choice",
2459
+ "repeats": 1,
2460
+ "should_decontaminate": false,
2461
+ "metadata": {
2462
+ "version": 1.0
2463
+ }
2464
+ }
2465
+ },
2466
+ "versions": {
2467
+ "ceval-valid": "N/A",
2468
+ "ceval-valid_accountant": 1.0,
2469
+ "ceval-valid_advanced_mathematics": 1.0,
2470
+ "ceval-valid_art_studies": 1.0,
2471
+ "ceval-valid_basic_medicine": 1.0,
2472
+ "ceval-valid_business_administration": 1.0,
2473
+ "ceval-valid_chinese_language_and_literature": 1.0,
2474
+ "ceval-valid_civil_servant": 1.0,
2475
+ "ceval-valid_clinical_medicine": 1.0,
2476
+ "ceval-valid_college_chemistry": 1.0,
2477
+ "ceval-valid_college_economics": 1.0,
2478
+ "ceval-valid_college_physics": 1.0,
2479
+ "ceval-valid_college_programming": 1.0,
2480
+ "ceval-valid_computer_architecture": 1.0,
2481
+ "ceval-valid_computer_network": 1.0,
2482
+ "ceval-valid_discrete_mathematics": 1.0,
2483
+ "ceval-valid_education_science": 1.0,
2484
+ "ceval-valid_electrical_engineer": 1.0,
2485
+ "ceval-valid_environmental_impact_assessment_engineer": 1.0,
2486
+ "ceval-valid_fire_engineer": 1.0,
2487
+ "ceval-valid_high_school_biology": 1.0,
2488
+ "ceval-valid_high_school_chemistry": 1.0,
2489
+ "ceval-valid_high_school_chinese": 1.0,
2490
+ "ceval-valid_high_school_geography": 1.0,
2491
+ "ceval-valid_high_school_history": 1.0,
2492
+ "ceval-valid_high_school_mathematics": 1.0,
2493
+ "ceval-valid_high_school_physics": 1.0,
2494
+ "ceval-valid_high_school_politics": 1.0,
2495
+ "ceval-valid_ideological_and_moral_cultivation": 1.0,
2496
+ "ceval-valid_law": 1.0,
2497
+ "ceval-valid_legal_professional": 1.0,
2498
+ "ceval-valid_logic": 1.0,
2499
+ "ceval-valid_mao_zedong_thought": 1.0,
2500
+ "ceval-valid_marxism": 1.0,
2501
+ "ceval-valid_metrology_engineer": 1.0,
2502
+ "ceval-valid_middle_school_biology": 1.0,
2503
+ "ceval-valid_middle_school_chemistry": 1.0,
2504
+ "ceval-valid_middle_school_geography": 1.0,
2505
+ "ceval-valid_middle_school_history": 1.0,
2506
+ "ceval-valid_middle_school_mathematics": 1.0,
2507
+ "ceval-valid_middle_school_physics": 1.0,
2508
+ "ceval-valid_middle_school_politics": 1.0,
2509
+ "ceval-valid_modern_chinese_history": 1.0,
2510
+ "ceval-valid_operating_system": 1.0,
2511
+ "ceval-valid_physician": 1.0,
2512
+ "ceval-valid_plant_protection": 1.0,
2513
+ "ceval-valid_probability_and_statistics": 1.0,
2514
+ "ceval-valid_professional_tour_guide": 1.0,
2515
+ "ceval-valid_sports_science": 1.0,
2516
+ "ceval-valid_tax_accountant": 1.0,
2517
+ "ceval-valid_teacher_qualification": 1.0,
2518
+ "ceval-valid_urban_and_rural_planner": 1.0,
2519
+ "ceval-valid_veterinary_medicine": 1.0
2520
+ },
2521
+ "n-shot": {
2522
+ "ceval-valid": 0,
2523
+ "ceval-valid_accountant": 0,
2524
+ "ceval-valid_advanced_mathematics": 0,
2525
+ "ceval-valid_art_studies": 0,
2526
+ "ceval-valid_basic_medicine": 0,
2527
+ "ceval-valid_business_administration": 0,
2528
+ "ceval-valid_chinese_language_and_literature": 0,
2529
+ "ceval-valid_civil_servant": 0,
2530
+ "ceval-valid_clinical_medicine": 0,
2531
+ "ceval-valid_college_chemistry": 0,
2532
+ "ceval-valid_college_economics": 0,
2533
+ "ceval-valid_college_physics": 0,
2534
+ "ceval-valid_college_programming": 0,
2535
+ "ceval-valid_computer_architecture": 0,
2536
+ "ceval-valid_computer_network": 0,
2537
+ "ceval-valid_discrete_mathematics": 0,
2538
+ "ceval-valid_education_science": 0,
2539
+ "ceval-valid_electrical_engineer": 0,
2540
+ "ceval-valid_environmental_impact_assessment_engineer": 0,
2541
+ "ceval-valid_fire_engineer": 0,
2542
+ "ceval-valid_high_school_biology": 0,
2543
+ "ceval-valid_high_school_chemistry": 0,
2544
+ "ceval-valid_high_school_chinese": 0,
2545
+ "ceval-valid_high_school_geography": 0,
2546
+ "ceval-valid_high_school_history": 0,
2547
+ "ceval-valid_high_school_mathematics": 0,
2548
+ "ceval-valid_high_school_physics": 0,
2549
+ "ceval-valid_high_school_politics": 0,
2550
+ "ceval-valid_ideological_and_moral_cultivation": 0,
2551
+ "ceval-valid_law": 0,
2552
+ "ceval-valid_legal_professional": 0,
2553
+ "ceval-valid_logic": 0,
2554
+ "ceval-valid_mao_zedong_thought": 0,
2555
+ "ceval-valid_marxism": 0,
2556
+ "ceval-valid_metrology_engineer": 0,
2557
+ "ceval-valid_middle_school_biology": 0,
2558
+ "ceval-valid_middle_school_chemistry": 0,
2559
+ "ceval-valid_middle_school_geography": 0,
2560
+ "ceval-valid_middle_school_history": 0,
2561
+ "ceval-valid_middle_school_mathematics": 0,
2562
+ "ceval-valid_middle_school_physics": 0,
2563
+ "ceval-valid_middle_school_politics": 0,
2564
+ "ceval-valid_modern_chinese_history": 0,
2565
+ "ceval-valid_operating_system": 0,
2566
+ "ceval-valid_physician": 0,
2567
+ "ceval-valid_plant_protection": 0,
2568
+ "ceval-valid_probability_and_statistics": 0,
2569
+ "ceval-valid_professional_tour_guide": 0,
2570
+ "ceval-valid_sports_science": 0,
2571
+ "ceval-valid_tax_accountant": 0,
2572
+ "ceval-valid_teacher_qualification": 0,
2573
+ "ceval-valid_urban_and_rural_planner": 0,
2574
+ "ceval-valid_veterinary_medicine": 0
2575
+ },
2576
+ "config": {
2577
+ "model": "hf",
2578
+ "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
2579
+ "batch_size": "auto",
2580
+ "batch_sizes": [
2581
+ 32
2582
+ ],
2583
+ "device": null,
2584
+ "use_cache": null,
2585
+ "limit": null,
2586
+ "bootstrap_iters": 100000,
2587
+ "gen_kwargs": null
2588
+ },
2589
+ "git_hash": "4d19ea9"
2590
+ }
lm-eval-output/allenai/OLMo-7B/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f000647be6eea5529f603c75503e3332301c0bd4aa7f76edcbbb38001ab4cc9c
3
- size 79965
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19d198e1ad344b1d97e9c92189aaebee5c56c2a039a9f1565485bd6dc1700f9e
3
+ size 29064
lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
  "results": {
3
  "cmmlu": {
4
- "acc,none": 0.2485753755828009,
5
- "acc_stderr,none": 0.042815306797706565,
6
- "acc_norm,none": 0.2485753755828009,
7
- "acc_norm_stderr,none": 0.042815306797706565,
8
  "alias": "cmmlu"
9
  },
10
  "cmmlu_agronomy": {
@@ -15,10 +15,10 @@
15
  "alias": " - cmmlu_agronomy"
16
  },
17
  "cmmlu_anatomy": {
18
- "acc,none": 0.22297297297297297,
19
- "acc_stderr,none": 0.03433092518104002,
20
- "acc_norm,none": 0.22297297297297297,
21
- "acc_norm_stderr,none": 0.03433092518104002,
22
  "alias": " - cmmlu_anatomy"
23
  },
24
  "cmmlu_ancient_chinese": {
@@ -29,45 +29,45 @@
29
  "alias": " - cmmlu_ancient_chinese"
30
  },
31
  "cmmlu_arts": {
32
- "acc,none": 0.29375,
33
- "acc_stderr,none": 0.03612181848191273,
34
- "acc_norm,none": 0.29375,
35
- "acc_norm_stderr,none": 0.03612181848191273,
36
  "alias": " - cmmlu_arts"
37
  },
38
  "cmmlu_astronomy": {
39
- "acc,none": 0.20606060606060606,
40
- "acc_stderr,none": 0.031584153240477086,
41
- "acc_norm,none": 0.20606060606060606,
42
- "acc_norm_stderr,none": 0.031584153240477086,
43
  "alias": " - cmmlu_astronomy"
44
  },
45
  "cmmlu_business_ethics": {
46
- "acc,none": 0.23444976076555024,
47
- "acc_stderr,none": 0.029375148972005737,
48
- "acc_norm,none": 0.23444976076555024,
49
- "acc_norm_stderr,none": 0.029375148972005737,
50
  "alias": " - cmmlu_business_ethics"
51
  },
52
  "cmmlu_chinese_civil_service_exam": {
53
- "acc,none": 0.21875,
54
- "acc_stderr,none": 0.032784644885244255,
55
- "acc_norm,none": 0.21875,
56
- "acc_norm_stderr,none": 0.032784644885244255,
57
  "alias": " - cmmlu_chinese_civil_service_exam"
58
  },
59
  "cmmlu_chinese_driving_rule": {
60
- "acc,none": 0.25190839694656486,
61
- "acc_stderr,none": 0.03807387116306086,
62
- "acc_norm,none": 0.25190839694656486,
63
- "acc_norm_stderr,none": 0.03807387116306086,
64
  "alias": " - cmmlu_chinese_driving_rule"
65
  },
66
  "cmmlu_chinese_food_culture": {
67
  "acc,none": 0.20588235294117646,
68
- "acc_stderr,none": 0.034800469312350674,
69
  "acc_norm,none": 0.20588235294117646,
70
- "acc_norm_stderr,none": 0.034800469312350674,
71
  "alias": " - cmmlu_chinese_food_culture"
72
  },
73
  "cmmlu_chinese_foreign_policy": {
@@ -78,73 +78,73 @@
78
  "alias": " - cmmlu_chinese_foreign_policy"
79
  },
80
  "cmmlu_chinese_history": {
81
- "acc,none": 0.29721362229102166,
82
- "acc_stderr,none": 0.025469363219004768,
83
- "acc_norm,none": 0.29721362229102166,
84
- "acc_norm_stderr,none": 0.025469363219004768,
85
  "alias": " - cmmlu_chinese_history"
86
  },
87
  "cmmlu_chinese_literature": {
88
  "acc,none": 0.22058823529411764,
89
- "acc_stderr,none": 0.02910225438967408,
90
  "acc_norm,none": 0.22058823529411764,
91
- "acc_norm_stderr,none": 0.02910225438967408,
92
  "alias": " - cmmlu_chinese_literature"
93
  },
94
  "cmmlu_chinese_teacher_qualification": {
95
- "acc,none": 0.22346368715083798,
96
- "acc_stderr,none": 0.031222980919579764,
97
- "acc_norm,none": 0.22346368715083798,
98
- "acc_norm_stderr,none": 0.031222980919579764,
99
  "alias": " - cmmlu_chinese_teacher_qualification"
100
  },
101
  "cmmlu_clinical_knowledge": {
102
- "acc,none": 0.22784810126582278,
103
- "acc_stderr,none": 0.027303484599069443,
104
- "acc_norm,none": 0.22784810126582278,
105
- "acc_norm_stderr,none": 0.027303484599069443,
106
  "alias": " - cmmlu_clinical_knowledge"
107
  },
108
  "cmmlu_college_actuarial_science": {
109
- "acc,none": 0.2830188679245283,
110
- "acc_stderr,none": 0.04396093377439375,
111
- "acc_norm,none": 0.2830188679245283,
112
- "acc_norm_stderr,none": 0.04396093377439375,
113
  "alias": " - cmmlu_college_actuarial_science"
114
  },
115
  "cmmlu_college_education": {
116
- "acc,none": 0.2897196261682243,
117
- "acc_stderr,none": 0.0440606533474851,
118
- "acc_norm,none": 0.2897196261682243,
119
- "acc_norm_stderr,none": 0.0440606533474851,
120
  "alias": " - cmmlu_college_education"
121
  },
122
  "cmmlu_college_engineering_hydrology": {
123
  "acc,none": 0.2641509433962264,
124
- "acc_stderr,none": 0.043025487739590106,
125
  "acc_norm,none": 0.2641509433962264,
126
- "acc_norm_stderr,none": 0.043025487739590106,
127
  "alias": " - cmmlu_college_engineering_hydrology"
128
  },
129
  "cmmlu_college_law": {
130
  "acc,none": 0.2962962962962963,
131
- "acc_stderr,none": 0.044143436668549335,
132
  "acc_norm,none": 0.2962962962962963,
133
- "acc_norm_stderr,none": 0.044143436668549335,
134
  "alias": " - cmmlu_college_law"
135
  },
136
  "cmmlu_college_mathematics": {
137
- "acc,none": 0.3047619047619048,
138
- "acc_stderr,none": 0.04513676718168311,
139
- "acc_norm,none": 0.3047619047619048,
140
- "acc_norm_stderr,none": 0.04513676718168311,
141
  "alias": " - cmmlu_college_mathematics"
142
  },
143
  "cmmlu_college_medical_statistics": {
144
- "acc,none": 0.19811320754716982,
145
- "acc_stderr,none": 0.0388972228831855,
146
- "acc_norm,none": 0.19811320754716982,
147
- "acc_norm_stderr,none": 0.0388972228831855,
148
  "alias": " - cmmlu_college_medical_statistics"
149
  },
150
  "cmmlu_college_medicine": {
@@ -155,24 +155,24 @@
155
  "alias": " - cmmlu_college_medicine"
156
  },
157
  "cmmlu_computer_science": {
158
- "acc,none": 0.24509803921568626,
159
- "acc_stderr,none": 0.030190282453501964,
160
- "acc_norm,none": 0.24509803921568626,
161
- "acc_norm_stderr,none": 0.030190282453501964,
162
  "alias": " - cmmlu_computer_science"
163
  },
164
  "cmmlu_computer_security": {
165
- "acc,none": 0.26900584795321636,
166
- "acc_stderr,none": 0.03401052620104088,
167
- "acc_norm,none": 0.26900584795321636,
168
- "acc_norm_stderr,none": 0.03401052620104088,
169
  "alias": " - cmmlu_computer_security"
170
  },
171
  "cmmlu_conceptual_physics": {
172
- "acc,none": 0.23129251700680273,
173
- "acc_stderr,none": 0.034896744812616155,
174
- "acc_norm,none": 0.23129251700680273,
175
- "acc_norm_stderr,none": 0.034896744812616155,
176
  "alias": " - cmmlu_conceptual_physics"
177
  },
178
  "cmmlu_construction_project_management": {
@@ -184,156 +184,156 @@
184
  },
185
  "cmmlu_economics": {
186
  "acc,none": 0.27672955974842767,
187
- "acc_stderr,none": 0.03559177035707934,
188
  "acc_norm,none": 0.27672955974842767,
189
- "acc_norm_stderr,none": 0.03559177035707934,
190
  "alias": " - cmmlu_economics"
191
  },
192
  "cmmlu_education": {
193
- "acc,none": 0.27607361963190186,
194
- "acc_stderr,none": 0.0351238528370505,
195
- "acc_norm,none": 0.27607361963190186,
196
- "acc_norm_stderr,none": 0.0351238528370505,
197
  "alias": " - cmmlu_education"
198
  },
199
  "cmmlu_electrical_engineering": {
200
- "acc,none": 0.27325581395348836,
201
- "acc_stderr,none": 0.03407826167337437,
202
- "acc_norm,none": 0.27325581395348836,
203
- "acc_norm_stderr,none": 0.03407826167337437,
204
  "alias": " - cmmlu_electrical_engineering"
205
  },
206
  "cmmlu_elementary_chinese": {
207
- "acc,none": 0.23412698412698413,
208
- "acc_stderr,none": 0.0267280489993024,
209
- "acc_norm,none": 0.23412698412698413,
210
- "acc_norm_stderr,none": 0.0267280489993024,
211
  "alias": " - cmmlu_elementary_chinese"
212
  },
213
  "cmmlu_elementary_commonsense": {
214
  "acc,none": 0.23737373737373738,
215
- "acc_stderr,none": 0.030313710538198896,
216
  "acc_norm,none": 0.23737373737373738,
217
- "acc_norm_stderr,none": 0.030313710538198896,
218
  "alias": " - cmmlu_elementary_commonsense"
219
  },
220
  "cmmlu_elementary_information_and_technology": {
221
- "acc,none": 0.24369747899159663,
222
- "acc_stderr,none": 0.02788682807838058,
223
- "acc_norm,none": 0.24369747899159663,
224
- "acc_norm_stderr,none": 0.02788682807838058,
225
  "alias": " - cmmlu_elementary_information_and_technology"
226
  },
227
  "cmmlu_elementary_mathematics": {
228
- "acc,none": 0.21304347826086956,
229
- "acc_stderr,none": 0.027057754389936194,
230
- "acc_norm,none": 0.21304347826086956,
231
- "acc_norm_stderr,none": 0.027057754389936194,
232
  "alias": " - cmmlu_elementary_mathematics"
233
  },
234
  "cmmlu_ethnology": {
235
- "acc,none": 0.2518518518518518,
236
- "acc_stderr,none": 0.037498507091740234,
237
- "acc_norm,none": 0.2518518518518518,
238
- "acc_norm_stderr,none": 0.037498507091740234,
239
  "alias": " - cmmlu_ethnology"
240
  },
241
  "cmmlu_food_science": {
242
- "acc,none": 0.2867132867132867,
243
- "acc_stderr,none": 0.03795000212801782,
244
- "acc_norm,none": 0.2867132867132867,
245
- "acc_norm_stderr,none": 0.03795000212801782,
246
  "alias": " - cmmlu_food_science"
247
  },
248
  "cmmlu_genetics": {
249
  "acc,none": 0.2897727272727273,
250
- "acc_stderr,none": 0.034293230802398766,
251
  "acc_norm,none": 0.2897727272727273,
252
- "acc_norm_stderr,none": 0.034293230802398766,
253
  "alias": " - cmmlu_genetics"
254
  },
255
  "cmmlu_global_facts": {
256
- "acc,none": 0.2483221476510067,
257
- "acc_stderr,none": 0.0355134404169743,
258
- "acc_norm,none": 0.2483221476510067,
259
- "acc_norm_stderr,none": 0.0355134404169743,
260
  "alias": " - cmmlu_global_facts"
261
  },
262
  "cmmlu_high_school_biology": {
263
- "acc,none": 0.23076923076923078,
264
- "acc_stderr,none": 0.03250593287417369,
265
- "acc_norm,none": 0.23076923076923078,
266
- "acc_norm_stderr,none": 0.03250593287417369,
267
  "alias": " - cmmlu_high_school_biology"
268
  },
269
  "cmmlu_high_school_chemistry": {
270
- "acc,none": 0.22727272727272727,
271
- "acc_stderr,none": 0.03661433360410719,
272
- "acc_norm,none": 0.22727272727272727,
273
- "acc_norm_stderr,none": 0.03661433360410719,
274
  "alias": " - cmmlu_high_school_chemistry"
275
  },
276
  "cmmlu_high_school_geography": {
277
- "acc,none": 0.2457627118644068,
278
- "acc_stderr,none": 0.03980329854920432,
279
- "acc_norm,none": 0.2457627118644068,
280
- "acc_norm_stderr,none": 0.03980329854920432,
281
  "alias": " - cmmlu_high_school_geography"
282
  },
283
  "cmmlu_high_school_mathematics": {
284
  "acc,none": 0.2621951219512195,
285
- "acc_stderr,none": 0.0344500028917346,
286
  "acc_norm,none": 0.2621951219512195,
287
- "acc_norm_stderr,none": 0.0344500028917346,
288
  "alias": " - cmmlu_high_school_mathematics"
289
  },
290
  "cmmlu_high_school_physics": {
291
- "acc,none": 0.2818181818181818,
292
- "acc_stderr,none": 0.04309118709946458,
293
- "acc_norm,none": 0.2818181818181818,
294
- "acc_norm_stderr,none": 0.04309118709946458,
295
  "alias": " - cmmlu_high_school_physics"
296
  },
297
  "cmmlu_high_school_politics": {
298
- "acc,none": 0.1888111888111888,
299
- "acc_stderr,none": 0.03284208093616429,
300
- "acc_norm,none": 0.1888111888111888,
301
- "acc_norm_stderr,none": 0.03284208093616429,
302
  "alias": " - cmmlu_high_school_politics"
303
  },
304
  "cmmlu_human_sexuality": {
305
- "acc,none": 0.19047619047619047,
306
- "acc_stderr,none": 0.035122074123020534,
307
- "acc_norm,none": 0.19047619047619047,
308
- "acc_norm_stderr,none": 0.035122074123020534,
309
  "alias": " - cmmlu_human_sexuality"
310
  },
311
  "cmmlu_international_law": {
312
- "acc,none": 0.25405405405405407,
313
- "acc_stderr,none": 0.03209281645145386,
314
- "acc_norm,none": 0.25405405405405407,
315
- "acc_norm_stderr,none": 0.03209281645145386,
316
  "alias": " - cmmlu_international_law"
317
  },
318
  "cmmlu_journalism": {
319
  "acc,none": 0.22674418604651161,
320
- "acc_stderr,none": 0.032020758995849365,
321
  "acc_norm,none": 0.22674418604651161,
322
- "acc_norm_stderr,none": 0.032020758995849365,
323
  "alias": " - cmmlu_journalism"
324
  },
325
  "cmmlu_jurisprudence": {
326
- "acc,none": 0.26763990267639903,
327
- "acc_stderr,none": 0.021864816663672668,
328
- "acc_norm,none": 0.26763990267639903,
329
- "acc_norm_stderr,none": 0.021864816663672668,
330
  "alias": " - cmmlu_jurisprudence"
331
  },
332
  "cmmlu_legal_and_moral_basis": {
333
- "acc,none": 0.3037383177570093,
334
- "acc_stderr,none": 0.03150984286811783,
335
- "acc_norm,none": 0.3037383177570093,
336
- "acc_norm_stderr,none": 0.03150984286811783,
337
  "alias": " - cmmlu_legal_and_moral_basis"
338
  },
339
  "cmmlu_logical": {
@@ -344,17 +344,17 @@
344
  "alias": " - cmmlu_logical"
345
  },
346
  "cmmlu_machine_learning": {
347
- "acc,none": 0.28688524590163933,
348
- "acc_stderr,none": 0.041118866352671826,
349
- "acc_norm,none": 0.28688524590163933,
350
- "acc_norm_stderr,none": 0.041118866352671826,
351
  "alias": " - cmmlu_machine_learning"
352
  },
353
  "cmmlu_management": {
354
- "acc,none": 0.19523809523809524,
355
- "acc_stderr,none": 0.027418446398346896,
356
- "acc_norm,none": 0.19523809523809524,
357
- "acc_norm_stderr,none": 0.027418446398346896,
358
  "alias": " - cmmlu_management"
359
  },
360
  "cmmlu_marketing": {
@@ -365,52 +365,52 @@
365
  "alias": " - cmmlu_marketing"
366
  },
367
  "cmmlu_marxist_theory": {
368
- "acc,none": 0.24338624338624337,
369
- "acc_stderr,none": 0.031297251928558506,
370
- "acc_norm,none": 0.24338624338624337,
371
- "acc_norm_stderr,none": 0.031297251928558506,
372
  "alias": " - cmmlu_marxist_theory"
373
  },
374
  "cmmlu_modern_chinese": {
375
- "acc,none": 0.23275862068965517,
376
- "acc_stderr,none": 0.039406691683376995,
377
- "acc_norm,none": 0.23275862068965517,
378
- "acc_norm_stderr,none": 0.039406691683376995,
379
  "alias": " - cmmlu_modern_chinese"
380
  },
381
  "cmmlu_nutrition": {
382
- "acc,none": 0.2689655172413793,
383
- "acc_stderr,none": 0.036951833116502325,
384
- "acc_norm,none": 0.2689655172413793,
385
- "acc_norm_stderr,none": 0.036951833116502325,
386
  "alias": " - cmmlu_nutrition"
387
  },
388
  "cmmlu_philosophy": {
389
- "acc,none": 0.3047619047619048,
390
- "acc_stderr,none": 0.0451367671816831,
391
- "acc_norm,none": 0.3047619047619048,
392
- "acc_norm_stderr,none": 0.0451367671816831,
393
  "alias": " - cmmlu_philosophy"
394
  },
395
  "cmmlu_professional_accounting": {
396
- "acc,none": 0.2342857142857143,
397
- "acc_stderr,none": 0.032109360396926204,
398
- "acc_norm,none": 0.2342857142857143,
399
- "acc_norm_stderr,none": 0.032109360396926204,
400
  "alias": " - cmmlu_professional_accounting"
401
  },
402
  "cmmlu_professional_law": {
403
- "acc,none": 0.2843601895734597,
404
- "acc_stderr,none": 0.031129489323148667,
405
- "acc_norm,none": 0.2843601895734597,
406
- "acc_norm_stderr,none": 0.031129489323148667,
407
  "alias": " - cmmlu_professional_law"
408
  },
409
  "cmmlu_professional_medicine": {
410
- "acc,none": 0.26595744680851063,
411
- "acc_stderr,none": 0.022816607010135298,
412
- "acc_norm,none": 0.26595744680851063,
413
- "acc_norm_stderr,none": 0.022816607010135298,
414
  "alias": " - cmmlu_professional_medicine"
415
  },
416
  "cmmlu_professional_psychology": {
@@ -428,61 +428,61 @@
428
  "alias": " - cmmlu_public_relations"
429
  },
430
  "cmmlu_security_study": {
431
- "acc,none": 0.24444444444444444,
432
- "acc_stderr,none": 0.03712537833614866,
433
- "acc_norm,none": 0.24444444444444444,
434
- "acc_norm_stderr,none": 0.03712537833614866,
435
  "alias": " - cmmlu_security_study"
436
  },
437
  "cmmlu_sociology": {
438
- "acc,none": 0.22566371681415928,
439
- "acc_stderr,none": 0.027867910955296744,
440
- "acc_norm,none": 0.22566371681415928,
441
- "acc_norm_stderr,none": 0.027867910955296744,
442
  "alias": " - cmmlu_sociology"
443
  },
444
  "cmmlu_sports_science": {
445
  "acc,none": 0.23030303030303031,
446
- "acc_stderr,none": 0.03287666758603489,
447
  "acc_norm,none": 0.23030303030303031,
448
- "acc_norm_stderr,none": 0.03287666758603489,
449
  "alias": " - cmmlu_sports_science"
450
  },
451
  "cmmlu_traditional_chinese_medicine": {
452
  "acc,none": 0.23243243243243245,
453
- "acc_stderr,none": 0.031138505170794653,
454
  "acc_norm,none": 0.23243243243243245,
455
- "acc_norm_stderr,none": 0.031138505170794653,
456
  "alias": " - cmmlu_traditional_chinese_medicine"
457
  },
458
  "cmmlu_virology": {
459
- "acc,none": 0.24260355029585798,
460
- "acc_stderr,none": 0.03307162750323179,
461
- "acc_norm,none": 0.24260355029585798,
462
- "acc_norm_stderr,none": 0.03307162750323179,
463
  "alias": " - cmmlu_virology"
464
  },
465
  "cmmlu_world_history": {
466
- "acc,none": 0.2236024844720497,
467
- "acc_stderr,none": 0.03293975688757214,
468
- "acc_norm,none": 0.2236024844720497,
469
- "acc_norm_stderr,none": 0.03293975688757214,
470
  "alias": " - cmmlu_world_history"
471
  },
472
  "cmmlu_world_religions": {
473
- "acc,none": 0.26875,
474
- "acc_stderr,none": 0.035156741348767645,
475
- "acc_norm,none": 0.26875,
476
- "acc_norm_stderr,none": 0.035156741348767645,
477
  "alias": " - cmmlu_world_religions"
478
  }
479
  },
480
  "groups": {
481
  "cmmlu": {
482
- "acc,none": 0.2485753755828009,
483
- "acc_stderr,none": 0.042815306797706565,
484
- "acc_norm,none": 0.2485753755828009,
485
- "acc_norm_stderr,none": 0.042815306797706565,
486
  "alias": "cmmlu"
487
  }
488
  },
@@ -3313,7 +3313,7 @@
3313
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
3314
  "batch_size": "auto",
3315
  "batch_sizes": [
3316
- 8
3317
  ],
3318
  "device": null,
3319
  "use_cache": null,
@@ -3321,5 +3321,5 @@
3321
  "bootstrap_iters": 100000,
3322
  "gen_kwargs": null
3323
  },
3324
- "git_hash": "2e3ceb0"
3325
  }
 
1
  {
2
  "results": {
3
  "cmmlu": {
4
+ "acc,none": 0.25004317043688495,
5
+ "acc_stderr,none": 0.043083884282466484,
6
+ "acc_norm,none": 0.25004317043688495,
7
+ "acc_norm_stderr,none": 0.043083884282466484,
8
  "alias": "cmmlu"
9
  },
10
  "cmmlu_agronomy": {
 
15
  "alias": " - cmmlu_agronomy"
16
  },
17
  "cmmlu_anatomy": {
18
+ "acc,none": 0.23648648648648649,
19
+ "acc_stderr,none": 0.035047162412504336,
20
+ "acc_norm,none": 0.23648648648648649,
21
+ "acc_norm_stderr,none": 0.035047162412504336,
22
  "alias": " - cmmlu_anatomy"
23
  },
24
  "cmmlu_ancient_chinese": {
 
29
  "alias": " - cmmlu_ancient_chinese"
30
  },
31
  "cmmlu_arts": {
32
+ "acc,none": 0.2875,
33
+ "acc_stderr,none": 0.035893251060583956,
34
+ "acc_norm,none": 0.2875,
35
+ "acc_norm_stderr,none": 0.035893251060583956,
36
  "alias": " - cmmlu_arts"
37
  },
38
  "cmmlu_astronomy": {
39
+ "acc,none": 0.2,
40
+ "acc_stderr,none": 0.031234752377721164,
41
+ "acc_norm,none": 0.2,
42
+ "acc_norm_stderr,none": 0.031234752377721164,
43
  "alias": " - cmmlu_astronomy"
44
  },
45
  "cmmlu_business_ethics": {
46
+ "acc,none": 0.24880382775119617,
47
+ "acc_stderr,none": 0.02997599063670254,
48
+ "acc_norm,none": 0.24880382775119617,
49
+ "acc_norm_stderr,none": 0.02997599063670254,
50
  "alias": " - cmmlu_business_ethics"
51
  },
52
  "cmmlu_chinese_civil_service_exam": {
53
+ "acc,none": 0.225,
54
+ "acc_stderr,none": 0.03311643267635493,
55
+ "acc_norm,none": 0.225,
56
+ "acc_norm_stderr,none": 0.03311643267635493,
57
  "alias": " - cmmlu_chinese_civil_service_exam"
58
  },
59
  "cmmlu_chinese_driving_rule": {
60
+ "acc,none": 0.2366412213740458,
61
+ "acc_stderr,none": 0.037276735755969195,
62
+ "acc_norm,none": 0.2366412213740458,
63
+ "acc_norm_stderr,none": 0.037276735755969195,
64
  "alias": " - cmmlu_chinese_driving_rule"
65
  },
66
  "cmmlu_chinese_food_culture": {
67
  "acc,none": 0.20588235294117646,
68
+ "acc_stderr,none": 0.03480046931235067,
69
  "acc_norm,none": 0.20588235294117646,
70
+ "acc_norm_stderr,none": 0.03480046931235067,
71
  "alias": " - cmmlu_chinese_food_culture"
72
  },
73
  "cmmlu_chinese_foreign_policy": {
 
78
  "alias": " - cmmlu_chinese_foreign_policy"
79
  },
80
  "cmmlu_chinese_history": {
81
+ "acc,none": 0.2848297213622291,
82
+ "acc_stderr,none": 0.02515182168617952,
83
+ "acc_norm,none": 0.2848297213622291,
84
+ "acc_norm_stderr,none": 0.02515182168617952,
85
  "alias": " - cmmlu_chinese_history"
86
  },
87
  "cmmlu_chinese_literature": {
88
  "acc,none": 0.22058823529411764,
89
+ "acc_stderr,none": 0.0291022543896741,
90
  "acc_norm,none": 0.22058823529411764,
91
+ "acc_norm_stderr,none": 0.0291022543896741,
92
  "alias": " - cmmlu_chinese_literature"
93
  },
94
  "cmmlu_chinese_teacher_qualification": {
95
+ "acc,none": 0.2346368715083799,
96
+ "acc_stderr,none": 0.03176302794175762,
97
+ "acc_norm,none": 0.2346368715083799,
98
+ "acc_norm_stderr,none": 0.03176302794175762,
99
  "alias": " - cmmlu_chinese_teacher_qualification"
100
  },
101
  "cmmlu_clinical_knowledge": {
102
+ "acc,none": 0.25316455696202533,
103
+ "acc_stderr,none": 0.028304657943035286,
104
+ "acc_norm,none": 0.25316455696202533,
105
+ "acc_norm_stderr,none": 0.028304657943035286,
106
  "alias": " - cmmlu_clinical_knowledge"
107
  },
108
  "cmmlu_college_actuarial_science": {
109
+ "acc,none": 0.29245283018867924,
110
+ "acc_stderr,none": 0.044392639061996274,
111
+ "acc_norm,none": 0.29245283018867924,
112
+ "acc_norm_stderr,none": 0.044392639061996274,
113
  "alias": " - cmmlu_college_actuarial_science"
114
  },
115
  "cmmlu_college_education": {
116
+ "acc,none": 0.3177570093457944,
117
+ "acc_stderr,none": 0.0452235007738203,
118
+ "acc_norm,none": 0.3177570093457944,
119
+ "acc_norm_stderr,none": 0.0452235007738203,
120
  "alias": " - cmmlu_college_education"
121
  },
122
  "cmmlu_college_engineering_hydrology": {
123
  "acc,none": 0.2641509433962264,
124
+ "acc_stderr,none": 0.0430254877395901,
125
  "acc_norm,none": 0.2641509433962264,
126
+ "acc_norm_stderr,none": 0.0430254877395901,
127
  "alias": " - cmmlu_college_engineering_hydrology"
128
  },
129
  "cmmlu_college_law": {
130
  "acc,none": 0.2962962962962963,
131
+ "acc_stderr,none": 0.04414343666854933,
132
  "acc_norm,none": 0.2962962962962963,
133
+ "acc_norm_stderr,none": 0.04414343666854933,
134
  "alias": " - cmmlu_college_law"
135
  },
136
  "cmmlu_college_mathematics": {
137
+ "acc,none": 0.3142857142857143,
138
+ "acc_stderr,none": 0.045521571818039494,
139
+ "acc_norm,none": 0.3142857142857143,
140
+ "acc_norm_stderr,none": 0.045521571818039494,
141
  "alias": " - cmmlu_college_mathematics"
142
  },
143
  "cmmlu_college_medical_statistics": {
144
+ "acc,none": 0.1792452830188679,
145
+ "acc_stderr,none": 0.037431386312552786,
146
+ "acc_norm,none": 0.1792452830188679,
147
+ "acc_norm_stderr,none": 0.037431386312552786,
148
  "alias": " - cmmlu_college_medical_statistics"
149
  },
150
  "cmmlu_college_medicine": {
 
155
  "alias": " - cmmlu_college_medicine"
156
  },
157
  "cmmlu_computer_science": {
158
+ "acc,none": 0.23529411764705882,
159
+ "acc_stderr,none": 0.02977177522814565,
160
+ "acc_norm,none": 0.23529411764705882,
161
+ "acc_norm_stderr,none": 0.02977177522814565,
162
  "alias": " - cmmlu_computer_science"
163
  },
164
  "cmmlu_computer_security": {
165
+ "acc,none": 0.2631578947368421,
166
+ "acc_stderr,none": 0.03377310252209193,
167
+ "acc_norm,none": 0.2631578947368421,
168
+ "acc_norm_stderr,none": 0.03377310252209193,
169
  "alias": " - cmmlu_computer_security"
170
  },
171
  "cmmlu_conceptual_physics": {
172
+ "acc,none": 0.22448979591836735,
173
+ "acc_stderr,none": 0.034531515032766795,
174
+ "acc_norm,none": 0.22448979591836735,
175
+ "acc_norm_stderr,none": 0.034531515032766795,
176
  "alias": " - cmmlu_conceptual_physics"
177
  },
178
  "cmmlu_construction_project_management": {
 
184
  },
185
  "cmmlu_economics": {
186
  "acc,none": 0.27672955974842767,
187
+ "acc_stderr,none": 0.03559177035707935,
188
  "acc_norm,none": 0.27672955974842767,
189
+ "acc_norm_stderr,none": 0.03559177035707935,
190
  "alias": " - cmmlu_economics"
191
  },
192
  "cmmlu_education": {
193
+ "acc,none": 0.2822085889570552,
194
+ "acc_stderr,none": 0.03536117886664743,
195
+ "acc_norm,none": 0.2822085889570552,
196
+ "acc_norm_stderr,none": 0.03536117886664743,
197
  "alias": " - cmmlu_education"
198
  },
199
  "cmmlu_electrical_engineering": {
200
+ "acc,none": 0.25,
201
+ "acc_stderr,none": 0.033113308926626096,
202
+ "acc_norm,none": 0.25,
203
+ "acc_norm_stderr,none": 0.033113308926626096,
204
  "alias": " - cmmlu_electrical_engineering"
205
  },
206
  "cmmlu_elementary_chinese": {
207
+ "acc,none": 0.24206349206349206,
208
+ "acc_stderr,none": 0.027036109679236982,
209
+ "acc_norm,none": 0.24206349206349206,
210
+ "acc_norm_stderr,none": 0.027036109679236982,
211
  "alias": " - cmmlu_elementary_chinese"
212
  },
213
  "cmmlu_elementary_commonsense": {
214
  "acc,none": 0.23737373737373738,
215
+ "acc_stderr,none": 0.0303137105381989,
216
  "acc_norm,none": 0.23737373737373738,
217
+ "acc_norm_stderr,none": 0.0303137105381989,
218
  "alias": " - cmmlu_elementary_commonsense"
219
  },
220
  "cmmlu_elementary_information_and_technology": {
221
+ "acc,none": 0.25630252100840334,
222
+ "acc_stderr,none": 0.02835962087053395,
223
+ "acc_norm,none": 0.25630252100840334,
224
+ "acc_norm_stderr,none": 0.02835962087053395,
225
  "alias": " - cmmlu_elementary_information_and_technology"
226
  },
227
  "cmmlu_elementary_mathematics": {
228
+ "acc,none": 0.2217391304347826,
229
+ "acc_stderr,none": 0.027451496604058923,
230
+ "acc_norm,none": 0.2217391304347826,
231
+ "acc_norm_stderr,none": 0.027451496604058923,
232
  "alias": " - cmmlu_elementary_mathematics"
233
  },
234
  "cmmlu_ethnology": {
235
+ "acc,none": 0.25925925925925924,
236
+ "acc_stderr,none": 0.03785714465066652,
237
+ "acc_norm,none": 0.25925925925925924,
238
+ "acc_norm_stderr,none": 0.03785714465066652,
239
  "alias": " - cmmlu_ethnology"
240
  },
241
  "cmmlu_food_science": {
242
+ "acc,none": 0.3006993006993007,
243
+ "acc_stderr,none": 0.03848167949490064,
244
+ "acc_norm,none": 0.3006993006993007,
245
+ "acc_norm_stderr,none": 0.03848167949490064,
246
  "alias": " - cmmlu_food_science"
247
  },
248
  "cmmlu_genetics": {
249
  "acc,none": 0.2897727272727273,
250
+ "acc_stderr,none": 0.034293230802398746,
251
  "acc_norm,none": 0.2897727272727273,
252
+ "acc_norm_stderr,none": 0.034293230802398746,
253
  "alias": " - cmmlu_genetics"
254
  },
255
  "cmmlu_global_facts": {
256
+ "acc,none": 0.2550335570469799,
257
+ "acc_stderr,none": 0.035829121651111746,
258
+ "acc_norm,none": 0.2550335570469799,
259
+ "acc_norm_stderr,none": 0.035829121651111746,
260
  "alias": " - cmmlu_global_facts"
261
  },
262
  "cmmlu_high_school_biology": {
263
+ "acc,none": 0.21893491124260356,
264
+ "acc_stderr,none": 0.03190409884491231,
265
+ "acc_norm,none": 0.21893491124260356,
266
+ "acc_norm_stderr,none": 0.03190409884491231,
267
  "alias": " - cmmlu_high_school_biology"
268
  },
269
  "cmmlu_high_school_chemistry": {
270
+ "acc,none": 0.25,
271
+ "acc_stderr,none": 0.037832495422898876,
272
+ "acc_norm,none": 0.25,
273
+ "acc_norm_stderr,none": 0.037832495422898876,
274
  "alias": " - cmmlu_high_school_chemistry"
275
  },
276
  "cmmlu_high_school_geography": {
277
+ "acc,none": 0.2288135593220339,
278
+ "acc_stderr,none": 0.03883538724538848,
279
+ "acc_norm,none": 0.2288135593220339,
280
+ "acc_norm_stderr,none": 0.03883538724538848,
281
  "alias": " - cmmlu_high_school_geography"
282
  },
283
  "cmmlu_high_school_mathematics": {
284
  "acc,none": 0.2621951219512195,
285
+ "acc_stderr,none": 0.03445000289173461,
286
  "acc_norm,none": 0.2621951219512195,
287
+ "acc_norm_stderr,none": 0.03445000289173461,
288
  "alias": " - cmmlu_high_school_mathematics"
289
  },
290
  "cmmlu_high_school_physics": {
291
+ "acc,none": 0.2727272727272727,
292
+ "acc_stderr,none": 0.04265792110940588,
293
+ "acc_norm,none": 0.2727272727272727,
294
+ "acc_norm_stderr,none": 0.04265792110940588,
295
  "alias": " - cmmlu_high_school_physics"
296
  },
297
  "cmmlu_high_school_politics": {
298
+ "acc,none": 0.2097902097902098,
299
+ "acc_stderr,none": 0.03416800637471349,
300
+ "acc_norm,none": 0.2097902097902098,
301
+ "acc_norm_stderr,none": 0.03416800637471349,
302
  "alias": " - cmmlu_high_school_politics"
303
  },
304
  "cmmlu_human_sexuality": {
305
+ "acc,none": 0.18253968253968253,
306
+ "acc_stderr,none": 0.034550710191021475,
307
+ "acc_norm,none": 0.18253968253968253,
308
+ "acc_norm_stderr,none": 0.034550710191021475,
309
  "alias": " - cmmlu_human_sexuality"
310
  },
311
  "cmmlu_international_law": {
312
+ "acc,none": 0.24324324324324326,
313
+ "acc_stderr,none": 0.03162930395697948,
314
+ "acc_norm,none": 0.24324324324324326,
315
+ "acc_norm_stderr,none": 0.03162930395697948,
316
  "alias": " - cmmlu_international_law"
317
  },
318
  "cmmlu_journalism": {
319
  "acc,none": 0.22674418604651161,
320
+ "acc_stderr,none": 0.03202075899584939,
321
  "acc_norm,none": 0.22674418604651161,
322
+ "acc_norm_stderr,none": 0.03202075899584939,
323
  "alias": " - cmmlu_journalism"
324
  },
325
  "cmmlu_jurisprudence": {
326
+ "acc,none": 0.2773722627737226,
327
+ "acc_stderr,none": 0.02211041530412192,
328
+ "acc_norm,none": 0.2773722627737226,
329
+ "acc_norm_stderr,none": 0.02211041530412192,
330
  "alias": " - cmmlu_jurisprudence"
331
  },
332
  "cmmlu_legal_and_moral_basis": {
333
+ "acc,none": 0.29439252336448596,
334
+ "acc_stderr,none": 0.0312287911542499,
335
+ "acc_norm,none": 0.29439252336448596,
336
+ "acc_norm_stderr,none": 0.0312287911542499,
337
  "alias": " - cmmlu_legal_and_moral_basis"
338
  },
339
  "cmmlu_logical": {
 
344
  "alias": " - cmmlu_logical"
345
  },
346
  "cmmlu_machine_learning": {
347
+ "acc,none": 0.30327868852459017,
348
+ "acc_stderr,none": 0.041788598786318756,
349
+ "acc_norm,none": 0.30327868852459017,
350
+ "acc_norm_stderr,none": 0.041788598786318756,
351
  "alias": " - cmmlu_machine_learning"
352
  },
353
  "cmmlu_management": {
354
+ "acc,none": 0.19047619047619047,
355
+ "acc_stderr,none": 0.02716201711702204,
356
+ "acc_norm,none": 0.19047619047619047,
357
+ "acc_norm_stderr,none": 0.02716201711702204,
358
  "alias": " - cmmlu_management"
359
  },
360
  "cmmlu_marketing": {
 
365
  "alias": " - cmmlu_marketing"
366
  },
367
  "cmmlu_marxist_theory": {
368
+ "acc,none": 0.24867724867724866,
369
+ "acc_stderr,none": 0.03152480234871163,
370
+ "acc_norm,none": 0.24867724867724866,
371
+ "acc_norm_stderr,none": 0.03152480234871163,
372
  "alias": " - cmmlu_marxist_theory"
373
  },
374
  "cmmlu_modern_chinese": {
375
+ "acc,none": 0.22413793103448276,
376
+ "acc_stderr,none": 0.03888669370117825,
377
+ "acc_norm,none": 0.22413793103448276,
378
+ "acc_norm_stderr,none": 0.03888669370117825,
379
  "alias": " - cmmlu_modern_chinese"
380
  },
381
  "cmmlu_nutrition": {
382
+ "acc,none": 0.2620689655172414,
383
+ "acc_stderr,none": 0.036646663372252565,
384
+ "acc_norm,none": 0.2620689655172414,
385
+ "acc_norm_stderr,none": 0.036646663372252565,
386
  "alias": " - cmmlu_nutrition"
387
  },
388
  "cmmlu_philosophy": {
389
+ "acc,none": 0.29523809523809524,
390
+ "acc_stderr,none": 0.044729159560441434,
391
+ "acc_norm,none": 0.29523809523809524,
392
+ "acc_norm_stderr,none": 0.044729159560441434,
393
  "alias": " - cmmlu_philosophy"
394
  },
395
  "cmmlu_professional_accounting": {
396
+ "acc,none": 0.24571428571428572,
397
+ "acc_stderr,none": 0.03263687142627841,
398
+ "acc_norm,none": 0.24571428571428572,
399
+ "acc_norm_stderr,none": 0.03263687142627841,
400
  "alias": " - cmmlu_professional_accounting"
401
  },
402
  "cmmlu_professional_law": {
403
+ "acc,none": 0.2796208530805687,
404
+ "acc_stderr,none": 0.030971033440870908,
405
+ "acc_norm,none": 0.2796208530805687,
406
+ "acc_norm_stderr,none": 0.030971033440870908,
407
  "alias": " - cmmlu_professional_law"
408
  },
409
  "cmmlu_professional_medicine": {
410
+ "acc,none": 0.2712765957446808,
411
+ "acc_stderr,none": 0.02296000025237266,
412
+ "acc_norm,none": 0.2712765957446808,
413
+ "acc_norm_stderr,none": 0.02296000025237266,
414
  "alias": " - cmmlu_professional_medicine"
415
  },
416
  "cmmlu_professional_psychology": {
 
428
  "alias": " - cmmlu_public_relations"
429
  },
430
  "cmmlu_security_study": {
431
+ "acc,none": 0.23703703703703705,
432
+ "acc_stderr,none": 0.03673731683969506,
433
+ "acc_norm,none": 0.23703703703703705,
434
+ "acc_norm_stderr,none": 0.03673731683969506,
435
  "alias": " - cmmlu_security_study"
436
  },
437
  "cmmlu_sociology": {
438
+ "acc,none": 0.23008849557522124,
439
+ "acc_stderr,none": 0.02805928483916018,
440
+ "acc_norm,none": 0.23008849557522124,
441
+ "acc_norm_stderr,none": 0.02805928483916018,
442
  "alias": " - cmmlu_sociology"
443
  },
444
  "cmmlu_sports_science": {
445
  "acc,none": 0.23030303030303031,
446
+ "acc_stderr,none": 0.0328766675860349,
447
  "acc_norm,none": 0.23030303030303031,
448
+ "acc_norm_stderr,none": 0.0328766675860349,
449
  "alias": " - cmmlu_sports_science"
450
  },
451
  "cmmlu_traditional_chinese_medicine": {
452
  "acc,none": 0.23243243243243245,
453
+ "acc_stderr,none": 0.03113850517079465,
454
  "acc_norm,none": 0.23243243243243245,
455
+ "acc_norm_stderr,none": 0.03113850517079465,
456
  "alias": " - cmmlu_traditional_chinese_medicine"
457
  },
458
  "cmmlu_virology": {
459
+ "acc,none": 0.2485207100591716,
460
+ "acc_stderr,none": 0.033341501981019636,
461
+ "acc_norm,none": 0.2485207100591716,
462
+ "acc_norm_stderr,none": 0.033341501981019636,
463
  "alias": " - cmmlu_virology"
464
  },
465
  "cmmlu_world_history": {
466
+ "acc,none": 0.2360248447204969,
467
+ "acc_stderr,none": 0.03357055232967969,
468
+ "acc_norm,none": 0.2360248447204969,
469
+ "acc_norm_stderr,none": 0.03357055232967969,
470
  "alias": " - cmmlu_world_history"
471
  },
472
  "cmmlu_world_religions": {
473
+ "acc,none": 0.2625,
474
+ "acc_stderr,none": 0.034893706520187605,
475
+ "acc_norm,none": 0.2625,
476
+ "acc_norm_stderr,none": 0.034893706520187605,
477
  "alias": " - cmmlu_world_religions"
478
  }
479
  },
480
  "groups": {
481
  "cmmlu": {
482
+ "acc,none": 0.25004317043688495,
483
+ "acc_stderr,none": 0.043083884282466484,
484
+ "acc_norm,none": 0.25004317043688495,
485
+ "acc_norm_stderr,none": 0.043083884282466484,
486
  "alias": "cmmlu"
487
  }
488
  },
 
3313
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
3314
  "batch_size": "auto",
3315
  "batch_sizes": [
3316
+ 32
3317
  ],
3318
  "device": null,
3319
  "use_cache": null,
 
3321
  "bootstrap_iters": 100000,
3322
  "gen_kwargs": null
3323
  },
3324
+ "git_hash": "4d19ea9"
3325
  }
lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cc69f38ccd1ff5eed2fe9dbdfb6e94b5d94c6c9d0c737ef0a4f268f3b0b4733c
3
- size 111568
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40520ce1ff8b1a23d138453962452ffe73d3be0c1507186529252b17242fec2a
3
+ size 115191
lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -2,7 +2,7 @@
2
  "results": {
3
  "cola": {
4
  "mcc,none": 0.003737743780434562,
5
- "mcc_stderr,none": 0.031171364680531898,
6
  "alias": "cola"
7
  }
8
  },
@@ -56,5 +56,5 @@
56
  "bootstrap_iters": 100000,
57
  "gen_kwargs": null
58
  },
59
- "git_hash": "2e3ceb0"
60
  }
 
2
  "results": {
3
  "cola": {
4
  "mcc,none": 0.003737743780434562,
5
+ "mcc_stderr,none": 0.031103768987297463,
6
  "alias": "cola"
7
  }
8
  },
 
56
  "bootstrap_iters": 100000,
57
  "gen_kwargs": null
58
  },
59
+ "git_hash": "4d19ea9"
60
  }
lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8ed94ed13ea33ac892b6c4b24f6f02443171923335c00eff64af1c1d89568425
3
- size 14251
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:daf1352c6866f32b198ffcb4c5221104e1dfde2a3267303c7e9cfa6b13861cc6
3
+ size 5472
lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "results": {
3
  "copa": {
4
- "acc,none": 0.86,
5
- "acc_stderr,none": 0.03487350880197769,
6
  "alias": "copa"
7
  }
8
  },
@@ -54,5 +54,5 @@
54
  "bootstrap_iters": 100000,
55
  "gen_kwargs": null
56
  },
57
- "git_hash": "2e3ceb0"
58
  }
 
1
  {
2
  "results": {
3
  "copa": {
4
+ "acc,none": 0.87,
5
+ "acc_stderr,none": 0.03379976689896309,
6
  "alias": "copa"
7
  }
8
  },
 
54
  "bootstrap_iters": 100000,
55
  "gen_kwargs": null
56
  },
57
+ "git_hash": "4d19ea9"
58
  }
lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1a5c1f036d78a7f1a94f65d7815126b3479c18e1be244c11d591c0be8c2b2526
3
- size 12890
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d706b949a88e229279cc1e0f890b69e20325d0b8b4c019c811f470cfa4b7632
3
+ size 2748
lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,173 +1,173 @@
1
  {
2
  "results": {
3
  "crows_pairs": {
4
- "likelihood_diff,none": 3.4815705128205128,
5
- "likelihood_diff_stderr,none": 0.47663007871133223,
6
- "pct_stereotype,none": 0.5603756708407871,
7
- "pct_stereotype_stderr,none": 0.0944668961491123,
8
  "alias": "crows_pairs"
9
  },
10
  "crows_pairs_english": {
11
- "likelihood_diff,none": 3.4655635062611805,
12
- "likelihood_diff_stderr,none": 0.08189260928515339,
13
- "pct_stereotype,none": 0.6469886702444841,
14
- "pct_stereotype_stderr,none": 0.011673622705751152,
15
  "alias": " - crows_pairs_english"
16
  },
17
  "crows_pairs_english_age": {
18
- "likelihood_diff,none": 3.7747252747252746,
19
- "likelihood_diff_stderr,none": 0.3733542426333088,
20
- "pct_stereotype,none": 0.7252747252747253,
21
- "pct_stereotype_stderr,none": 0.047052133987784364,
22
  "alias": " - crows_pairs_english_age"
23
  },
24
  "crows_pairs_english_autre": {
25
- "likelihood_diff,none": 5.715909090909091,
26
- "likelihood_diff_stderr,none": 1.6801050857531363,
27
  "pct_stereotype,none": 0.8181818181818182,
28
- "pct_stereotype_stderr,none": 0.12196734422726124,
29
  "alias": " - crows_pairs_english_autre"
30
  },
31
  "crows_pairs_english_disability": {
32
- "likelihood_diff,none": 6.015384615384615,
33
- "likelihood_diff_stderr,none": 0.5966623141775952,
34
- "pct_stereotype,none": 0.6923076923076923,
35
- "pct_stereotype_stderr,none": 0.05769230769230768,
36
  "alias": " - crows_pairs_english_disability"
37
  },
38
  "crows_pairs_english_gender": {
39
- "likelihood_diff,none": 2.60703125,
40
- "likelihood_diff_stderr,none": 0.157265210678921,
41
- "pct_stereotype,none": 0.6625,
42
- "pct_stereotype_stderr,none": 0.026474909752348248,
43
  "alias": " - crows_pairs_english_gender"
44
  },
45
  "crows_pairs_english_nationality": {
46
- "likelihood_diff,none": 3.392361111111111,
47
- "likelihood_diff_stderr,none": 0.22640739112896405,
48
- "pct_stereotype,none": 0.6018518518518519,
49
- "pct_stereotype_stderr,none": 0.033384734032074016,
50
  "alias": " - crows_pairs_english_nationality"
51
  },
52
  "crows_pairs_english_physical_appearance": {
53
- "likelihood_diff,none": 3.779513888888889,
54
- "likelihood_diff_stderr,none": 0.3056216112222913,
55
  "pct_stereotype,none": 0.7638888888888888,
56
  "pct_stereotype_stderr,none": 0.050401578099733044,
57
  "alias": " - crows_pairs_english_physical_appearance"
58
  },
59
  "crows_pairs_english_race_color": {
60
- "likelihood_diff,none": 3.1764271653543306,
61
- "likelihood_diff_stderr,none": 0.1388300491868073,
62
- "pct_stereotype,none": 0.5413385826771654,
63
- "pct_stereotype_stderr,none": 0.022129755490549068,
64
  "alias": " - crows_pairs_english_race_color"
65
  },
66
  "crows_pairs_english_religion": {
67
- "likelihood_diff,none": 3.310810810810811,
68
- "likelihood_diff_stderr,none": 0.28480970226640806,
69
- "pct_stereotype,none": 0.7657657657657657,
70
- "pct_stereotype_stderr,none": 0.04038097636567093,
71
  "alias": " - crows_pairs_english_religion"
72
  },
73
  "crows_pairs_english_sexual_orientation": {
74
- "likelihood_diff,none": 4.282258064516129,
75
- "likelihood_diff_stderr,none": 0.4391141133147601,
76
  "pct_stereotype,none": 0.8602150537634409,
77
  "pct_stereotype_stderr,none": 0.036152622588464155,
78
  "alias": " - crows_pairs_english_sexual_orientation"
79
  },
80
  "crows_pairs_english_socioeconomic": {
81
- "likelihood_diff,none": 4.142763157894737,
82
- "likelihood_diff_stderr,none": 0.23880639175146945,
83
- "pct_stereotype,none": 0.6578947368421053,
84
- "pct_stereotype_stderr,none": 0.03450858738901065,
85
  "alias": " - crows_pairs_english_socioeconomic"
86
  },
87
  "crows_pairs_french": {
88
- "likelihood_diff,none": 3.497875670840787,
89
- "likelihood_diff_stderr,none": 0.08141456707301374,
90
- "pct_stereotype,none": 0.4752534287418008,
91
- "pct_stereotype_stderr,none": 0.012198331374086789,
92
  "alias": " - crows_pairs_french"
93
  },
94
  "crows_pairs_french_age": {
95
- "likelihood_diff,none": 3.227777777777778,
96
- "likelihood_diff_stderr,none": 0.34426468266405463,
97
- "pct_stereotype,none": 0.4444444444444444,
98
- "pct_stereotype_stderr,none": 0.052671718126664185,
99
  "alias": " - crows_pairs_french_age"
100
  },
101
  "crows_pairs_french_autre": {
102
- "likelihood_diff,none": 2.9615384615384617,
103
- "likelihood_diff_stderr,none": 0.704180598697418,
104
  "pct_stereotype,none": 0.5384615384615384,
105
  "pct_stereotype_stderr,none": 0.14390989949130545,
106
  "alias": " - crows_pairs_french_autre"
107
  },
108
  "crows_pairs_french_disability": {
109
- "likelihood_diff,none": 5.371212121212121,
110
- "likelihood_diff_stderr,none": 0.5079565770380846,
111
  "pct_stereotype,none": 0.6212121212121212,
112
  "pct_stereotype_stderr,none": 0.0601674102524024,
113
  "alias": " - crows_pairs_french_disability"
114
  },
115
  "crows_pairs_french_gender": {
116
- "likelihood_diff,none": 2.80607476635514,
117
- "likelihood_diff_stderr,none": 0.1529387267167066,
118
- "pct_stereotype,none": 0.48909657320872274,
119
- "pct_stereotype_stderr,none": 0.027944203070818643,
120
  "alias": " - crows_pairs_french_gender"
121
  },
122
  "crows_pairs_french_nationality": {
123
- "likelihood_diff,none": 4.414031620553359,
124
- "likelihood_diff_stderr,none": 0.23840899623275094,
125
- "pct_stereotype,none": 0.31225296442687744,
126
- "pct_stereotype_stderr,none": 0.02919223713357907,
127
  "alias": " - crows_pairs_french_nationality"
128
  },
129
  "crows_pairs_french_physical_appearance": {
130
- "likelihood_diff,none": 3.6770833333333335,
131
- "likelihood_diff_stderr,none": 0.4180562948095515,
132
- "pct_stereotype,none": 0.5138888888888888,
133
- "pct_stereotype_stderr,none": 0.05931618532716555,
134
  "alias": " - crows_pairs_french_physical_appearance"
135
  },
136
  "crows_pairs_french_race_color": {
137
- "likelihood_diff,none": 3.0456521739130435,
138
- "likelihood_diff_stderr,none": 0.1433029684316863,
139
- "pct_stereotype,none": 0.41304347826086957,
140
- "pct_stereotype_stderr,none": 0.022982353907431446,
141
  "alias": " - crows_pairs_french_race_color"
142
  },
143
  "crows_pairs_french_religion": {
144
- "likelihood_diff,none": 3.6543478260869566,
145
- "likelihood_diff_stderr,none": 0.33007440000905736,
146
  "pct_stereotype,none": 0.6,
147
- "pct_stereotype_stderr,none": 0.04588314677411234,
148
  "alias": " - crows_pairs_french_religion"
149
  },
150
  "crows_pairs_french_sexual_orientation": {
151
- "likelihood_diff,none": 4.087912087912088,
152
- "likelihood_diff_stderr,none": 0.3075508243527189,
153
- "pct_stereotype,none": 0.7472527472527473,
154
- "pct_stereotype_stderr,none": 0.04580951853732889,
155
  "alias": " - crows_pairs_french_sexual_orientation"
156
  },
157
  "crows_pairs_french_socioeconomic": {
158
- "likelihood_diff,none": 3.6463647959183674,
159
- "likelihood_diff_stderr,none": 0.24452528878254828,
160
- "pct_stereotype,none": 0.5459183673469388,
161
- "pct_stereotype_stderr,none": 0.035654431417332814,
162
  "alias": " - crows_pairs_french_socioeconomic"
163
  }
164
  },
165
  "groups": {
166
  "crows_pairs": {
167
- "likelihood_diff,none": 3.4815705128205128,
168
- "likelihood_diff_stderr,none": 0.47663007871133223,
169
- "pct_stereotype,none": 0.5603756708407871,
170
- "pct_stereotype_stderr,none": 0.0944668961491123,
171
  "alias": "crows_pairs"
172
  }
173
  },
@@ -1048,5 +1048,5 @@
1048
  "bootstrap_iters": 100000,
1049
  "gen_kwargs": null
1050
  },
1051
- "git_hash": "2e3ceb0"
1052
  }
 
1
  {
2
  "results": {
3
  "crows_pairs": {
4
+ "likelihood_diff,none": 3.4772659511031603,
5
+ "likelihood_diff_stderr,none": 0.49088835451046214,
6
+ "pct_stereotype,none": 0.5608228980322003,
7
+ "pct_stereotype_stderr,none": 0.09848087406351029,
8
  "alias": "crows_pairs"
9
  },
10
  "crows_pairs_english": {
11
+ "likelihood_diff,none": 3.465638044126416,
12
+ "likelihood_diff_stderr,none": 0.0817674558279069,
13
+ "pct_stereotype,none": 0.6446034585569469,
14
+ "pct_stereotype_stderr,none": 0.011691383517451213,
15
  "alias": " - crows_pairs_english"
16
  },
17
  "crows_pairs_english_age": {
18
+ "likelihood_diff,none": 3.79532967032967,
19
+ "likelihood_diff_stderr,none": 0.3741520555832592,
20
+ "pct_stereotype,none": 0.7362637362637363,
21
+ "pct_stereotype_stderr,none": 0.046449428524973954,
22
  "alias": " - crows_pairs_english_age"
23
  },
24
  "crows_pairs_english_autre": {
25
+ "likelihood_diff,none": 5.693181818181818,
26
+ "likelihood_diff_stderr,none": 1.6696967978331319,
27
  "pct_stereotype,none": 0.8181818181818182,
28
+ "pct_stereotype_stderr,none": 0.12196734422726127,
29
  "alias": " - crows_pairs_english_autre"
30
  },
31
  "crows_pairs_english_disability": {
32
+ "likelihood_diff,none": 5.996153846153846,
33
+ "likelihood_diff_stderr,none": 0.5963188938712826,
34
+ "pct_stereotype,none": 0.676923076923077,
35
+ "pct_stereotype_stderr,none": 0.05845647751373334,
36
  "alias": " - crows_pairs_english_disability"
37
  },
38
  "crows_pairs_english_gender": {
39
+ "likelihood_diff,none": 2.61015625,
40
+ "likelihood_diff_stderr,none": 0.1573761763903728,
41
+ "pct_stereotype,none": 0.659375,
42
+ "pct_stereotype_stderr,none": 0.026534392975531496,
43
  "alias": " - crows_pairs_english_gender"
44
  },
45
  "crows_pairs_english_nationality": {
46
+ "likelihood_diff,none": 3.419560185185185,
47
+ "likelihood_diff_stderr,none": 0.22476506938859697,
48
+ "pct_stereotype,none": 0.6111111111111112,
49
+ "pct_stereotype_stderr,none": 0.03324708911809117,
50
  "alias": " - crows_pairs_english_nationality"
51
  },
52
  "crows_pairs_english_physical_appearance": {
53
+ "likelihood_diff,none": 3.828125,
54
+ "likelihood_diff_stderr,none": 0.3062704344289238,
55
  "pct_stereotype,none": 0.7638888888888888,
56
  "pct_stereotype_stderr,none": 0.050401578099733044,
57
  "alias": " - crows_pairs_english_physical_appearance"
58
  },
59
  "crows_pairs_english_race_color": {
60
+ "likelihood_diff,none": 3.176919291338583,
61
+ "likelihood_diff_stderr,none": 0.13862152723353277,
62
+ "pct_stereotype,none": 0.5393700787401575,
63
+ "pct_stereotype_stderr,none": 0.022136834498576036,
64
  "alias": " - crows_pairs_english_race_color"
65
  },
66
  "crows_pairs_english_religion": {
67
+ "likelihood_diff,none": 3.3355855855855854,
68
+ "likelihood_diff_stderr,none": 0.2889364817644626,
69
+ "pct_stereotype,none": 0.7477477477477478,
70
+ "pct_stereotype_stderr,none": 0.04140938118194942,
71
  "alias": " - crows_pairs_english_religion"
72
  },
73
  "crows_pairs_english_sexual_orientation": {
74
+ "likelihood_diff,none": 4.259408602150538,
75
+ "likelihood_diff_stderr,none": 0.43254953377275895,
76
  "pct_stereotype,none": 0.8602150537634409,
77
  "pct_stereotype_stderr,none": 0.036152622588464155,
78
  "alias": " - crows_pairs_english_sexual_orientation"
79
  },
80
  "crows_pairs_english_socioeconomic": {
81
+ "likelihood_diff,none": 4.128289473684211,
82
+ "likelihood_diff_stderr,none": 0.23747047298192855,
83
+ "pct_stereotype,none": 0.6631578947368421,
84
+ "pct_stereotype_stderr,none": 0.03437880340748323,
85
  "alias": " - crows_pairs_english_socioeconomic"
86
  },
87
  "crows_pairs_french": {
88
+ "likelihood_diff,none": 3.4888938580799045,
89
+ "likelihood_diff_stderr,none": 0.08156240137076916,
90
+ "pct_stereotype,none": 0.4770423375074538,
91
+ "pct_stereotype_stderr,none": 0.01220041828317914,
92
  "alias": " - crows_pairs_french"
93
  },
94
  "crows_pairs_french_age": {
95
+ "likelihood_diff,none": 3.2333333333333334,
96
+ "likelihood_diff_stderr,none": 0.34675236624373657,
97
+ "pct_stereotype,none": 0.45555555555555555,
98
+ "pct_stereotype_stderr,none": 0.05279009646630345,
99
  "alias": " - crows_pairs_french_age"
100
  },
101
  "crows_pairs_french_autre": {
102
+ "likelihood_diff,none": 2.9423076923076925,
103
+ "likelihood_diff_stderr,none": 0.6800463350033433,
104
  "pct_stereotype,none": 0.5384615384615384,
105
  "pct_stereotype_stderr,none": 0.14390989949130545,
106
  "alias": " - crows_pairs_french_autre"
107
  },
108
  "crows_pairs_french_disability": {
109
+ "likelihood_diff,none": 5.412878787878788,
110
+ "likelihood_diff_stderr,none": 0.4983507889678439,
111
  "pct_stereotype,none": 0.6212121212121212,
112
  "pct_stereotype_stderr,none": 0.0601674102524024,
113
  "alias": " - crows_pairs_french_disability"
114
  },
115
  "crows_pairs_french_gender": {
116
+ "likelihood_diff,none": 2.787772585669782,
117
+ "likelihood_diff_stderr,none": 0.15454171718993162,
118
+ "pct_stereotype,none": 0.4735202492211838,
119
+ "pct_stereotype_stderr,none": 0.027911625198936637,
120
  "alias": " - crows_pairs_french_gender"
121
  },
122
  "crows_pairs_french_nationality": {
123
+ "likelihood_diff,none": 4.394268774703558,
124
+ "likelihood_diff_stderr,none": 0.2393255401900638,
125
+ "pct_stereotype,none": 0.30434782608695654,
126
+ "pct_stereotype_stderr,none": 0.028985507246376746,
127
  "alias": " - crows_pairs_french_nationality"
128
  },
129
  "crows_pairs_french_physical_appearance": {
130
+ "likelihood_diff,none": 3.7291666666666665,
131
+ "likelihood_diff_stderr,none": 0.42202395141887994,
132
+ "pct_stereotype,none": 0.5416666666666666,
133
+ "pct_stereotype_stderr,none": 0.05913268547421809,
134
  "alias": " - crows_pairs_french_physical_appearance"
135
  },
136
  "crows_pairs_french_race_color": {
137
+ "likelihood_diff,none": 3.0375,
138
+ "likelihood_diff_stderr,none": 0.14411452077441483,
139
+ "pct_stereotype,none": 0.43043478260869567,
140
+ "pct_stereotype_stderr,none": 0.023111017495849547,
141
  "alias": " - crows_pairs_french_race_color"
142
  },
143
  "crows_pairs_french_religion": {
144
+ "likelihood_diff,none": 3.626086956521739,
145
+ "likelihood_diff_stderr,none": 0.33040341695448183,
146
  "pct_stereotype,none": 0.6,
147
+ "pct_stereotype_stderr,none": 0.04588314677411235,
148
  "alias": " - crows_pairs_french_religion"
149
  },
150
  "crows_pairs_french_sexual_orientation": {
151
+ "likelihood_diff,none": 4.024725274725275,
152
+ "likelihood_diff_stderr,none": 0.30015906815281673,
153
+ "pct_stereotype,none": 0.7362637362637363,
154
+ "pct_stereotype_stderr,none": 0.046449428524973954,
155
  "alias": " - crows_pairs_french_sexual_orientation"
156
  },
157
  "crows_pairs_french_socioeconomic": {
158
+ "likelihood_diff,none": 3.6160714285714284,
159
+ "likelihood_diff_stderr,none": 0.24417182803609577,
160
+ "pct_stereotype,none": 0.5561224489795918,
161
+ "pct_stereotype_stderr,none": 0.035579471949536604,
162
  "alias": " - crows_pairs_french_socioeconomic"
163
  }
164
  },
165
  "groups": {
166
  "crows_pairs": {
167
+ "likelihood_diff,none": 3.4772659511031603,
168
+ "likelihood_diff_stderr,none": 0.49088835451046214,
169
+ "pct_stereotype,none": 0.5608228980322003,
170
+ "pct_stereotype_stderr,none": 0.09848087406351029,
171
  "alias": "crows_pairs"
172
  }
173
  },
 
1048
  "bootstrap_iters": 100000,
1049
  "gen_kwargs": null
1050
  },
1051
+ "git_hash": "4d19ea9"
1052
  }
lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ebfa42286734a17eafc8caed854007f738497389888b4b6d224186002d0c5645
3
- size 113857
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20cf09501a056adb16025c05bdd901db65a6aabf35f44eec48fdd8538535cbcf
3
+ size 31645
lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,20 +1,20 @@
1
  {
2
  "results": {
3
  "freebase": {
4
- "exact_match,none": 0.028543307086614175,
5
- "exact_match_stderr,none": 0.0036949528903927557,
6
  "alias": "freebase"
7
  },
8
  "webqs": {
9
- "exact_match,none": 0.028543307086614175,
10
- "exact_match_stderr,none": 0.0036949528903927557,
11
  "alias": " - webqs"
12
  }
13
  },
14
  "groups": {
15
  "freebase": {
16
- "exact_match,none": 0.028543307086614175,
17
- "exact_match_stderr,none": 0.0036949528903927557,
18
  "alias": "freebase"
19
  }
20
  },
@@ -70,5 +70,5 @@
70
  "bootstrap_iters": 100000,
71
  "gen_kwargs": null
72
  },
73
- "git_hash": "2e3ceb0"
74
  }
 
1
  {
2
  "results": {
3
  "freebase": {
4
+ "exact_match,none": 0.029035433070866142,
5
+ "exact_match_stderr,none": 0.0037257257477226868,
6
  "alias": "freebase"
7
  },
8
  "webqs": {
9
+ "exact_match,none": 0.029035433070866142,
10
+ "exact_match_stderr,none": 0.0037257257477226868,
11
  "alias": " - webqs"
12
  }
13
  },
14
  "groups": {
15
  "freebase": {
16
+ "exact_match,none": 0.029035433070866142,
17
+ "exact_match_stderr,none": 0.0037257257477226868,
18
  "alias": "freebase"
19
  }
20
  },
 
70
  "bootstrap_iters": 100000,
71
  "gen_kwargs": null
72
  },
73
+ "git_hash": "4d19ea9"
74
  }
lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8d914584ef70ee0864d3088a1baf759abdc27445c582b251bfe3119a8d31f6f4
3
- size 12737
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a335fe8dc570aa92826ceb9807da45e21044ae57e02699d5cea12dac18e3c26
3
+ size 7322
lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,56 +1,56 @@
1
  {
2
  "results": {
3
  "glue": {
4
- "acc,none": 0.47458134817465414,
5
- "acc_stderr,none": 0.07691289703225153,
6
- "f1,none": 0.444689593964169,
7
- "f1_stderr,none": 0.0013063875646580627,
8
  "mcc,none": 0.003737743780434562,
9
- "mcc_stderr,none": 0.0009741365404113992,
10
  "alias": "glue"
11
  },
12
  "cola": {
13
  "mcc,none": 0.003737743780434562,
14
- "mcc_stderr,none": 0.03121116051048726,
15
  "alias": " - cola"
16
  },
17
  "mnli": {
18
- "acc,none": 0.32929190015282733,
19
- "acc_stderr,none": 0.004743886315223882,
20
  "alias": " - mnli"
21
  },
22
  "mnli_mismatch": {
23
- "acc,none": 0.33350284784377543,
24
- "acc_stderr,none": 0.0047549951070959134,
25
  "alias": " - mnli_mismatch"
26
  },
27
  "mrpc": {
28
  "acc,none": 0.6838235294117647,
29
- "acc_stderr,none": 0.02304833666842021,
30
  "f1,none": 0.8122270742358079,
31
- "f1_stderr,none": 0.016218335300780515,
32
  "alias": " - mrpc"
33
  },
34
  "qnli": {
35
- "acc,none": 0.49697968149368477,
36
- "acc_stderr,none": 0.0067652871181183415,
37
  "alias": " - qnli"
38
  },
39
  "qqp": {
40
- "acc,none": 0.5346772198862231,
41
- "acc_stderr,none": 0.002480712860000902,
42
- "f1,none": 0.4411039481892992,
43
- "f1_stderr,none": 0.003378627814144618,
44
  "alias": " - qqp"
45
  },
46
  "rte": {
47
- "acc,none": 0.5451263537906137,
48
- "acc_stderr,none": 0.029973636495415252,
49
  "alias": " - rte"
50
  },
51
  "sst2": {
52
  "acc,none": 0.573394495412844,
53
- "acc_stderr,none": 0.016758336618033463,
54
  "alias": " - sst2"
55
  },
56
  "wnli": {
@@ -61,12 +61,12 @@
61
  },
62
  "groups": {
63
  "glue": {
64
- "acc,none": 0.47458134817465414,
65
- "acc_stderr,none": 0.07691289703225153,
66
- "f1,none": 0.444689593964169,
67
- "f1_stderr,none": 0.0013063875646580627,
68
  "mcc,none": 0.003737743780434562,
69
- "mcc_stderr,none": 0.0009741365404113992,
70
  "alias": "glue"
71
  }
72
  },
@@ -362,7 +362,7 @@
362
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
363
  "batch_size": "auto",
364
  "batch_sizes": [
365
- 16
366
  ],
367
  "device": null,
368
  "use_cache": null,
@@ -370,5 +370,5 @@
370
  "bootstrap_iters": 100000,
371
  "gen_kwargs": null
372
  },
373
- "git_hash": "2e3ceb0"
374
  }
 
1
  {
2
  "results": {
3
  "glue": {
4
+ "acc,none": 0.47397570271557893,
5
+ "acc_stderr,none": 0.006104309039981269,
6
+ "f1,none": 0.4450259440671591,
7
+ "f1_stderr,none": 0.0011613802806875427,
8
  "mcc,none": 0.003737743780434562,
9
+ "mcc_stderr,none": 0.031103768987297463,
10
  "alias": "glue"
11
  },
12
  "cola": {
13
  "mcc,none": 0.003737743780434562,
14
+ "mcc_stderr,none": 0.031103768987297463,
15
  "alias": " - cola"
16
  },
17
  "mnli": {
18
+ "acc,none": 0.3295975547631177,
19
+ "acc_stderr,none": 0.004745005919447844,
20
  "alias": " - mnli"
21
  },
22
  "mnli_mismatch": {
23
+ "acc,none": 0.3350284784377543,
24
+ "acc_stderr,none": 0.004760400998434307,
25
  "alias": " - mnli_mismatch"
26
  },
27
  "mrpc": {
28
  "acc,none": 0.6838235294117647,
29
+ "acc_stderr,none": 0.023048336668420193,
30
  "f1,none": 0.8122270742358079,
31
+ "f1_stderr,none": 0.016275484057001473,
32
  "alias": " - mrpc"
33
  },
34
  "qnli": {
35
+ "acc,none": 0.4962474830679114,
36
+ "acc_stderr,none": 0.006765220016415221,
37
  "alias": " - qnli"
38
  },
39
  "qqp": {
40
+ "acc,none": 0.5348256245362355,
41
+ "acc_stderr,none": 0.0024806614372752606,
42
+ "f1,none": 0.44184597121234603,
43
+ "f1_stderr,none": 0.0033853946882924225,
44
  "alias": " - qqp"
45
  },
46
  "rte": {
47
+ "acc,none": 0.5523465703971119,
48
+ "acc_stderr,none": 0.02993107036293953,
49
  "alias": " - rte"
50
  },
51
  "sst2": {
52
  "acc,none": 0.573394495412844,
53
+ "acc_stderr,none": 0.016758336618033467,
54
  "alias": " - sst2"
55
  },
56
  "wnli": {
 
61
  },
62
  "groups": {
63
  "glue": {
64
+ "acc,none": 0.47397570271557893,
65
+ "acc_stderr,none": 0.006104309039981269,
66
+ "f1,none": 0.4450259440671591,
67
+ "f1_stderr,none": 0.0011613802806875427,
68
  "mcc,none": 0.003737743780434562,
69
+ "mcc_stderr,none": 0.031103768987297463,
70
  "alias": "glue"
71
  }
72
  },
 
362
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
363
  "batch_size": "auto",
364
  "batch_sizes": [
365
+ 64
366
  ],
367
  "device": null,
368
  "use_cache": null,
 
370
  "bootstrap_iters": 100000,
371
  "gen_kwargs": null
372
  },
373
+ "git_hash": "4d19ea9"
374
  }
lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:00eab5e98b32ede57f9fcbc2e65ab8d2b3686aba7dbb4279ac2da0dae0d2dd25
3
- size 104469
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5eee6e04dbf2d353f826c53c313deeaab74af28c3ea4e3da4eed1d5e7e96863a
3
+ size 173047
lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "results": {
3
  "gsm8k": {
4
- "exact_match,get-answer": 0.04245640636846096,
5
- "exact_match_stderr,get-answer": 0.005553837749990044,
6
  "alias": "gsm8k"
7
  }
8
  },
@@ -84,5 +84,5 @@
84
  "bootstrap_iters": 100000,
85
  "gen_kwargs": null
86
  },
87
- "git_hash": "2e3ceb0"
88
  }
 
1
  {
2
  "results": {
3
  "gsm8k": {
4
+ "exact_match,get-answer": 0.050037907505686124,
5
+ "exact_match_stderr,get-answer": 0.006005442354577737,
6
  "alias": "gsm8k"
7
  }
8
  },
 
84
  "bootstrap_iters": 100000,
85
  "gen_kwargs": null
86
  },
87
+ "git_hash": "4d19ea9"
88
  }
lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:10fe1ee1afc331d8586fd52aa8cac721c131f3700001f647dcfdd24fed23d6d9
3
- size 15430
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78ca123d7f8c685cd690b01779f1aaa9d548ab5573e33a3e6bedf2093ababe38
3
+ size 7584
lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -2,9 +2,9 @@
2
  "results": {
3
  "hellaswag": {
4
  "acc,none": 0.5571599283011353,
5
- "acc_stderr,none": 0.0049570683775165105,
6
- "acc_norm,none": 0.755327623979287,
7
- "acc_norm_stderr,none": 0.0042901420299216834,
8
  "alias": "hellaswag"
9
  }
10
  },
@@ -55,7 +55,7 @@
55
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
56
  "batch_size": "auto",
57
  "batch_sizes": [
58
- 32
59
  ],
60
  "device": null,
61
  "use_cache": null,
@@ -63,5 +63,5 @@
63
  "bootstrap_iters": 100000,
64
  "gen_kwargs": null
65
  },
66
- "git_hash": "2e3ceb0"
67
  }
 
2
  "results": {
3
  "hellaswag": {
4
  "acc,none": 0.5571599283011353,
5
+ "acc_stderr,none": 0.004957068377516497,
6
+ "acc_norm,none": 0.7550288787094205,
7
+ "acc_norm_stderr,none": 0.004291911350430623,
8
  "alias": "hellaswag"
9
  }
10
  },
 
55
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
56
  "batch_size": "auto",
57
  "batch_sizes": [
58
+ 64
59
  ],
60
  "device": null,
61
  "use_cache": null,
 
63
  "bootstrap_iters": 100000,
64
  "gen_kwargs": null
65
  },
66
+ "git_hash": "4d19ea9"
67
  }
lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4b624507a14b86e057bb2c74dfd645cf274d3ca96d6ae412b2c2621c270c1193
3
- size 24651
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95ce074d7f77a8be709130d3e3daf446f7b0b2a967bbc04ebb6c9804ec968a4c
3
+ size 43589
lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,87 +1,87 @@
1
  {
2
  "results": {
3
  "kmmlu": {
4
- "acc,none": 0.26936182500721906,
5
- "acc_stderr,none": 0.024299898599051413,
6
- "acc_norm,none": 0.26936182500721906,
7
- "acc_norm_stderr,none": 0.024299898599051413,
8
  "alias": "kmmlu"
9
  },
10
  "kmmlu_accounting": {
11
- "acc,none": 0.29,
12
- "acc_stderr,none": 0.045604802157206845,
13
- "acc_norm,none": 0.29,
14
- "acc_norm_stderr,none": 0.045604802157206845,
15
  "alias": " - kmmlu_accounting"
16
  },
17
  "kmmlu_agricultural_sciences": {
18
- "acc,none": 0.261,
19
- "acc_stderr,none": 0.013895037677965136,
20
- "acc_norm,none": 0.261,
21
- "acc_norm_stderr,none": 0.013895037677965136,
22
  "alias": " - kmmlu_agricultural_sciences"
23
  },
24
  "kmmlu_aviation_engineering_and_maintenance": {
25
- "acc,none": 0.27,
26
- "acc_stderr,none": 0.014046255632633913,
27
- "acc_norm,none": 0.27,
28
- "acc_norm_stderr,none": 0.014046255632633913,
29
  "alias": " - kmmlu_aviation_engineering_and_maintenance"
30
  },
31
  "kmmlu_biology": {
32
- "acc,none": 0.26,
33
- "acc_stderr,none": 0.013877773329774164,
34
- "acc_norm,none": 0.26,
35
- "acc_norm_stderr,none": 0.013877773329774164,
36
  "alias": " - kmmlu_biology"
37
  },
38
  "kmmlu_chemical_engineering": {
39
- "acc,none": 0.27,
40
- "acc_stderr,none": 0.014046255632633916,
41
- "acc_norm,none": 0.27,
42
- "acc_norm_stderr,none": 0.014046255632633916,
43
  "alias": " - kmmlu_chemical_engineering"
44
  },
45
  "kmmlu_chemistry": {
46
- "acc,none": 0.26666666666666666,
47
- "acc_stderr,none": 0.01806848202433441,
48
- "acc_norm,none": 0.26666666666666666,
49
- "acc_norm_stderr,none": 0.01806848202433441,
50
  "alias": " - kmmlu_chemistry"
51
  },
52
  "kmmlu_civil_engineering": {
53
- "acc,none": 0.315,
54
- "acc_stderr,none": 0.014696631960792498,
55
- "acc_norm,none": 0.315,
56
- "acc_norm_stderr,none": 0.014696631960792498,
57
  "alias": " - kmmlu_civil_engineering"
58
  },
59
  "kmmlu_computer_science": {
60
- "acc,none": 0.279,
61
- "acc_stderr,none": 0.014190150117612033,
62
- "acc_norm,none": 0.279,
63
- "acc_norm_stderr,none": 0.014190150117612033,
64
  "alias": " - kmmlu_computer_science"
65
  },
66
  "kmmlu_construction": {
67
- "acc,none": 0.284,
68
- "acc_stderr,none": 0.014267009061031314,
69
- "acc_norm,none": 0.284,
70
- "acc_norm_stderr,none": 0.014267009061031314,
71
  "alias": " - kmmlu_construction"
72
  },
73
  "kmmlu_criminal_law": {
74
- "acc,none": 0.26,
75
- "acc_stderr,none": 0.03109395714370027,
76
- "acc_norm,none": 0.26,
77
- "acc_norm_stderr,none": 0.03109395714370027,
78
  "alias": " - kmmlu_criminal_law"
79
  },
80
  "kmmlu_ecology": {
81
- "acc,none": 0.276,
82
- "acc_stderr,none": 0.014142984975740663,
83
- "acc_norm,none": 0.276,
84
- "acc_norm_stderr,none": 0.014142984975740663,
85
  "alias": " - kmmlu_ecology"
86
  },
87
  "kmmlu_economics": {
@@ -92,66 +92,66 @@
92
  "alias": " - kmmlu_economics"
93
  },
94
  "kmmlu_education": {
95
- "acc,none": 0.23,
96
- "acc_stderr,none": 0.04229525846816505,
97
- "acc_norm,none": 0.23,
98
- "acc_norm_stderr,none": 0.04229525846816505,
99
  "alias": " - kmmlu_education"
100
  },
101
  "kmmlu_electrical_engineering": {
102
- "acc,none": 0.309,
103
- "acc_stderr,none": 0.014619600977206493,
104
- "acc_norm,none": 0.309,
105
- "acc_norm_stderr,none": 0.014619600977206493,
106
  "alias": " - kmmlu_electrical_engineering"
107
  },
108
  "kmmlu_electronics_engineering": {
109
- "acc,none": 0.271,
110
- "acc_stderr,none": 0.014062601350986182,
111
- "acc_norm,none": 0.271,
112
- "acc_norm_stderr,none": 0.014062601350986182,
113
  "alias": " - kmmlu_electronics_engineering"
114
  },
115
  "kmmlu_energy_management": {
116
- "acc,none": 0.275,
117
- "acc_stderr,none": 0.014127086556490524,
118
- "acc_norm,none": 0.275,
119
- "acc_norm_stderr,none": 0.014127086556490524,
120
  "alias": " - kmmlu_energy_management"
121
  },
122
  "kmmlu_environmental_science": {
123
- "acc,none": 0.3,
124
- "acc_stderr,none": 0.014498627873361428,
125
- "acc_norm,none": 0.3,
126
- "acc_norm_stderr,none": 0.014498627873361428,
127
  "alias": " - kmmlu_environmental_science"
128
  },
129
  "kmmlu_fashion": {
130
- "acc,none": 0.261,
131
- "acc_stderr,none": 0.01389503767796512,
132
- "acc_norm,none": 0.261,
133
- "acc_norm_stderr,none": 0.01389503767796512,
134
  "alias": " - kmmlu_fashion"
135
  },
136
  "kmmlu_food_processing": {
137
- "acc,none": 0.239,
138
- "acc_stderr,none": 0.013493000446937587,
139
- "acc_norm,none": 0.239,
140
- "acc_norm_stderr,none": 0.013493000446937587,
141
  "alias": " - kmmlu_food_processing"
142
  },
143
  "kmmlu_gas_technology_and_engineering": {
144
- "acc,none": 0.268,
145
- "acc_stderr,none": 0.014013292702729482,
146
- "acc_norm,none": 0.268,
147
- "acc_norm_stderr,none": 0.014013292702729482,
148
  "alias": " - kmmlu_gas_technology_and_engineering"
149
  },
150
  "kmmlu_geomatics": {
151
- "acc,none": 0.264,
152
- "acc_stderr,none": 0.013946271849440469,
153
- "acc_norm,none": 0.264,
154
- "acc_norm_stderr,none": 0.013946271849440469,
155
  "alias": " - kmmlu_geomatics"
156
  },
157
  "kmmlu_health": {
@@ -162,80 +162,80 @@
162
  "alias": " - kmmlu_health"
163
  },
164
  "kmmlu_industrial_engineer": {
165
- "acc,none": 0.276,
166
- "acc_stderr,none": 0.014142984975740671,
167
- "acc_norm,none": 0.276,
168
- "acc_norm_stderr,none": 0.014142984975740671,
169
  "alias": " - kmmlu_industrial_engineer"
170
  },
171
  "kmmlu_information_technology": {
172
- "acc,none": 0.271,
173
- "acc_stderr,none": 0.014062601350986186,
174
- "acc_norm,none": 0.271,
175
- "acc_norm_stderr,none": 0.014062601350986186,
176
  "alias": " - kmmlu_information_technology"
177
  },
178
  "kmmlu_interior_architecture_and_design": {
179
- "acc,none": 0.29,
180
- "acc_stderr,none": 0.014356395999905694,
181
- "acc_norm,none": 0.29,
182
- "acc_norm_stderr,none": 0.014356395999905694,
183
  "alias": " - kmmlu_interior_architecture_and_design"
184
  },
185
  "kmmlu_law": {
186
- "acc,none": 0.257,
187
- "acc_stderr,none": 0.013825416526895031,
188
- "acc_norm,none": 0.257,
189
- "acc_norm_stderr,none": 0.013825416526895031,
190
  "alias": " - kmmlu_law"
191
  },
192
  "kmmlu_machine_design_and_manufacturing": {
193
- "acc,none": 0.254,
194
- "acc_stderr,none": 0.013772206565168543,
195
- "acc_norm,none": 0.254,
196
- "acc_norm_stderr,none": 0.013772206565168543,
197
  "alias": " - kmmlu_machine_design_and_manufacturing"
198
  },
199
  "kmmlu_management": {
200
- "acc,none": 0.274,
201
- "acc_stderr,none": 0.01411109928825958,
202
- "acc_norm,none": 0.274,
203
- "acc_norm_stderr,none": 0.01411109928825958,
204
  "alias": " - kmmlu_management"
205
  },
206
  "kmmlu_maritime_engineering": {
207
  "acc,none": 0.23666666666666666,
208
- "acc_stderr,none": 0.01736649795856464,
209
  "acc_norm,none": 0.23666666666666666,
210
- "acc_norm_stderr,none": 0.01736649795856464,
211
  "alias": " - kmmlu_maritime_engineering"
212
  },
213
  "kmmlu_marketing": {
214
- "acc,none": 0.293,
215
- "acc_stderr,none": 0.014399942998441271,
216
- "acc_norm,none": 0.293,
217
- "acc_norm_stderr,none": 0.014399942998441271,
218
  "alias": " - kmmlu_marketing"
219
  },
220
  "kmmlu_materials_engineering": {
221
- "acc,none": 0.234,
222
- "acc_stderr,none": 0.01339490288966001,
223
- "acc_norm,none": 0.234,
224
- "acc_norm_stderr,none": 0.01339490288966001,
225
  "alias": " - kmmlu_materials_engineering"
226
  },
227
  "kmmlu_mechanical_engineering": {
228
  "acc,none": 0.269,
229
- "acc_stderr,none": 0.014029819522568193,
230
  "acc_norm,none": 0.269,
231
- "acc_norm_stderr,none": 0.014029819522568193,
232
  "alias": " - kmmlu_mechanical_engineering"
233
  },
234
  "kmmlu_nondestructive_testing": {
235
  "acc,none": 0.229,
236
- "acc_stderr,none": 0.013294199326613595,
237
  "acc_norm,none": 0.229,
238
- "acc_norm_stderr,none": 0.013294199326613595,
239
  "alias": " - kmmlu_nondestructive_testing"
240
  },
241
  "kmmlu_patent": {
@@ -253,68 +253,68 @@
253
  "alias": " - kmmlu_political_science_and_sociology"
254
  },
255
  "kmmlu_psychology": {
256
- "acc,none": 0.234,
257
- "acc_stderr,none": 0.01339490288966001,
258
- "acc_norm,none": 0.234,
259
- "acc_norm_stderr,none": 0.01339490288966001,
260
  "alias": " - kmmlu_psychology"
261
  },
262
  "kmmlu_public_safety": {
263
- "acc,none": 0.301,
264
- "acc_stderr,none": 0.014512395033543159,
265
- "acc_norm,none": 0.301,
266
- "acc_norm_stderr,none": 0.014512395033543159,
267
  "alias": " - kmmlu_public_safety"
268
  },
269
  "kmmlu_railway_and_automotive_engineering": {
270
- "acc,none": 0.25,
271
- "acc_stderr,none": 0.013699915608779773,
272
- "acc_norm,none": 0.25,
273
- "acc_norm_stderr,none": 0.013699915608779773,
274
  "alias": " - kmmlu_railway_and_automotive_engineering"
275
  },
276
  "kmmlu_real_estate": {
277
- "acc,none": 0.275,
278
- "acc_stderr,none": 0.031652557907861936,
279
- "acc_norm,none": 0.275,
280
- "acc_norm_stderr,none": 0.031652557907861936,
281
  "alias": " - kmmlu_real_estate"
282
  },
283
  "kmmlu_refrigerating_machinery": {
284
- "acc,none": 0.263,
285
- "acc_stderr,none": 0.013929286594259748,
286
- "acc_norm,none": 0.263,
287
- "acc_norm_stderr,none": 0.013929286594259748,
288
  "alias": " - kmmlu_refrigerating_machinery"
289
  },
290
  "kmmlu_social_welfare": {
291
- "acc,none": 0.258,
292
- "acc_stderr,none": 0.013842963108656604,
293
- "acc_norm,none": 0.258,
294
- "acc_norm_stderr,none": 0.013842963108656604,
295
  "alias": " - kmmlu_social_welfare"
296
  },
297
  "kmmlu_taxation": {
298
- "acc,none": 0.295,
299
- "acc_stderr,none": 0.03232801420614269,
300
- "acc_norm,none": 0.295,
301
- "acc_norm_stderr,none": 0.03232801420614269,
302
  "alias": " - kmmlu_taxation"
303
  },
304
  "kmmlu_telecommunications_and_wireless_technology": {
305
- "acc,none": 0.282,
306
- "acc_stderr,none": 0.014236526215291341,
307
- "acc_norm,none": 0.282,
308
- "acc_norm_stderr,none": 0.014236526215291341,
309
  "alias": " - kmmlu_telecommunications_and_wireless_technology"
310
  }
311
  },
312
  "groups": {
313
  "kmmlu": {
314
- "acc,none": 0.26936182500721906,
315
- "acc_stderr,none": 0.024299898599051413,
316
- "acc_norm,none": 0.26936182500721906,
317
- "acc_norm_stderr,none": 0.024299898599051413,
318
  "alias": "kmmlu"
319
  }
320
  },
@@ -2094,7 +2094,7 @@
2094
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
2095
  "batch_size": "auto",
2096
  "batch_sizes": [
2097
- 4
2098
  ],
2099
  "device": null,
2100
  "use_cache": null,
@@ -2102,5 +2102,5 @@
2102
  "bootstrap_iters": 100000,
2103
  "gen_kwargs": null
2104
  },
2105
- "git_hash": "2e3ceb0"
2106
  }
 
1
  {
2
  "results": {
3
  "kmmlu": {
4
+ "acc,none": 0.26988160554432566,
5
+ "acc_stderr,none": 0.02356571648698782,
6
+ "acc_norm,none": 0.26988160554432566,
7
+ "acc_norm_stderr,none": 0.02356571648698782,
8
  "alias": "kmmlu"
9
  },
10
  "kmmlu_accounting": {
11
+ "acc,none": 0.3,
12
+ "acc_stderr,none": 0.046056618647183814,
13
+ "acc_norm,none": 0.3,
14
+ "acc_norm_stderr,none": 0.046056618647183814,
15
  "alias": " - kmmlu_accounting"
16
  },
17
  "kmmlu_agricultural_sciences": {
18
+ "acc,none": 0.256,
19
+ "acc_stderr,none": 0.013807775152234185,
20
+ "acc_norm,none": 0.256,
21
+ "acc_norm_stderr,none": 0.013807775152234185,
22
  "alias": " - kmmlu_agricultural_sciences"
23
  },
24
  "kmmlu_aviation_engineering_and_maintenance": {
25
+ "acc,none": 0.271,
26
+ "acc_stderr,none": 0.014062601350986186,
27
+ "acc_norm,none": 0.271,
28
+ "acc_norm_stderr,none": 0.014062601350986186,
29
  "alias": " - kmmlu_aviation_engineering_and_maintenance"
30
  },
31
  "kmmlu_biology": {
32
+ "acc,none": 0.257,
33
+ "acc_stderr,none": 0.01382541652689504,
34
+ "acc_norm,none": 0.257,
35
+ "acc_norm_stderr,none": 0.01382541652689504,
36
  "alias": " - kmmlu_biology"
37
  },
38
  "kmmlu_chemical_engineering": {
39
+ "acc,none": 0.276,
40
+ "acc_stderr,none": 0.014142984975740671,
41
+ "acc_norm,none": 0.276,
42
+ "acc_norm_stderr,none": 0.014142984975740671,
43
  "alias": " - kmmlu_chemical_engineering"
44
  },
45
  "kmmlu_chemistry": {
46
+ "acc,none": 0.27166666666666667,
47
+ "acc_stderr,none": 0.018174809149686416,
48
+ "acc_norm,none": 0.27166666666666667,
49
+ "acc_norm_stderr,none": 0.018174809149686416,
50
  "alias": " - kmmlu_chemistry"
51
  },
52
  "kmmlu_civil_engineering": {
53
+ "acc,none": 0.308,
54
+ "acc_stderr,none": 0.01460648312734276,
55
+ "acc_norm,none": 0.308,
56
+ "acc_norm_stderr,none": 0.01460648312734276,
57
  "alias": " - kmmlu_civil_engineering"
58
  },
59
  "kmmlu_computer_science": {
60
+ "acc,none": 0.282,
61
+ "acc_stderr,none": 0.01423652621529135,
62
+ "acc_norm,none": 0.282,
63
+ "acc_norm_stderr,none": 0.01423652621529135,
64
  "alias": " - kmmlu_computer_science"
65
  },
66
  "kmmlu_construction": {
67
+ "acc,none": 0.283,
68
+ "acc_stderr,none": 0.014251810906481744,
69
+ "acc_norm,none": 0.283,
70
+ "acc_norm_stderr,none": 0.014251810906481744,
71
  "alias": " - kmmlu_construction"
72
  },
73
  "kmmlu_criminal_law": {
74
+ "acc,none": 0.265,
75
+ "acc_stderr,none": 0.03128528159088722,
76
+ "acc_norm,none": 0.265,
77
+ "acc_norm_stderr,none": 0.03128528159088722,
78
  "alias": " - kmmlu_criminal_law"
79
  },
80
  "kmmlu_ecology": {
81
+ "acc,none": 0.27,
82
+ "acc_stderr,none": 0.014046255632633918,
83
+ "acc_norm,none": 0.27,
84
+ "acc_norm_stderr,none": 0.014046255632633918,
85
  "alias": " - kmmlu_ecology"
86
  },
87
  "kmmlu_economics": {
 
92
  "alias": " - kmmlu_economics"
93
  },
94
  "kmmlu_education": {
95
+ "acc,none": 0.24,
96
+ "acc_stderr,none": 0.04292346959909282,
97
+ "acc_norm,none": 0.24,
98
+ "acc_norm_stderr,none": 0.04292346959909282,
99
  "alias": " - kmmlu_education"
100
  },
101
  "kmmlu_electrical_engineering": {
102
+ "acc,none": 0.298,
103
+ "acc_stderr,none": 0.014470846741134717,
104
+ "acc_norm,none": 0.298,
105
+ "acc_norm_stderr,none": 0.014470846741134717,
106
  "alias": " - kmmlu_electrical_engineering"
107
  },
108
  "kmmlu_electronics_engineering": {
109
+ "acc,none": 0.269,
110
+ "acc_stderr,none": 0.014029819522568196,
111
+ "acc_norm,none": 0.269,
112
+ "acc_norm_stderr,none": 0.014029819522568196,
113
  "alias": " - kmmlu_electronics_engineering"
114
  },
115
  "kmmlu_energy_management": {
116
+ "acc,none": 0.274,
117
+ "acc_stderr,none": 0.014111099288259588,
118
+ "acc_norm,none": 0.274,
119
+ "acc_norm_stderr,none": 0.014111099288259588,
120
  "alias": " - kmmlu_energy_management"
121
  },
122
  "kmmlu_environmental_science": {
123
+ "acc,none": 0.301,
124
+ "acc_stderr,none": 0.014512395033543152,
125
+ "acc_norm,none": 0.301,
126
+ "acc_norm_stderr,none": 0.014512395033543152,
127
  "alias": " - kmmlu_environmental_science"
128
  },
129
  "kmmlu_fashion": {
130
+ "acc,none": 0.264,
131
+ "acc_stderr,none": 0.01394627184944046,
132
+ "acc_norm,none": 0.264,
133
+ "acc_norm_stderr,none": 0.01394627184944046,
134
  "alias": " - kmmlu_fashion"
135
  },
136
  "kmmlu_food_processing": {
137
+ "acc,none": 0.243,
138
+ "acc_stderr,none": 0.013569640199177451,
139
+ "acc_norm,none": 0.243,
140
+ "acc_norm_stderr,none": 0.013569640199177451,
141
  "alias": " - kmmlu_food_processing"
142
  },
143
  "kmmlu_gas_technology_and_engineering": {
144
+ "acc,none": 0.276,
145
+ "acc_stderr,none": 0.01414298497574067,
146
+ "acc_norm,none": 0.276,
147
+ "acc_norm_stderr,none": 0.01414298497574067,
148
  "alias": " - kmmlu_gas_technology_and_engineering"
149
  },
150
  "kmmlu_geomatics": {
151
+ "acc,none": 0.268,
152
+ "acc_stderr,none": 0.014013292702729494,
153
+ "acc_norm,none": 0.268,
154
+ "acc_norm_stderr,none": 0.014013292702729494,
155
  "alias": " - kmmlu_geomatics"
156
  },
157
  "kmmlu_health": {
 
162
  "alias": " - kmmlu_health"
163
  },
164
  "kmmlu_industrial_engineer": {
165
+ "acc,none": 0.278,
166
+ "acc_stderr,none": 0.014174516461485242,
167
+ "acc_norm,none": 0.278,
168
+ "acc_norm_stderr,none": 0.014174516461485242,
169
  "alias": " - kmmlu_industrial_engineer"
170
  },
171
  "kmmlu_information_technology": {
172
+ "acc,none": 0.266,
173
+ "acc_stderr,none": 0.013979965645145153,
174
+ "acc_norm,none": 0.266,
175
+ "acc_norm_stderr,none": 0.013979965645145153,
176
  "alias": " - kmmlu_information_technology"
177
  },
178
  "kmmlu_interior_architecture_and_design": {
179
+ "acc,none": 0.291,
180
+ "acc_stderr,none": 0.014370995982377932,
181
+ "acc_norm,none": 0.291,
182
+ "acc_norm_stderr,none": 0.014370995982377932,
183
  "alias": " - kmmlu_interior_architecture_and_design"
184
  },
185
  "kmmlu_law": {
186
+ "acc,none": 0.253,
187
+ "acc_stderr,none": 0.01375427861358708,
188
+ "acc_norm,none": 0.253,
189
+ "acc_norm_stderr,none": 0.01375427861358708,
190
  "alias": " - kmmlu_law"
191
  },
192
  "kmmlu_machine_design_and_manufacturing": {
193
+ "acc,none": 0.261,
194
+ "acc_stderr,none": 0.01389503767796513,
195
+ "acc_norm,none": 0.261,
196
+ "acc_norm_stderr,none": 0.01389503767796513,
197
  "alias": " - kmmlu_machine_design_and_manufacturing"
198
  },
199
  "kmmlu_management": {
200
+ "acc,none": 0.277,
201
+ "acc_stderr,none": 0.014158794845306265,
202
+ "acc_norm,none": 0.277,
203
+ "acc_norm_stderr,none": 0.014158794845306265,
204
  "alias": " - kmmlu_management"
205
  },
206
  "kmmlu_maritime_engineering": {
207
  "acc,none": 0.23666666666666666,
208
+ "acc_stderr,none": 0.017366497958564646,
209
  "acc_norm,none": 0.23666666666666666,
210
+ "acc_norm_stderr,none": 0.017366497958564646,
211
  "alias": " - kmmlu_maritime_engineering"
212
  },
213
  "kmmlu_marketing": {
214
+ "acc,none": 0.296,
215
+ "acc_stderr,none": 0.014442734941575018,
216
+ "acc_norm,none": 0.296,
217
+ "acc_norm_stderr,none": 0.014442734941575018,
218
  "alias": " - kmmlu_marketing"
219
  },
220
  "kmmlu_materials_engineering": {
221
+ "acc,none": 0.241,
222
+ "acc_stderr,none": 0.013531522534515433,
223
+ "acc_norm,none": 0.241,
224
+ "acc_norm_stderr,none": 0.013531522534515433,
225
  "alias": " - kmmlu_materials_engineering"
226
  },
227
  "kmmlu_mechanical_engineering": {
228
  "acc,none": 0.269,
229
+ "acc_stderr,none": 0.014029819522568198,
230
  "acc_norm,none": 0.269,
231
+ "acc_norm_stderr,none": 0.014029819522568198,
232
  "alias": " - kmmlu_mechanical_engineering"
233
  },
234
  "kmmlu_nondestructive_testing": {
235
  "acc,none": 0.229,
236
+ "acc_stderr,none": 0.013294199326613597,
237
  "acc_norm,none": 0.229,
238
+ "acc_norm_stderr,none": 0.013294199326613597,
239
  "alias": " - kmmlu_nondestructive_testing"
240
  },
241
  "kmmlu_patent": {
 
253
  "alias": " - kmmlu_political_science_and_sociology"
254
  },
255
  "kmmlu_psychology": {
256
+ "acc,none": 0.237,
257
+ "acc_stderr,none": 0.013454070462577941,
258
+ "acc_norm,none": 0.237,
259
+ "acc_norm_stderr,none": 0.013454070462577941,
260
  "alias": " - kmmlu_psychology"
261
  },
262
  "kmmlu_public_safety": {
263
+ "acc,none": 0.3,
264
+ "acc_stderr,none": 0.014498627873361427,
265
+ "acc_norm,none": 0.3,
266
+ "acc_norm_stderr,none": 0.014498627873361427,
267
  "alias": " - kmmlu_public_safety"
268
  },
269
  "kmmlu_railway_and_automotive_engineering": {
270
+ "acc,none": 0.254,
271
+ "acc_stderr,none": 0.01377220656516854,
272
+ "acc_norm,none": 0.254,
273
+ "acc_norm_stderr,none": 0.01377220656516854,
274
  "alias": " - kmmlu_railway_and_automotive_engineering"
275
  },
276
  "kmmlu_real_estate": {
277
+ "acc,none": 0.28,
278
+ "acc_stderr,none": 0.031828687164775826,
279
+ "acc_norm,none": 0.28,
280
+ "acc_norm_stderr,none": 0.031828687164775826,
281
  "alias": " - kmmlu_real_estate"
282
  },
283
  "kmmlu_refrigerating_machinery": {
284
+ "acc,none": 0.261,
285
+ "acc_stderr,none": 0.01389503767796512,
286
+ "acc_norm,none": 0.261,
287
+ "acc_norm_stderr,none": 0.01389503767796512,
288
  "alias": " - kmmlu_refrigerating_machinery"
289
  },
290
  "kmmlu_social_welfare": {
291
+ "acc,none": 0.257,
292
+ "acc_stderr,none": 0.013825416526895036,
293
+ "acc_norm,none": 0.257,
294
+ "acc_norm_stderr,none": 0.013825416526895036,
295
  "alias": " - kmmlu_social_welfare"
296
  },
297
  "kmmlu_taxation": {
298
+ "acc,none": 0.29,
299
+ "acc_stderr,none": 0.03216633903375033,
300
+ "acc_norm,none": 0.29,
301
+ "acc_norm_stderr,none": 0.03216633903375033,
302
  "alias": " - kmmlu_taxation"
303
  },
304
  "kmmlu_telecommunications_and_wireless_technology": {
305
+ "acc,none": 0.283,
306
+ "acc_stderr,none": 0.014251810906481765,
307
+ "acc_norm,none": 0.283,
308
+ "acc_norm_stderr,none": 0.014251810906481765,
309
  "alias": " - kmmlu_telecommunications_and_wireless_technology"
310
  }
311
  },
312
  "groups": {
313
  "kmmlu": {
314
+ "acc,none": 0.26988160554432566,
315
+ "acc_stderr,none": 0.02356571648698782,
316
+ "acc_norm,none": 0.26988160554432566,
317
+ "acc_norm_stderr,none": 0.02356571648698782,
318
  "alias": "kmmlu"
319
  }
320
  },
 
2094
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
2095
  "batch_size": "auto",
2096
  "batch_sizes": [
2097
+ 16
2098
  ],
2099
  "device": null,
2100
  "use_cache": null,
 
2102
  "bootstrap_iters": 100000,
2103
  "gen_kwargs": null
2104
  },
2105
+ "git_hash": "4d19ea9"
2106
  }
lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8cefbf4f0873f2ef42cefc718c36a09c0c63d3e4b7431485ef4846dadf6694b
3
- size 208768
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:322c52059ccde2c052fcb1d4a0e5ff80e358628bf38e0e0ac3a949566296d671
3
+ size 595303
lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,47 +1,47 @@
1
  {
2
  "results": {
3
  "kobest": {
4
- "acc,none": 0.48629686472264855,
5
- "acc_stderr,none": 0.03845705102970606,
6
- "f1,none": 0.3941498184514018,
7
  "f1_stderr,none": "N/A",
8
- "acc_norm,none": 0.464,
9
- "acc_norm_stderr,none": 0.0004984048096192345,
10
  "alias": "kobest"
11
  },
12
  "kobest_boolq": {
13
- "acc,none": 0.5042735042735043,
14
- "acc_stderr,none": 0.013348279916769821,
15
- "f1,none": 0.35467032967032963,
16
  "f1_stderr,none": "N/A",
17
  "alias": " - kobest_boolq"
18
  },
19
  "kobest_copa": {
20
- "acc,none": 0.524,
21
- "acc_stderr,none": 0.015801065586651758,
22
- "f1,none": 0.5229906482116158,
23
  "f1_stderr,none": "N/A",
24
  "alias": " - kobest_copa"
25
  },
26
  "kobest_hellaswag": {
27
- "acc,none": 0.37,
28
- "acc_stderr,none": 0.021613289165165785,
29
- "f1,none": 0.3655287187202081,
30
  "f1_stderr,none": "N/A",
31
- "acc_norm,none": 0.464,
32
- "acc_norm_stderr,none": 0.022324981738385243,
33
  "alias": " - kobest_hellaswag"
34
  },
35
  "kobest_sentineg": {
36
- "acc,none": 0.46851385390428213,
37
- "acc_stderr,none": 0.025076077305681316,
38
- "f1,none": 0.4552271323122947,
39
  "f1_stderr,none": "N/A",
40
  "alias": " - kobest_sentineg"
41
  },
42
  "kobest_wic": {
43
  "acc,none": 0.4880952380952381,
44
- "acc_stderr,none": 0.014087502464604053,
45
  "f1,none": 0.328,
46
  "f1_stderr,none": "N/A",
47
  "alias": " - kobest_wic"
@@ -49,12 +49,12 @@
49
  },
50
  "groups": {
51
  "kobest": {
52
- "acc,none": 0.48629686472264855,
53
- "acc_stderr,none": 0.03845705102970606,
54
- "f1,none": 0.3941498184514018,
55
  "f1_stderr,none": "N/A",
56
- "acc_norm,none": 0.464,
57
- "acc_norm_stderr,none": 0.0004984048096192345,
58
  "alias": "kobest"
59
  }
60
  },
@@ -281,7 +281,7 @@
281
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
282
  "batch_size": "auto",
283
  "batch_sizes": [
284
- 8
285
  ],
286
  "device": null,
287
  "use_cache": null,
@@ -289,5 +289,5 @@
289
  "bootstrap_iters": 100000,
290
  "gen_kwargs": null
291
  },
292
- "git_hash": "2e3ceb0"
293
  }
 
1
  {
2
  "results": {
3
  "kobest": {
4
+ "acc,none": 0.48541986406489807,
5
+ "acc_stderr,none": 0.0366339172650086,
6
+ "f1,none": 0.39393512492621,
7
  "f1_stderr,none": "N/A",
8
+ "acc_norm,none": 0.462,
9
+ "acc_norm_stderr,none": 0.0004981082164328657,
10
  "alias": "kobest"
11
  },
12
  "kobest_boolq": {
13
+ "acc,none": 0.5035612535612536,
14
+ "acc_stderr,none": 0.013348428901951027,
15
+ "f1,none": 0.3554657321046679,
16
  "f1_stderr,none": "N/A",
17
  "alias": " - kobest_boolq"
18
  },
19
  "kobest_copa": {
20
+ "acc,none": 0.522,
21
+ "acc_stderr,none": 0.015803979428161946,
22
+ "f1,none": 0.520896144717429,
23
  "f1_stderr,none": "N/A",
24
  "alias": " - kobest_copa"
25
  },
26
  "kobest_hellaswag": {
27
+ "acc,none": 0.366,
28
+ "acc_stderr,none": 0.021564276850201618,
29
+ "f1,none": 0.3618163879785442,
30
  "f1_stderr,none": "N/A",
31
+ "acc_norm,none": 0.462,
32
+ "acc_norm_stderr,none": 0.022318338119870523,
33
  "alias": " - kobest_hellaswag"
34
  },
35
  "kobest_sentineg": {
36
+ "acc,none": 0.47103274559193953,
37
+ "acc_stderr,none": 0.025083743486632542,
38
+ "f1,none": 0.4598989375485877,
39
  "f1_stderr,none": "N/A",
40
  "alias": " - kobest_sentineg"
41
  },
42
  "kobest_wic": {
43
  "acc,none": 0.4880952380952381,
44
+ "acc_stderr,none": 0.014087502464604038,
45
  "f1,none": 0.328,
46
  "f1_stderr,none": "N/A",
47
  "alias": " - kobest_wic"
 
49
  },
50
  "groups": {
51
  "kobest": {
52
+ "acc,none": 0.48541986406489807,
53
+ "acc_stderr,none": 0.0366339172650086,
54
+ "f1,none": 0.39393512492621,
55
  "f1_stderr,none": "N/A",
56
+ "acc_norm,none": 0.462,
57
+ "acc_norm_stderr,none": 0.0004981082164328657,
58
  "alias": "kobest"
59
  }
60
  },
 
281
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
282
  "batch_size": "auto",
283
  "batch_sizes": [
284
+ 32
285
  ],
286
  "device": null,
287
  "use_cache": null,
 
289
  "bootstrap_iters": 100000,
290
  "gen_kwargs": null
291
  },
292
+ "git_hash": "4d19ea9"
293
  }
lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da74f1796d0de08111e8b232ced3122134f5e53a0dcd6fe1bf87e24e4228bf65
3
- size 38221
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c1bc8ee74f17dcec203e1ae47221a328fb041afedef6fc8aaa043899d37b0eb
3
+ size 24630
lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,33 +1,33 @@
1
  {
2
  "results": {
3
  "lambada": {
4
- "perplexity,none": 4.581086434995066,
5
- "perplexity_stderr,none": 0.24535209996272006,
6
- "acc,none": 0.6639821463225306,
7
- "acc_stderr,none": 0.014008914961907022,
8
  "alias": "lambada"
9
  },
10
  "lambada_openai": {
11
- "perplexity,none": 4.132928377411266,
12
- "perplexity_stderr,none": 0.0870725561065737,
13
- "acc,none": 0.6887250145546284,
14
- "acc_stderr,none": 0.006450703968778299,
15
  "alias": " - lambada_openai"
16
  },
17
  "lambada_standard": {
18
- "perplexity,none": 5.029244492578866,
19
- "perplexity_stderr,none": 0.11127913513447098,
20
- "acc,none": 0.6392392780904328,
21
- "acc_stderr,none": 0.006690420625907091,
22
  "alias": " - lambada_standard"
23
  }
24
  },
25
  "groups": {
26
  "lambada": {
27
- "perplexity,none": 4.581086434995066,
28
- "perplexity_stderr,none": 0.24535209996272006,
29
- "acc,none": 0.6639821463225306,
30
- "acc_stderr,none": 0.014008914961907022,
31
  "alias": "lambada"
32
  }
33
  },
@@ -114,7 +114,7 @@
114
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
115
  "batch_size": "auto",
116
  "batch_sizes": [
117
- 32
118
  ],
119
  "device": null,
120
  "use_cache": null,
@@ -122,5 +122,5 @@
122
  "bootstrap_iters": 100000,
123
  "gen_kwargs": null
124
  },
125
- "git_hash": "2e3ceb0"
126
  }
 
1
  {
2
  "results": {
3
  "lambada": {
4
+ "perplexity,none": 4.581309477921781,
5
+ "perplexity_stderr,none": 0.24450204428335026,
6
+ "acc,none": 0.6634969920434698,
7
+ "acc_stderr,none": 0.013626250353494346,
8
  "alias": "lambada"
9
  },
10
  "lambada_openai": {
11
+ "perplexity,none": 4.135508409053762,
12
+ "perplexity_stderr,none": 0.08754367372026851,
13
+ "acc,none": 0.6873665825732583,
14
+ "acc_stderr,none": 0.006458385716767283,
15
  "alias": " - lambada_openai"
16
  },
17
  "lambada_standard": {
18
+ "perplexity,none": 5.0271105467898005,
19
+ "perplexity_stderr,none": 0.11190006251917957,
20
+ "acc,none": 0.6396274015136814,
21
+ "acc_stderr,none": 0.006688850414338584,
22
  "alias": " - lambada_standard"
23
  }
24
  },
25
  "groups": {
26
  "lambada": {
27
+ "perplexity,none": 4.581309477921781,
28
+ "perplexity_stderr,none": 0.24450204428335026,
29
+ "acc,none": 0.6634969920434698,
30
+ "acc_stderr,none": 0.013626250353494346,
31
  "alias": "lambada"
32
  }
33
  },
 
114
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
115
  "batch_size": "auto",
116
  "batch_sizes": [
117
+ 64
118
  ],
119
  "device": null,
120
  "use_cache": null,
 
122
  "bootstrap_iters": 100000,
123
  "gen_kwargs": null
124
  },
125
+ "git_hash": "4d19ea9"
126
  }
lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6bb2ecfc031d37c3cf86cd8f928d40154cad20956535afca696e2a64a6195c2
3
- size 26754
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8095a18253d00ff62fad815caf102d0cf54f3362e5f6c690c70cc73c1b60aaa3
3
+ size 14892
lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,33 +1,33 @@
1
  {
2
  "results": {
3
  "lambada_cloze": {
4
- "perplexity,none": 202.60084862007227,
5
- "perplexity_stderr,none": 6.427806219106391,
6
- "acc,none": 0.09411993013778382,
7
- "acc_stderr,none": 0.008590078218125246,
8
  "alias": "lambada_cloze"
9
  },
10
  "lambada_openai_cloze_yaml": {
11
- "perplexity,none": 201.1914466369863,
12
- "perplexity_stderr,none": 6.506244049635805,
13
- "acc,none": 0.07898311663108869,
14
- "acc_stderr,none": 0.0037576212389559436,
15
  "alias": " - lambada_openai_cloze_yaml"
16
  },
17
  "lambada_standard_cloze_yaml": {
18
- "perplexity,none": 204.01025060315825,
19
- "perplexity_stderr,none": 6.270310497569642,
20
  "acc,none": 0.10925674364447895,
21
- "acc_stderr,none": 0.004346227651722467,
22
  "alias": " - lambada_standard_cloze_yaml"
23
  }
24
  },
25
  "groups": {
26
  "lambada_cloze": {
27
- "perplexity,none": 202.60084862007227,
28
- "perplexity_stderr,none": 6.427806219106391,
29
- "acc,none": 0.09411993013778382,
30
- "acc_stderr,none": 0.008590078218125246,
31
  "alias": "lambada_cloze"
32
  }
33
  },
@@ -114,7 +114,7 @@
114
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
115
  "batch_size": "auto",
116
  "batch_sizes": [
117
- 32
118
  ],
119
  "device": null,
120
  "use_cache": null,
@@ -122,5 +122,5 @@
122
  "bootstrap_iters": 100000,
123
  "gen_kwargs": null
124
  },
125
- "git_hash": "2e3ceb0"
126
  }
 
1
  {
2
  "results": {
3
  "lambada_cloze": {
4
+ "perplexity,none": 202.80427845102082,
5
+ "perplexity_stderr,none": 6.42809109784278,
6
+ "acc,none": 0.09441102270522025,
7
+ "acc_stderr,none": 0.008464918604058229,
8
  "alias": "lambada_cloze"
9
  },
10
  "lambada_openai_cloze_yaml": {
11
+ "perplexity,none": 201.40675886932158,
12
+ "perplexity_stderr,none": 6.518722997817119,
13
+ "acc,none": 0.07956530176596158,
14
+ "acc_stderr,none": 0.0037702523650452176,
15
  "alias": " - lambada_openai_cloze_yaml"
16
  },
17
  "lambada_standard_cloze_yaml": {
18
+ "perplexity,none": 204.20179803272006,
19
+ "perplexity_stderr,none": 6.259254030853473,
20
  "acc,none": 0.10925674364447895,
21
+ "acc_stderr,none": 0.004346227651722471,
22
  "alias": " - lambada_standard_cloze_yaml"
23
  }
24
  },
25
  "groups": {
26
  "lambada_cloze": {
27
+ "perplexity,none": 202.80427845102082,
28
+ "perplexity_stderr,none": 6.42809109784278,
29
+ "acc,none": 0.09441102270522025,
30
+ "acc_stderr,none": 0.008464918604058229,
31
  "alias": "lambada_cloze"
32
  }
33
  },
 
114
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
115
  "batch_size": "auto",
116
  "batch_sizes": [
117
+ 64
118
  ],
119
  "device": null,
120
  "use_cache": null,
 
122
  "bootstrap_iters": 100000,
123
  "gen_kwargs": null
124
  },
125
+ "git_hash": "4d19ea9"
126
  }
lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13a1ea7ff27a744f6832ce754169e7999a544270af382a89c145bee14e266eaa
3
- size 19729
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0215396b73b698ba81d69dd365b50d1311395881a989f4b1a1113be8721dd79c
3
+ size 15076
lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "results": {
3
  "logieval": {
4
- "exact_match,get-answer": 0.2455470737913486,
5
- "exact_match_stderr,get-answer": 0.010859138259206532,
6
  "alias": "logieval"
7
  }
8
  },
@@ -71,5 +71,5 @@
71
  "bootstrap_iters": 100000,
72
  "gen_kwargs": null
73
  },
74
- "git_hash": "2e3ceb0"
75
  }
 
1
  {
2
  "results": {
3
  "logieval": {
4
+ "exact_match,get-answer": 0.2684478371501272,
5
+ "exact_match_stderr,get-answer": 0.011180584582096637,
6
  "alias": "logieval"
7
  }
8
  },
 
71
  "bootstrap_iters": 100000,
72
  "gen_kwargs": null
73
  },
74
+ "git_hash": "4d19ea9"
75
  }
lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a1f89f1bd20cb0d9581ebca08b3d7e47091c6fde766a2ed79edbe53b8d80f01
3
- size 22052
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f4f24e3aab87e3f887ba30422f70190388fdd4b4cefce5804d138e2ea3b4d4c
3
+ size 9264
lm-eval-output/allenai/OLMo-7B/logiqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "logiqa": {
4
+ "acc,none": 0.23195084485407066,
5
+ "acc_stderr,none": 0.016555252497925894,
6
+ "acc_norm,none": 0.2749615975422427,
7
+ "acc_norm_stderr,none": 0.01751297178222521,
8
+ "alias": "logiqa"
9
+ }
10
+ },
11
+ "configs": {
12
+ "logiqa": {
13
+ "task": "logiqa",
14
+ "dataset_path": "EleutherAI/logiqa",
15
+ "dataset_name": "logiqa",
16
+ "training_split": "train",
17
+ "validation_split": "validation",
18
+ "test_split": "test",
19
+ "doc_to_text": "def doc_to_text(doc) -> str:\n \"\"\"\n Passage: <passage>\n Question: <question>\n Choices:\n A. <choice1>\n B. <choice2>\n C. <choice3>\n D. <choice4>\n Answer:\n \"\"\"\n choices = [\"a\", \"b\", \"c\", \"d\"]\n prompt = \"Passage: \" + doc[\"context\"] + \"\\n\"\n prompt += \"Question: \" + doc[\"question\"] + \"\\nChoices:\\n\"\n for choice, option in zip(choices, doc[\"options\"]):\n prompt += f\"{choice.upper()}. {option}\\n\"\n prompt += \"Answer:\"\n return prompt\n",
20
+ "doc_to_target": "def doc_to_target(doc) -> int:\n choices = [\"a\", \"b\", \"c\", \"d\"]\n return choices.index(doc[\"label\"].strip())\n",
21
+ "doc_to_choice": "{{options}}",
22
+ "description": "",
23
+ "target_delimiter": " ",
24
+ "fewshot_delimiter": "\n\n",
25
+ "metric_list": [
26
+ {
27
+ "metric": "acc",
28
+ "aggregation": "mean",
29
+ "higher_is_better": true
30
+ },
31
+ {
32
+ "metric": "acc_norm",
33
+ "aggregation": "mean",
34
+ "higher_is_better": true
35
+ }
36
+ ],
37
+ "output_type": "multiple_choice",
38
+ "repeats": 1,
39
+ "should_decontaminate": true,
40
+ "doc_to_decontamination_query": "{{context}}",
41
+ "metadata": {
42
+ "version": 1.0
43
+ }
44
+ }
45
+ },
46
+ "versions": {
47
+ "logiqa": 1.0
48
+ },
49
+ "n-shot": {
50
+ "logiqa": 0
51
+ },
52
+ "config": {
53
+ "model": "hf",
54
+ "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
55
+ "batch_size": "auto",
56
+ "batch_sizes": [
57
+ 32
58
+ ],
59
+ "device": null,
60
+ "use_cache": null,
61
+ "limit": null,
62
+ "bootstrap_iters": 100000,
63
+ "gen_kwargs": null
64
+ },
65
+ "git_hash": "4d19ea9"
66
+ }
lm-eval-output/allenai/OLMo-7B/logiqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6d813aeb02f78d7c002a843f01b29fa2d93e26155de266673b77dff9b133248
3
- size 22189
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8277def6e3e1e896b8bf2f3e655a70b86d7a982f7284b51939940c1f46e93b4
3
+ size 8002
lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
  "results": {
3
  "logiqa2": {
4
- "acc,none": 0.24872773536895673,
5
- "acc_stderr,none": 0.010906180806103564,
6
- "acc_norm,none": 0.2907124681933842,
7
- "acc_norm_stderr,none": 0.011456577557813215,
8
  "alias": "logiqa2"
9
  }
10
  },
@@ -54,7 +54,7 @@
54
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
55
  "batch_size": "auto",
56
  "batch_sizes": [
57
- 8
58
  ],
59
  "device": null,
60
  "use_cache": null,
@@ -62,5 +62,5 @@
62
  "bootstrap_iters": 100000,
63
  "gen_kwargs": null
64
  },
65
- "git_hash": "2e3ceb0"
66
  }
 
1
  {
2
  "results": {
3
  "logiqa2": {
4
+ "acc,none": 0.2506361323155216,
5
+ "acc_stderr,none": 0.010934026494722665,
6
+ "acc_norm,none": 0.2881679389312977,
7
+ "acc_norm_stderr,none": 0.011426770634965258,
8
  "alias": "logiqa2"
9
  }
10
  },
 
54
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
55
  "batch_size": "auto",
56
  "batch_sizes": [
57
+ 32
58
  ],
59
  "device": null,
60
  "use_cache": null,
 
62
  "bootstrap_iters": 100000,
63
  "gen_kwargs": null
64
  },
65
+ "git_hash": "4d19ea9"
66
  }
lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:61f0b5913ee7c74e6d4b779cb1e963f0acccdebf06ba0781ce8c8d25afebd7b4
3
- size 23189
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e71f9d99e62bcc19d25f90d9a3d5547f0d7968f2ca5b60d7a5cfbc70afd441b
3
+ size 15217
lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
  "results": {
3
  "mathqa": {
4
- "acc,none": 0.25896147403685094,
5
- "acc_stderr,none": 0.008019338828219905,
6
- "acc_norm,none": 0.254606365159129,
7
- "acc_norm_stderr,none": 0.007974951653806829,
8
  "alias": "mathqa"
9
  }
10
  },
@@ -56,7 +56,7 @@
56
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
57
  "batch_size": "auto",
58
  "batch_sizes": [
59
- 32
60
  ],
61
  "device": null,
62
  "use_cache": null,
@@ -64,5 +64,5 @@
64
  "bootstrap_iters": 100000,
65
  "gen_kwargs": null
66
  },
67
- "git_hash": "2e3ceb0"
68
  }
 
1
  {
2
  "results": {
3
  "mathqa": {
4
+ "acc,none": 0.25996649916247905,
5
+ "acc_stderr,none": 0.008029434758777935,
6
+ "acc_norm,none": 0.2562814070351759,
7
+ "acc_norm_stderr,none": 0.007992146938217008,
8
  "alias": "mathqa"
9
  }
10
  },
 
56
  "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True",
57
  "batch_size": "auto",
58
  "batch_sizes": [
59
+ 64
60
  ],
61
  "device": null,
62
  "use_cache": null,
 
64
  "bootstrap_iters": 100000,
65
  "gen_kwargs": null
66
  },
67
+ "git_hash": "4d19ea9"
68
  }
lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6da0aab79c838012114e306652dbf9afe0d2d5fc797600d1298d4ffb21a02fa7
3
- size 14214
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86caeec82d12cbdedff681e862549f68df0c055020e299e40836168ce33918c7
3
+ size 18748