dataset
stringlengths
5
115
config
stringlengths
1
162
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
0
77.9k
null_count
int64
0
62.9M
null_proportion
float64
0
1
min
int64
0
9.25M
max
int64
0
1.07B
mean
float64
0
90.4M
median
float64
0
80.1M
std
float64
0
130M
histogram
dict
partial
bool
2 classes
Windy/ultr4_4rewriter_m0_vs_real
default
test
208
prompt
0
0
376
8,142
2,034.625
1,649
1,353.97124
{ "bin_edges": [ 376, 1153, 1930, 2707, 3484, 4261, 5038, 5815, 6592, 7369, 8142 ], "hist": [ 53, 74, 38, 18, 8, 11, 1, 1, 1, 3 ] }
false
Windy/ultr4_4rewriter_m0_vs_real
default
train
10,212
completion
0
0
24
11,137
2,292.50431
2,474
1,364.77852
{ "bin_edges": [ 24, 1136, 2248, 3360, 4472, 5584, 6696, 7808, 8920, 10032, 11137 ], "hist": [ 2674, 1886, 3155, 2132, 320, 32, 8, 1, 1, 3 ] }
false
Windy/ultr4_4rewriter_m0_vs_real
default
train
10,212
prompt
0
0
322
17,552
1,958.67058
1,685
1,244.61622
{ "bin_edges": [ 322, 2046, 3770, 5494, 7218, 8942, 10666, 12390, 14114, 15838, 17552 ], "hist": [ 6616, 2765, 596, 192, 34, 5, 3, 0, 0, 1 ] }
false
lmms-lab/LiveBenchDetailedResults
2024-07
llama3_llava_next_8b
250
criteria
0
0
157
1,709
854.196
834.5
277.88823
{ "bin_edges": [ 157, 313, 469, 625, 781, 937, 1093, 1249, 1405, 1561, 1709 ], "hist": [ 5, 7, 33, 57, 60, 44, 24, 10, 6, 4 ] }
false
lmms-lab/LiveBenchDetailedResults
2024-07
llama3_llava_next_8b
250
ground_truth
0
0
4
2,474
912.112
993.5
599.60142
{ "bin_edges": [ 4, 252, 500, 748, 996, 1244, 1492, 1740, 1988, 2236, 2474 ], "hist": [ 49, 29, 18, 29, 47, 36, 21, 13, 5, 3 ] }
false
lmms-lab/LiveBenchDetailedResults
2024-07
llama3_llava_next_8b
250
question
0
0
70
626
229.736
225
86.50265
{ "bin_edges": [ 70, 126, 182, 238, 294, 350, 406, 462, 518, 574, 626 ], "hist": [ 22, 58, 61, 63, 28, 10, 2, 4, 1, 1 ] }
false
lmms-lab/LiveBenchDetailedResults
2024-07
llama3_llava_next_8b
250
reason
0
0
119
1,270
495.78
513
174.66446
{ "bin_edges": [ 119, 235, 351, 467, 583, 699, 815, 931, 1047, 1163, 1270 ], "hist": [ 24, 26, 46, 80, 52, 14, 5, 2, 0, 1 ] }
false
lmms-lab/LiveBenchDetailedResults
2024-07
llama3_llava_next_8b
250
response
0
0
5
5,450
1,403.708
1,500
881.235
{ "bin_edges": [ 5, 550, 1095, 1640, 2185, 2730, 3275, 3820, 4365, 4910, 5450 ], "hist": [ 63, 31, 49, 54, 41, 10, 0, 1, 0, 1 ] }
false
ssirikon/AESLC_Unsloth_Train
default
train
14,455
input
0
0
12
32,759
785.86095
425
1,873.25893
{ "bin_edges": [ 12, 3287, 6562, 9837, 13112, 16387, 19662, 22937, 26212, 29487, 32759 ], "hist": [ 14130, 210, 39, 25, 6, 3, 1, 0, 22, 19 ] }
false
ssirikon/AESLC_Unsloth_Train
default
train
14,455
output
19
0.00131
2
1,874
29.51025
24
67.33923
{ "bin_edges": [ 2, 190, 378, 566, 754, 942, 1130, 1318, 1506, 1694, 1874 ], "hist": [ 14418, 0, 0, 0, 0, 0, 0, 0, 0, 18 ] }
false
ycfNTU/Animal_llama2_zeroshot
default
test
1,333
exid
0
0
9
12
11.48237
12
0.60332
{ "bin_edges": [ 9, 10, 11, 12, 12 ], "hist": [ 7, 55, 559, 712 ] }
false
ycfNTU/Animal_llama2_zeroshot
default
test
1,333
output
0
0
0
3,572
271.93548
95
431.66047
{ "bin_edges": [ 0, 358, 716, 1074, 1432, 1790, 2148, 2506, 2864, 3222, 3572 ], "hist": [ 986, 192, 87, 29, 18, 8, 7, 3, 1, 2 ] }
false
ycfNTU/Animal_llama2_zeroshot
default
test
1,333
summary
0
0
4
2,529
715.87322
660
388.54467
{ "bin_edges": [ 4, 257, 510, 763, 1016, 1269, 1522, 1775, 2028, 2281, 2529 ], "hist": [ 131, 202, 459, 348, 85, 56, 22, 18, 6, 6 ] }
false
ycfNTU/Animal_llama2_zeroshot
default
test
1,333
text
0
0
0
3,978
2,057.86647
2,264
996.23483
{ "bin_edges": [ 0, 398, 796, 1194, 1592, 1990, 2388, 2786, 3184, 3582, 3978 ], "hist": [ 69, 119, 132, 129, 151, 106, 312, 111, 125, 79 ] }
false
ssirikon/AESLC_Unsloth_Train_Subset
default
train
967
input
0
0
124
28,796
756.212
424
1,710.41347
{ "bin_edges": [ 124, 2992, 5860, 8728, 11596, 14464, 17332, 20200, 23068, 25936, 28796 ], "hist": [ 951, 10, 3, 0, 0, 0, 0, 0, 0, 3 ] }
false
ssirikon/AESLC_Unsloth_Train_Subset
default
train
967
output
0
0
3
109
27.41675
24
17.44963
{ "bin_edges": [ 3, 14, 25, 36, 47, 58, 69, 80, 91, 102, 109 ], "hist": [ 220, 283, 210, 121, 72, 30, 18, 6, 4, 3 ] }
false
Windy/sft4rewriter_m0_vs_real
default
train
10,212
completion
0
0
24
11,137
2,292.50431
2,474
1,364.77852
{ "bin_edges": [ 24, 1136, 2248, 3360, 4472, 5584, 6696, 7808, 8920, 10032, 11137 ], "hist": [ 2674, 1886, 3155, 2132, 320, 32, 8, 1, 1, 3 ] }
false
Windy/sft4rewriter_m0_vs_real
default
train
10,212
prompt
0
0
322
17,552
1,958.67058
1,685
1,244.61622
{ "bin_edges": [ 322, 2046, 3770, 5494, 7218, 8942, 10666, 12390, 14114, 15838, 17552 ], "hist": [ 6616, 2765, 596, 192, 34, 5, 3, 0, 0, 1 ] }
false
Windy/sft4rewriter_m0_vs_real
default
test
208
completion
0
0
37
5,525
2,427.12019
2,656.5
1,377.91341
{ "bin_edges": [ 37, 586, 1135, 1684, 2233, 2782, 3331, 3880, 4429, 4978, 5525 ], "hist": [ 30, 24, 12, 17, 29, 32, 31, 24, 6, 3 ] }
false
Windy/sft4rewriter_m0_vs_real
default
test
208
prompt
0
0
376
8,142
2,034.625
1,649
1,353.97124
{ "bin_edges": [ 376, 1153, 1930, 2707, 3484, 4261, 5038, 5815, 6592, 7369, 8142 ], "hist": [ 53, 74, 38, 18, 8, 11, 1, 1, 1, 3 ] }
false
ycfNTU/Album_llama2_zeroshot
default
test
278
exid
0
0
8
12
11.49281
12
0.62292
{ "bin_edges": [ 8, 9, 10, 11, 12, 12 ], "hist": [ 1, 1, 10, 114, 152 ] }
false
ycfNTU/Album_llama2_zeroshot
default
test
278
output
0
0
0
3,935
597.21583
422.5
597.61626
{ "bin_edges": [ 0, 394, 788, 1182, 1576, 1970, 2364, 2758, 3152, 3546, 3935 ], "hist": [ 133, 63, 38, 22, 13, 5, 2, 0, 1, 1 ] }
false
ycfNTU/Album_llama2_zeroshot
default
test
278
summary
0
0
4
2,631
838.31295
728.5
466.64161
{ "bin_edges": [ 4, 267, 530, 793, 1056, 1319, 1582, 1845, 2108, 2371, 2631 ], "hist": [ 12, 80, 65, 47, 33, 15, 15, 7, 2, 2 ] }
false
ycfNTU/Album_llama2_zeroshot
default
test
278
text
0
0
0
3,994
1,987.57914
2,105.5
1,241.24994
{ "bin_edges": [ 0, 400, 800, 1200, 1600, 2000, 2400, 2800, 3200, 3600, 3994 ], "hist": [ 46, 22, 20, 21, 21, 28, 22, 41, 34, 23 ] }
false
sndillmops/demo_train_prepared
default
train
40
horsepower
0
0
2
3
2.675
3
0.47434
{ "bin_edges": [ 2, 3, 3 ], "hist": [ 13, 27 ] }
false
ycfNTU/USB_llama3_zeroshot
default
test
398
id
0
0
27
32
30.88442
31
1.23619
{ "bin_edges": [ 27, 28, 29, 30, 31, 32, 32 ], "hist": [ 15, 9, 33, 26, 182, 133 ] }
false
ycfNTU/USB_llama3_zeroshot
default
test
398
output
0
0
25
834
180.5804
129.5
139.75069
{ "bin_edges": [ 25, 106, 187, 268, 349, 430, 511, 592, 673, 754, 834 ], "hist": [ 169, 73, 60, 49, 25, 11, 5, 3, 1, 2 ] }
false
ycfNTU/USB_llama3_zeroshot
default
test
398
text
0
0
397
3,973
2,612.75126
2,666.5
870.34393
{ "bin_edges": [ 397, 755, 1113, 1471, 1829, 2187, 2545, 2903, 3261, 3619, 3973 ], "hist": [ 3, 19, 30, 39, 38, 50, 50, 48, 71, 50 ] }
false
ycfNTU/USB_llama3_zeroshot
default
test
398
topic_name
0
0
3
68
15.20854
14
9.48467
{ "bin_edges": [ 3, 10, 17, 24, 31, 38, 45, 52, 59, 66, 68 ], "hist": [ 139, 120, 81, 34, 11, 6, 3, 2, 1, 1 ] }
false
ycfNTU/Album_llama3_zeroshot
default
test
278
exid
0
0
8
12
11.49281
12
0.62292
{ "bin_edges": [ 8, 9, 10, 11, 12, 12 ], "hist": [ 1, 1, 10, 114, 152 ] }
false
ycfNTU/Album_llama3_zeroshot
default
test
278
output
0
0
0
3,935
597.21583
422.5
597.61626
{ "bin_edges": [ 0, 394, 788, 1182, 1576, 1970, 2364, 2758, 3152, 3546, 3935 ], "hist": [ 133, 63, 38, 22, 13, 5, 2, 0, 1, 1 ] }
false
ycfNTU/Album_llama3_zeroshot
default
test
278
summary
0
0
0
2,713
620.8741
547.5
439.98243
{ "bin_edges": [ 0, 272, 544, 816, 1088, 1360, 1632, 1904, 2176, 2448, 2713 ], "hist": [ 48, 89, 79, 28, 16, 9, 3, 2, 3, 1 ] }
false
ycfNTU/Album_llama3_zeroshot
default
test
278
text
0
0
0
3,994
1,987.57914
2,105.5
1,241.24994
{ "bin_edges": [ 0, 400, 800, 1200, 1600, 2000, 2400, 2800, 3200, 3600, 3994 ], "hist": [ 46, 22, 20, 21, 21, 28, 22, 41, 34, 23 ] }
false
ycfNTU/Animal_llama3_zeroshot
default
test
1,333
exid
0
0
9
12
11.48237
12
0.60332
{ "bin_edges": [ 9, 10, 11, 12, 12 ], "hist": [ 7, 55, 559, 712 ] }
false
ycfNTU/Animal_llama3_zeroshot
default
test
1,333
output
0
0
0
3,572
271.93548
95
431.66047
{ "bin_edges": [ 0, 358, 716, 1074, 1432, 1790, 2148, 2506, 2864, 3222, 3572 ], "hist": [ 986, 192, 87, 29, 18, 8, 7, 3, 1, 2 ] }
false
ycfNTU/Animal_llama3_zeroshot
default
test
1,333
summary
0
0
0
2,984
553.57389
503
401.82343
{ "bin_edges": [ 0, 299, 598, 897, 1196, 1495, 1794, 2093, 2392, 2691, 2984 ], "hist": [ 270, 664, 253, 57, 38, 21, 11, 9, 6, 4 ] }
false
ycfNTU/Animal_llama3_zeroshot
default
test
1,333
text
0
0
0
3,978
2,057.86647
2,264
996.23483
{ "bin_edges": [ 0, 398, 796, 1194, 1592, 1990, 2388, 2786, 3184, 3582, 3978 ], "hist": [ 69, 119, 132, 129, 151, 106, 312, 111, 125, 79 ] }
false
vinay876/llm_studio_12
default
train
12
answers
0
0
11
110
46.16667
36.5
32.1073
{ "bin_edges": [ 11, 21, 31, 41, 51, 61, 71, 81, 91, 101, 110 ], "hist": [ 3, 1, 4, 0, 1, 0, 1, 0, 1, 1 ] }
false
vinay876/llm_studio_12
default
train
12
question
0
0
33
79
48.33333
43.5
13.4795
{ "bin_edges": [ 33, 38, 43, 48, 53, 58, 63, 68, 73, 78, 79 ], "hist": [ 2, 3, 2, 1, 1, 1, 1, 0, 0, 1 ] }
false
Asap7772/sft_prm800k_processed
default
test
1,000
text
0
0
261
2,992
1,197.327
1,102
516.34853
{ "bin_edges": [ 261, 535, 809, 1083, 1357, 1631, 1905, 2179, 2453, 2727, 2992 ], "hist": [ 46, 206, 232, 177, 127, 101, 60, 35, 8, 8 ] }
false
Asap7772/sft_prm800k_processed
default
train
11,563
text
0
0
217
3,080
1,186.27571
1,099
499.90127
{ "bin_edges": [ 217, 504, 791, 1078, 1365, 1652, 1939, 2226, 2513, 2800, 3080 ], "hist": [ 514, 2177, 2880, 2373, 1495, 1024, 658, 324, 93, 25 ] }
false
Asap7772/sft_math_processed
default
test
1,000
text
0
0
168
3,366
883.459
777.5
420.22808
{ "bin_edges": [ 168, 488, 808, 1128, 1448, 1768, 2088, 2408, 2728, 3048, 3366 ], "hist": [ 145, 376, 241, 136, 67, 25, 5, 1, 3, 1 ] }
false
Asap7772/sft_math_processed
default
train
393,944
text
0
0
71
6,844
876.26631
765
441.15523
{ "bin_edges": [ 71, 749, 1427, 2105, 2783, 3461, 4139, 4817, 5495, 6173, 6844 ], "hist": [ 189709, 159471, 38438, 4994, 1016, 229, 67, 12, 7, 1 ] }
false
Noorgha/pytest
default
train
981
input
0
0
30
2,728
235.90724
151
237.55319
{ "bin_edges": [ 30, 300, 570, 840, 1110, 1380, 1650, 1920, 2190, 2460, 2728 ], "hist": [ 732, 178, 40, 20, 7, 2, 0, 1, 0, 1 ] }
false
Noorgha/pytest
default
train
981
output
0
0
82
6,863
459.98981
371
371.58218
{ "bin_edges": [ 82, 761, 1440, 2119, 2798, 3477, 4156, 4835, 5514, 6193, 6863 ], "hist": [ 855, 112, 9, 3, 1, 0, 0, 0, 0, 1 ] }
false
b-brave/speech_disorders_voice
default
train
1,137
transcription
0
0
3
151
26.29551
15
24.49437
{ "bin_edges": [ 3, 18, 33, 48, 63, 78, 93, 108, 123, 138, 151 ], "hist": [ 659, 197, 99, 67, 51, 27, 17, 14, 5, 1 ] }
false
b-brave/speech_disorders_voice
default
test
260
transcription
0
0
3
255
24.95385
15
32.78616
{ "bin_edges": [ 3, 29, 55, 81, 107, 133, 159, 185, 211, 237, 255 ], "hist": [ 203, 37, 6, 4, 3, 4, 1, 0, 0, 2 ] }
false
lmms-lab/LiveBenchResults
2024-06
test
19
Model Name
0
0
6
27
15.57895
14
5.61066
{ "bin_edges": [ 6, 9, 12, 15, 18, 21, 24, 27, 27 ], "hist": [ 1, 2, 9, 1, 2, 2, 1, 1 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_arc_challenge
2024_06_21T13_14_59.631581.json
1,172
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 1172 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_arc_challenge
2024_06_21T13_14_59.631581.json
1,172
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 1172 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_arc_challenge
latest
1,172
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 1172 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_arc_challenge
latest
1,172
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 1172 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_boolean_expressions
2024_07_22T09_11_41.186841
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_boolean_expressions
2024_07_22T09_11_41.186841
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_boolean_expressions
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_boolean_expressions
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_causal_judgement
2024_07_22T09_11_41.186841
187
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 187 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_causal_judgement
2024_07_22T09_11_41.186841
187
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 187 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_causal_judgement
latest
187
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 187 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_causal_judgement
latest
187
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 187 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_date_understanding
2024_07_22T09_11_41.186841
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_date_understanding
2024_07_22T09_11_41.186841
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_date_understanding
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_date_understanding
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_disambiguation_qa
2024_07_22T09_11_41.186841
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_disambiguation_qa
2024_07_22T09_11_41.186841
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_disambiguation_qa
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_disambiguation_qa
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_formal_fallacies
2024_07_22T09_11_41.186841
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_formal_fallacies
2024_07_22T09_11_41.186841
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_formal_fallacies
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_formal_fallacies
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_geometric_shapes
2024_07_22T09_11_41.186841
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_geometric_shapes
2024_07_22T09_11_41.186841
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_geometric_shapes
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_geometric_shapes
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_hyperbaton
2024_07_22T09_11_41.186841
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_hyperbaton
2024_07_22T09_11_41.186841
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_hyperbaton
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_hyperbaton
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_logical_deduction_five_objects
2024_07_22T09_11_41.186841
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_logical_deduction_five_objects
2024_07_22T09_11_41.186841
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_logical_deduction_five_objects
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_logical_deduction_five_objects
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_logical_deduction_seven_objects
2024_07_22T09_11_41.186841
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_logical_deduction_seven_objects
2024_07_22T09_11_41.186841
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_logical_deduction_seven_objects
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_logical_deduction_seven_objects
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_logical_deduction_three_objects
2024_07_22T09_11_41.186841
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_logical_deduction_three_objects
2024_07_22T09_11_41.186841
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_logical_deduction_three_objects
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_logical_deduction_three_objects
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_movie_recommendation
2024_07_22T09_11_41.186841
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_movie_recommendation
2024_07_22T09_11_41.186841
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_movie_recommendation
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_movie_recommendation
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_navigate
2024_07_22T09_11_41.186841
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_navigate
2024_07_22T09_11_41.186841
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_navigate
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/allenai__OLMo-7B-Instruct-hf-details
allenai__OLMo-7B-Instruct-hf__leaderboard_bbh_navigate
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false