dataset
stringlengths
5
115
config
stringlengths
1
162
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
0
77.9k
null_count
int64
0
62.9M
null_proportion
float64
0
1
min
int64
0
9.25M
max
int64
0
1.07B
mean
float64
0
90.4M
median
float64
0
80.1M
std
float64
0
130M
histogram
dict
partial
bool
2 classes
davidguzmanr/AfriInstruct
afr-fra
train
985
instruction
0
0
99
449
209.67919
203
47.06107
{ "bin_edges": [ 99, 135, 171, 207, 243, 279, 315, 351, 387, 423, 449 ], "hist": [ 25, 156, 345, 263, 117, 50, 17, 8, 3, 1 ] }
false
davidguzmanr/AfriInstruct
afr-fra
train
985
output
0
0
26
399
152.78274
147
52.02123
{ "bin_edges": [ 26, 64, 102, 140, 178, 216, 254, 292, 330, 368, 399 ], "hist": [ 16, 136, 272, 278, 174, 74, 20, 8, 5, 2 ] }
false
davidguzmanr/AfriInstruct
por
train
53,700
instruction
0
0
40
18,510
253.99577
148
538.13564
{ "bin_edges": [ 40, 1888, 3736, 5584, 7432, 9280, 11128, 12976, 14824, 16672, 18510 ], "hist": [ 52131, 1440, 45, 31, 21, 16, 6, 3, 5, 2 ] }
false
davidguzmanr/AfriInstruct
por
train
53,700
output
0
0
2
2,047
115.85561
79
211.79324
{ "bin_edges": [ 2, 207, 412, 617, 822, 1027, 1232, 1437, 1642, 1847, 2047 ], "hist": [ 49591, 2716, 366, 137, 349, 29, 37, 27, 27, 421 ] }
false
davidguzmanr/AfriInstruct
eng
train
53,700
instruction
0
0
40
18,510
253.99577
148
538.13564
{ "bin_edges": [ 40, 1888, 3736, 5584, 7432, 9280, 11128, 12976, 14824, 16672, 18510 ], "hist": [ 52131, 1440, 45, 31, 21, 16, 6, 3, 5, 2 ] }
false
davidguzmanr/AfriInstruct
eng
train
53,700
output
0
0
2
2,047
115.85561
79
211.79324
{ "bin_edges": [ 2, 207, 412, 617, 822, 1027, 1232, 1437, 1642, 1847, 2047 ], "hist": [ 49591, 2716, 366, 137, 349, 29, 37, 27, 27, 421 ] }
false
davidguzmanr/AfriInstruct
ibo-fra
train
53,700
instruction
0
0
40
18,510
253.99577
148
538.13564
{ "bin_edges": [ 40, 1888, 3736, 5584, 7432, 9280, 11128, 12976, 14824, 16672, 18510 ], "hist": [ 52131, 1440, 45, 31, 21, 16, 6, 3, 5, 2 ] }
false
davidguzmanr/AfriInstruct
ibo-fra
train
53,700
output
0
0
2
2,047
115.85561
79
211.79324
{ "bin_edges": [ 2, 207, 412, 617, 822, 1027, 1232, 1437, 1642, 1847, 2047 ], "hist": [ 49591, 2716, 366, 137, 349, 29, 37, 27, 27, 421 ] }
false
davidguzmanr/AfriInstruct
nya-eng
train
53,700
instruction
0
0
40
18,510
253.99577
148
538.13564
{ "bin_edges": [ 40, 1888, 3736, 5584, 7432, 9280, 11128, 12976, 14824, 16672, 18510 ], "hist": [ 52131, 1440, 45, 31, 21, 16, 6, 3, 5, 2 ] }
false
davidguzmanr/AfriInstruct
nya-eng
train
53,700
output
0
0
2
2,047
115.85561
79
211.79324
{ "bin_edges": [ 2, 207, 412, 617, 822, 1027, 1232, 1437, 1642, 1847, 2047 ], "hist": [ 49591, 2716, 366, 137, 349, 29, 37, 27, 27, 421 ] }
false
davidguzmanr/AfriInstruct
twi
train
53,830
instruction
0
0
21
18,510
253.64826
148
537.53727
{ "bin_edges": [ 21, 1870, 3719, 5568, 7417, 9266, 11115, 12964, 14813, 16662, 18510 ], "hist": [ 52255, 1444, 47, 31, 21, 16, 6, 3, 5, 2 ] }
false
davidguzmanr/AfriInstruct
twi
train
53,830
output
0
0
2
2,047
115.664
79
211.57441
{ "bin_edges": [ 2, 207, 412, 617, 822, 1027, 1232, 1437, 1642, 1847, 2047 ], "hist": [ 49721, 2716, 366, 137, 349, 29, 37, 27, 27, 421 ] }
false
davidguzmanr/AfriInstruct
ibo
train
95,300
instruction
0
0
3
18,510
238.85475
130
486.81557
{ "bin_edges": [ 3, 1854, 3705, 5556, 7407, 9258, 11109, 12960, 14811, 16662, 18510 ], "hist": [ 92441, 2700, 69, 36, 22, 16, 6, 3, 5, 2 ] }
false
davidguzmanr/AfriInstruct
ibo
train
95,300
output
0
0
0
6,730
153.61662
119
206.324
{ "bin_edges": [ 0, 674, 1348, 2022, 2696, 3370, 4044, 4718, 5392, 6066, 6730 ], "hist": [ 93533, 1165, 173, 402, 9, 11, 2, 2, 2, 1 ] }
false
davidguzmanr/AfriInstruct
zul-eng
train
53,700
instruction
0
0
40
18,510
253.99577
148
538.13564
{ "bin_edges": [ 40, 1888, 3736, 5584, 7432, 9280, 11128, 12976, 14824, 16672, 18510 ], "hist": [ 52131, 1440, 45, 31, 21, 16, 6, 3, 5, 2 ] }
false
davidguzmanr/AfriInstruct
zul-eng
train
53,700
output
0
0
2
2,047
115.85561
79
211.79324
{ "bin_edges": [ 2, 207, 412, 617, 822, 1027, 1232, 1437, 1642, 1847, 2047 ], "hist": [ 49591, 2716, 366, 137, 349, 29, 37, 27, 27, 421 ] }
false
davidguzmanr/AfriInstruct
wol
train
53,760
instruction
0
0
14
18,510
253.76802
148
537.88038
{ "bin_edges": [ 14, 1864, 3714, 5564, 7414, 9264, 11114, 12964, 14814, 16664, 18510 ], "hist": [ 52184, 1445, 45, 33, 21, 16, 6, 3, 5, 2 ] }
false
davidguzmanr/AfriInstruct
wol
train
53,760
output
0
0
2
2,047
115.7904
79
211.70638
{ "bin_edges": [ 2, 207, 412, 617, 822, 1027, 1232, 1437, 1642, 1847, 2047 ], "hist": [ 49645, 2722, 366, 137, 349, 29, 37, 27, 27, 421 ] }
false
ebayes/uhura-instruct-llama
default
train
548,561
text
0
0
99
46,527
917.47755
681
1,078.03012
{ "bin_edges": [ 99, 4742, 9385, 14028, 18671, 23314, 27957, 32600, 37243, 41886, 46527 ], "hist": [ 540374, 6885, 991, 229, 47, 22, 9, 2, 1, 1 ] }
false
ebayes/uhura-eval
ha_multiple_choice
test
155
a
0
0
1
60
12.76129
10
9.23298
{ "bin_edges": [ 1, 7, 13, 19, 25, 31, 37, 43, 49, 55, 60 ], "hist": [ 50, 44, 25, 16, 16, 2, 0, 0, 1, 1 ] }
false
ebayes/uhura-eval
ha_multiple_choice
test
155
b
0
0
2
42
11.92258
10
7.81983
{ "bin_edges": [ 2, 7, 12, 17, 22, 27, 32, 37, 42, 42 ], "hist": [ 49, 36, 36, 15, 8, 7, 3, 0, 1 ] }
false
ebayes/uhura-eval
ha_multiple_choice
test
155
c
0
0
2
41
12.26452
11
8.13369
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 34, 38, 41 ], "hist": [ 33, 38, 29, 21, 14, 7, 6, 1, 4, 2 ] }
false
ebayes/uhura-eval
ha_multiple_choice
test
155
d
0
0
2
38
12.0129
10
7.68029
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 34, 38, 38 ], "hist": [ 29, 45, 28, 21, 11, 8, 9, 2, 1, 1 ] }
false
ebayes/uhura-eval
ha_multiple_choice
test
155
question
0
0
5
334
53.25806
42
40.33896
{ "bin_edges": [ 5, 38, 71, 104, 137, 170, 203, 236, 269, 302, 334 ], "hist": [ 58, 68, 17, 4, 4, 3, 0, 0, 0, 1 ] }
false
ebayes/uhura-eval
sw_generation
test
73
a
0
0
0
232
41.31507
16
53.89982
{ "bin_edges": [ 0, 24, 48, 72, 96, 120, 144, 168, 192, 216, 232 ], "hist": [ 41, 11, 6, 3, 5, 3, 0, 2, 0, 2 ] }
false
ebayes/uhura-eval
sw_generation
test
73
q
0
0
0
305
72.0137
62
44.21365
{ "bin_edges": [ 0, 31, 62, 93, 124, 155, 186, 217, 248, 279, 305 ], "hist": [ 6, 28, 22, 10, 4, 2, 0, 0, 0, 1 ] }
false
ebayes/uhura-eval
yo_multiple_choice
test
258
a
0
0
3
45
11.68217
9
7.54956
{ "bin_edges": [ 3, 8, 13, 18, 23, 28, 33, 38, 43, 45 ], "hist": [ 88, 84, 40, 21, 14, 5, 2, 3, 1 ] }
false
ebayes/uhura-eval
yo_multiple_choice
test
258
b
0
0
2
43
11.93023
9.5
7.07072
{ "bin_edges": [ 2, 7, 12, 17, 22, 27, 32, 37, 42, 43 ], "hist": [ 55, 100, 47, 27, 17, 7, 4, 0, 1 ] }
false
ebayes/uhura-eval
yo_multiple_choice
test
258
c
0
0
2
40
12.23643
9
7.4475
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 34, 38, 40 ], "hist": [ 33, 100, 39, 25, 26, 22, 4, 3, 5, 1 ] }
false
ebayes/uhura-eval
yo_multiple_choice
test
258
d
0
0
0
46
12.37984
9.5
8.08726
{ "bin_edges": [ 0, 5, 10, 15, 20, 25, 30, 35, 40, 45, 46 ], "hist": [ 24, 105, 51, 32, 21, 12, 8, 1, 3, 1 ] }
false
ebayes/uhura-eval
yo_multiple_choice
test
258
question
0
0
4
74
33.92636
34
14.43321
{ "bin_edges": [ 4, 12, 20, 28, 36, 44, 52, 60, 68, 74 ], "hist": [ 8, 42, 44, 43, 66, 27, 9, 17, 2 ] }
false
ebayes/uhura-eval
am_multiple_choice
test
77
a
0
0
2
46
10.77922
8
9.48354
{ "bin_edges": [ 2, 7, 12, 17, 22, 27, 32, 37, 42, 46 ], "hist": [ 35, 18, 9, 6, 3, 1, 2, 2, 1 ] }
false
ebayes/uhura-eval
am_multiple_choice
test
77
b
0
0
2
49
10.20779
8
8.50903
{ "bin_edges": [ 2, 7, 12, 17, 22, 27, 32, 37, 42, 47, 49 ], "hist": [ 32, 22, 9, 7, 3, 2, 1, 0, 0, 1 ] }
false
ebayes/uhura-eval
am_multiple_choice
test
77
c
0
0
0
41
10.83117
8
9.18581
{ "bin_edges": [ 0, 5, 10, 15, 20, 25, 30, 35, 40, 41 ], "hist": [ 26, 20, 9, 10, 4, 5, 0, 2, 1 ] }
false
ebayes/uhura-eval
am_multiple_choice
test
77
d
0
0
0
42
9.45455
7
7.91326
{ "bin_edges": [ 0, 5, 10, 15, 20, 25, 30, 35, 40, 42 ], "hist": [ 27, 22, 9, 9, 6, 3, 0, 0, 1 ] }
false
ebayes/uhura-eval
am_multiple_choice
test
77
question
0
0
19
96
50.67532
45
18.86688
{ "bin_edges": [ 19, 27, 35, 43, 51, 59, 67, 75, 83, 91, 96 ], "hist": [ 5, 10, 19, 12, 5, 9, 8, 4, 1, 4 ] }
false
mesolitica/translate-DenseFusion-1M
default
train
489,430
caption
0
0
74
2,983
1,113.39383
1,126
276.65086
{ "bin_edges": [ 74, 365, 656, 947, 1238, 1529, 1820, 2111, 2402, 2693, 2983 ], "hist": [ 93, 19686, 122509, 183103, 132961, 27363, 3470, 214, 25, 6 ] }
false
mesolitica/translate-DenseFusion-1M
default
train
489,430
caption_ms
0
0
86
16,367
1,136.71214
1,146
285.29414
{ "bin_edges": [ 86, 1715, 3344, 4973, 6602, 8231, 9860, 11489, 13118, 14747, 16367 ], "hist": [ 478050, 11368, 7, 1, 2, 0, 0, 0, 0, 2 ] }
false
mesolitica/translate-DenseFusion-1M
default
train
489,430
image_id
0
0
7
13
12.95113
13
0.26378
{ "bin_edges": [ 7, 8, 9, 10, 11, 12, 13, 13 ], "hist": [ 2, 0, 58, 790, 2905, 15492, 470183 ] }
false
mesolitica/translate-DenseFusion-1M
default
train
489,430
url
0
0
25
3,802
107.27762
92
79.58816
{ "bin_edges": [ 25, 403, 781, 1159, 1537, 1915, 2293, 2671, 3049, 3427, 3802 ], "hist": [ 486075, 2550, 397, 2, 390, 0, 4, 0, 3, 9 ] }
false
neph1/Alpaca-Lora-GPT4-Swedish-Refined
default
train
52,002
id
0
0
6
10
9.78635
10
0.46402
{ "bin_edges": [ 6, 7, 8, 9, 10, 10 ], "hist": [ 10, 90, 900, 9000, 42002 ] }
false
neph1/Alpaca-Lora-GPT4-Swedish-Refined
default
train
52,002
input
0
0
0
2,518
22.67195
0
58.09918
{ "bin_edges": [ 0, 252, 504, 756, 1008, 1260, 1512, 1764, 2016, 2268, 2518 ], "hist": [ 51389, 523, 71, 8, 0, 7, 2, 1, 0, 1 ] }
false
neph1/Alpaca-Lora-GPT4-Swedish-Refined
default
train
52,002
instruction
0
0
10
1,529
58.72736
56
22.78988
{ "bin_edges": [ 10, 162, 314, 466, 618, 770, 922, 1074, 1226, 1378, 1529 ], "hist": [ 51890, 100, 10, 1, 0, 0, 0, 0, 0, 1 ] }
false
neph1/Alpaca-Lora-GPT4-Swedish-Refined
default
train
52,002
original_instruction
0
0
9
489
59.77874
57
21.19347
{ "bin_edges": [ 9, 58, 107, 156, 205, 254, 303, 352, 401, 450, 489 ], "hist": [ 26897, 23625, 1325, 121, 20, 3, 2, 2, 4, 3 ] }
false
neph1/Alpaca-Lora-GPT4-Swedish-Refined
default
train
52,002
output
0
0
0
6,094
677.95794
466
657.44035
{ "bin_edges": [ 0, 610, 1220, 1830, 2440, 3050, 3660, 4270, 4880, 5490, 6094 ], "hist": [ 29378, 10852, 8166, 3110, 480, 11, 2, 0, 2, 1 ] }
false
D4vidHuang/FSE2025_MappingWithExclusion_ZH
default
train
500
content
0
0
155
32,767
5,521.358
3,103
5,968.69075
{ "bin_edges": [ 155, 3417, 6679, 9941, 13203, 16465, 19727, 22989, 26251, 29513, 32767 ], "hist": [ 267, 96, 57, 26, 14, 16, 11, 9, 2, 2 ] }
false
D4vidHuang/FSE2025_MappingWithExclusion_ZH
default
train
500
file_id
0
0
5
9
7.274
7
0.53619
{ "bin_edges": [ 5, 6, 7, 8, 9, 9 ], "hist": [ 1, 14, 337, 143, 5 ] }
false
D4vidHuang/FSE2025_MappingWithExclusion_ZH
default
train
500
masked_comment
0
0
111
32,767
5,479.466
3,086.5
5,982.07204
{ "bin_edges": [ 111, 3377, 6643, 9909, 13175, 16441, 19707, 22973, 26239, 29505, 32767 ], "hist": [ 267, 97, 56, 26, 15, 15, 11, 9, 2, 2 ] }
false
D4vidHuang/FSE2025_MappingWithExclusion_ZH
default
train
500
original_comment
0
0
5
1,825
55.128
18
161.16476
{ "bin_edges": [ 5, 188, 371, 554, 737, 920, 1103, 1286, 1469, 1652, 1825 ], "hist": [ 473, 10, 6, 5, 3, 0, 0, 1, 1, 1 ] }
false
D4vidHuang/FSE2025_MappingWithExclusion_ZH
default
train
500
path
0
0
8
120
47.876
47
22.45237
{ "bin_edges": [ 8, 20, 32, 44, 56, 68, 80, 92, 104, 116, 120 ], "hist": [ 54, 81, 80, 96, 94, 56, 22, 10, 4, 3 ] }
false
D4vidHuang/FSE2025_MappingWithExclusion_ZH
default
train
500
repo
0
0
9
59
22.578
21
8.10922
{ "bin_edges": [ 9, 15, 21, 27, 33, 39, 45, 51, 57, 59 ], "hist": [ 53, 186, 149, 60, 29, 9, 8, 4, 2 ] }
false
neoneye/simon-arc-rle-image-v21
default
train
100,000
input
0
0
12
1,861
142.24282
127
125.95118
{ "bin_edges": [ 12, 197, 382, 567, 752, 937, 1122, 1307, 1492, 1677, 1861 ], "hist": [ 83832, 14918, 50, 407, 184, 43, 158, 195, 136, 77 ] }
false
neoneye/simon-arc-rle-image-v21
default
train
100,000
output
0
0
4
826
136.62727
132
83.08587
{ "bin_edges": [ 4, 87, 170, 253, 336, 419, 502, 585, 668, 751, 826 ], "hist": [ 26761, 42808, 23433, 6078, 110, 185, 294, 212, 84, 35 ] }
false
hubin/tape_gpt_pred_explain
default
train
169,343
text
0
0
33
1,841
778.6256
772
211.95282
{ "bin_edges": [ 33, 214, 395, 576, 757, 938, 1119, 1300, 1481, 1662, 1841 ], "hist": [ 78, 1836, 28573, 49482, 55033, 25156, 5615, 2294, 1245, 31 ] }
false
mesolitica/malaysian-politician-hansard-style
default
train
300,459
answer
0
0
0
4,240
328.1433
304
177.23512
{ "bin_edges": [ 0, 425, 850, 1275, 1700, 2125, 2550, 2975, 3400, 3825, 4240 ], "hist": [ 235782, 60595, 3290, 541, 154, 49, 24, 15, 6, 3 ] }
false
mesolitica/malaysian-politician-hansard-style
default
train
300,459
personality
0
0
0
259
20.08071
20
10.95255
{ "bin_edges": [ 0, 26, 52, 78, 104, 130, 156, 182, 208, 234, 259 ], "hist": [ 245192, 49723, 4923, 368, 148, 61, 13, 18, 8, 5 ] }
false
mesolitica/malaysian-politician-hansard-style
default
train
300,459
politician
0
0
10
121
29.1768
28
9.43208
{ "bin_edges": [ 10, 22, 34, 46, 58, 70, 82, 94, 106, 118, 121 ], "hist": [ 66034, 151329, 68621, 11461, 1904, 779, 276, 38, 11, 6 ] }
false
mesolitica/malaysian-politician-hansard-style
default
train
300,459
question
0
0
0
1,210
106.84421
86
69.59847
{ "bin_edges": [ 0, 122, 244, 366, 488, 610, 732, 854, 976, 1098, 1210 ], "hist": [ 220797, 64324, 12049, 2488, 578, 151, 47, 17, 5, 3 ] }
false
mesolitica/malaysian-politician-hansard-style
default
train
300,459
topic
0
0
0
243
18.33878
17
9.38994
{ "bin_edges": [ 0, 25, 50, 75, 100, 125, 150, 175, 200, 225, 243 ], "hist": [ 236509, 61466, 2289, 165, 27, 2, 0, 0, 0, 1 ] }
false
mesolitica/malaysian-politician-hansard-style
default
train
300,459
url
0
0
60
60
60
60
0
{ "bin_edges": [ 60, 60 ], "hist": [ 300459 ] }
false
prabal123/SQL_custom_dataset_4
default
train
289
text
0
0
465
888
625.88581
628
70.22413
{ "bin_edges": [ 465, 508, 551, 594, 637, 680, 723, 766, 809, 852, 888 ], "hist": [ 16, 28, 48, 67, 58, 60, 5, 5, 0, 2 ] }
false
mgrtsv/reddit_dataset_231
default
train
22,000,000
communityName
0
0
4
23
12.89629
12
4.32209
{ "bin_edges": [ 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 23 ], "hist": [ 531908, 1735734, 2995775, 3717392, 3841792, 3104405, 2505508, 1675585, 1146301, 745600 ] }
true
mgrtsv/reddit_dataset_231
default
train
22,000,000
datetime
0
0
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 22000000 ] }
true
mgrtsv/reddit_dataset_231
default
train
22,000,000
label
0
0
4
23
12.89629
12
4.32209
{ "bin_edges": [ 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 23 ], "hist": [ 531908, 1735734, 2995775, 3717392, 3841792, 3104405, 2505508, 1675585, 1146301, 745600 ] }
true
mgrtsv/reddit_dataset_231
default
train
22,000,000
text
0
0
0
54,442
221.48813
84
597.63736
{ "bin_edges": [ 0, 5445, 10890, 16335, 21780, 27225, 32670, 38115, 43560, 49005, 54442 ], "hist": [ 21964898, 26178, 4890, 1886, 1014, 553, 393, 187, 0, 1 ] }
true
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_boolean_expressions
2024_07_21T07_14_53.671049
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_boolean_expressions
2024_07_21T07_14_53.671049
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_boolean_expressions
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_boolean_expressions
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_date_understanding
2024_07_21T07_14_53.671049
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_date_understanding
2024_07_21T07_14_53.671049
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_date_understanding
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_date_understanding
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_disambiguation_qa
2024_07_21T07_14_53.671049
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_disambiguation_qa
2024_07_21T07_14_53.671049
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_disambiguation_qa
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_disambiguation_qa
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_formal_fallacies
2024_07_21T07_14_53.671049
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_formal_fallacies
2024_07_21T07_14_53.671049
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_formal_fallacies
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_formal_fallacies
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_logical_deduction_five_objects
2024_07_21T07_14_53.671049
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_logical_deduction_five_objects
2024_07_21T07_14_53.671049
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_logical_deduction_five_objects
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_logical_deduction_five_objects
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_hyperbaton
2024_07_21T07_14_53.671049
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_hyperbaton
2024_07_21T07_14_53.671049
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_hyperbaton
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_hyperbaton
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_logical_deduction_seven_objects
2024_07_21T07_14_53.671049
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_logical_deduction_seven_objects
2024_07_21T07_14_53.671049
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_logical_deduction_seven_objects
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_logical_deduction_seven_objects
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_logical_deduction_three_objects
2024_07_21T07_14_53.671049
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_logical_deduction_three_objects
2024_07_21T07_14_53.671049
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_logical_deduction_three_objects
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_logical_deduction_three_objects
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_navigate
2024_07_21T07_14_53.671049
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_navigate
2024_07_21T07_14_53.671049
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct-details
VAGOsolutions__Llama-3-SauerkrautLM-70b-Instruct__leaderboard_bbh_navigate
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false