dataset
stringlengths
5
115
config
stringlengths
1
162
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
0
77.9k
null_count
int64
0
62.9M
null_proportion
float64
0
1
min
int64
0
9.25M
max
int64
0
1.07B
mean
float64
0
90.4M
median
float64
0
80.1M
std
float64
0
130M
histogram
dict
partial
bool
2 classes
MrbBakh/AirlineSentiment
default
train
10,248
name
0
0
2
19
10.53649
11
2.618
{ "bin_edges": [ 2, 4, 6, 8, 10, 12, 14, 16, 18, 19 ], "hist": [ 8, 169, 1215, 2351, 2715, 2128, 1656, 0, 6 ] }
false
MrbBakh/AirlineSentiment
default
train
10,248
text
0
0
12
170
103.57904
114
36.39286
{ "bin_edges": [ 12, 28, 44, 60, 76, 92, 108, 124, 140, 156, 170 ], "hist": [ 328, 573, 686, 917, 1007, 1140, 1287, 3013, 1243, 54 ] }
false
MrbBakh/AirlineSentiment
default
train
10,248
tweet_created
0
0
14
15
14.67086
15
0.46992
{ "bin_edges": [ 14, 15, 15 ], "hist": [ 3373, 6875 ] }
false
MrbBakh/AirlineSentiment
default
train
10,248
tweet_location
3,289
0.32094
1
30
13.42434
13
6.37052
{ "bin_edges": [ 1, 4, 7, 10, 13, 16, 19, 22, 25, 28, 30 ], "hist": [ 380, 460, 936, 1603, 1568, 676, 462, 331, 275, 268 ] }
false
MrbBakh/AirlineSentiment
default
validation
2,928
name
0
0
3
15
10.53415
11
2.6294
{ "bin_edges": [ 3, 5, 7, 9, 11, 13, 15, 15 ], "hist": [ 18, 161, 564, 692, 740, 543, 210 ] }
false
MrbBakh/AirlineSentiment
default
validation
2,928
text
0
0
12
176
103.76298
114
36.14648
{ "bin_edges": [ 12, 29, 46, 63, 80, 97, 114, 131, 148, 165, 176 ], "hist": [ 99, 162, 245, 278, 318, 355, 449, 943, 78, 1 ] }
false
MrbBakh/AirlineSentiment
default
validation
2,928
tweet_created
0
0
14
15
14.66667
15
0.47149
{ "bin_edges": [ 14, 15, 15 ], "hist": [ 976, 1952 ] }
false
MrbBakh/AirlineSentiment
default
validation
2,928
tweet_location
979
0.33436
1
30
13.44433
13
6.35036
{ "bin_edges": [ 1, 4, 7, 10, 13, 16, 19, 22, 25, 28, 30 ], "hist": [ 123, 116, 235, 457, 450, 192, 132, 97, 81, 66 ] }
false
MrbBakh/FinancialClassification
default
test
585
text
0
0
9
295
119.13162
109
56.82234
{ "bin_edges": [ 9, 38, 67, 96, 125, 154, 183, 212, 241, 270, 295 ], "hist": [ 10, 104, 130, 105, 85, 65, 42, 21, 16, 7 ] }
false
MrbBakh/FinancialClassification
default
train
4,089
text
0
0
10
306
117.03693
107
56.33463
{ "bin_edges": [ 10, 40, 70, 100, 130, 160, 190, 220, 250, 280, 306 ], "hist": [ 131, 790, 914, 794, 589, 351, 272, 159, 63, 26 ] }
false
MrbBakh/FinancialClassification
default
validation
1,168
text
0
0
12
315
115.91438
106
56.53645
{ "bin_edges": [ 12, 43, 74, 105, 136, 167, 198, 229, 260, 291, 315 ], "hist": [ 59, 249, 267, 223, 161, 95, 53, 42, 14, 5 ] }
false
aifeifei798/DPO_Pairs-Roleplay-NSFW
default
train
3,428
chosen
0
0
49
4,073
777.04697
587
554.25698
{ "bin_edges": [ 49, 452, 855, 1258, 1661, 2064, 2467, 2870, 3273, 3676, 4073 ], "hist": [ 862, 1641, 452, 219, 110, 55, 42, 27, 8, 12 ] }
false
aifeifei798/DPO_Pairs-Roleplay-NSFW
default
train
3,428
prompt
0
0
472
1,565
1,038.04901
963
243.52683
{ "bin_edges": [ 472, 582, 692, 802, 912, 1022, 1132, 1242, 1352, 1462, 1565 ], "hist": [ 150, 201, 55, 256, 1380, 84, 208, 903, 75, 116 ] }
false
aifeifei798/DPO_Pairs-Roleplay-NSFW
default
train
3,428
rejected
0
0
0
3,045
571.53792
461.5
372.55421
{ "bin_edges": [ 0, 305, 610, 915, 1220, 1525, 1830, 2135, 2440, 2745, 3045 ], "hist": [ 723, 1645, 523, 286, 155, 61, 25, 5, 1, 4 ] }
false
prabal123/SQL_custom_dataset_5
default
train
273
text
0
0
467
890
629.49084
631
70.19092
{ "bin_edges": [ 467, 510, 553, 596, 639, 682, 725, 768, 811, 854, 890 ], "hist": [ 14, 26, 46, 62, 54, 59, 5, 5, 0, 2 ] }
false
Vipitis/Shadereval-experiments-dev
default
train
257
body
0
0
25
7,466
813.43969
499
998.65361
{ "bin_edges": [ 25, 770, 1515, 2260, 3005, 3750, 4495, 5240, 5985, 6730, 7466 ], "hist": [ 166, 56, 19, 5, 3, 5, 1, 1, 0, 1 ] }
false
Vipitis/Shadereval-experiments-dev
default
train
257
comment
0
0
3
999
123.07782
65
158.06931
{ "bin_edges": [ 3, 103, 203, 303, 403, 503, 603, 703, 803, 903, 999 ], "hist": [ 168, 46, 18, 13, 2, 2, 4, 1, 1, 2 ] }
false
Vipitis/Shadereval-experiments-dev
default
train
257
docstring
0
0
7
1,592
82.07004
51
122.12907
{ "bin_edges": [ 7, 166, 325, 484, 643, 802, 961, 1120, 1279, 1438, 1592 ], "hist": [ 236, 12, 7, 1, 0, 0, 0, 0, 0, 1 ] }
false
Vipitis/Shadereval-experiments-dev
default
train
257
full_code
0
0
721
36,674
8,726.87549
6,679
6,903.31944
{ "bin_edges": [ 721, 4317, 7913, 11509, 15105, 18701, 22297, 25893, 29489, 33085, 36674 ], "hist": [ 68, 95, 34, 15, 15, 13, 13, 0, 2, 2 ] }
false
Vipitis/Shadereval-experiments-dev
default
train
257
header
0
0
11
1,659
47.83658
36
103.69801
{ "bin_edges": [ 11, 176, 341, 506, 671, 836, 1001, 1166, 1331, 1496, 1659 ], "hist": [ 256, 0, 0, 0, 0, 0, 0, 0, 0, 1 ] }
false
Vipitis/Shadereval-experiments-dev
default
train
257
id
0
0
6
6
6
6
0
{ "bin_edges": [ 6, 6 ], "hist": [ 257 ] }
false
Vipitis/Shadereval-experiments-dev
default
train
257
input_both
0
0
47
2,078
252.98444
178
250.01075
{ "bin_edges": [ 47, 251, 455, 659, 863, 1067, 1271, 1475, 1679, 1883, 2078 ], "hist": [ 183, 46, 16, 3, 4, 3, 0, 0, 0, 2 ] }
false
Vipitis/Shadereval-experiments-dev
default
train
257
input_comment
0
0
31
1,982
170.9144
110
200.15687
{ "bin_edges": [ 31, 227, 423, 619, 815, 1011, 1207, 1403, 1599, 1795, 1982 ], "hist": [ 207, 35, 4, 6, 3, 1, 0, 0, 0, 1 ] }
false
Vipitis/Shadereval-experiments-dev
default
train
257
input_docstring
0
0
18
1,754
129.90661
93
159.29669
{ "bin_edges": [ 18, 192, 366, 540, 714, 888, 1062, 1236, 1410, 1584, 1754 ], "hist": [ 227, 20, 7, 1, 0, 0, 0, 0, 0, 2 ] }
false
Vipitis/Shadereval-experiments-dev
default
train
257
input_none
0
0
11
1,659
47.83658
36
103.69801
{ "bin_edges": [ 11, 176, 341, 506, 671, 836, 1001, 1166, 1331, 1496, 1659 ], "hist": [ 256, 0, 0, 0, 0, 0, 0, 0, 0, 1 ] }
false
Vipitis/Shadereval-experiments-dev
default
train
257
model_ctx
0
0
40
2,305
293.99222
176
345.46213
{ "bin_edges": [ 40, 267, 494, 721, 948, 1175, 1402, 1629, 1856, 2083, 2305 ], "hist": [ 172, 47, 16, 9, 2, 4, 2, 2, 1, 2 ] }
false
vinven7/PharmKG
default
test
10,000
completion
0
0
17
320
51.5216
45
23.75718
{ "bin_edges": [ 17, 48, 79, 110, 141, 172, 203, 234, 265, 296, 320 ], "hist": [ 5439, 3351, 942, 216, 36, 8, 5, 1, 1, 1 ] }
false
vinven7/PharmKG
default
test
10,000
prompt
0
0
484
763
512.0705
505
23.4097
{ "bin_edges": [ 484, 512, 540, 568, 596, 624, 652, 680, 708, 736, 763 ], "hist": [ 6080, 2691, 937, 236, 40, 10, 4, 0, 1, 1 ] }
false
vinven7/PharmKG
default
train
1,000,000
completion
0
0
17
442
64.98787
56
34.08503
{ "bin_edges": [ 17, 60, 103, 146, 189, 232, 275, 318, 361, 404, 442 ], "hist": [ 540020, 339067, 79589, 34811, 5094, 1193, 197, 26, 1, 2 ] }
false
vinven7/PharmKG
default
train
1,000,000
prompt
0
0
483
909
523.36078
515
31.61755
{ "bin_edges": [ 483, 526, 569, 612, 655, 698, 741, 784, 827, 870, 909 ], "hist": [ 619090, 288359, 72810, 17195, 2141, 349, 46, 9, 0, 1 ] }
false
DatPySci/tldr_comparisons_mistral7b
default
train
32
chosen
0
0
65
354
158.59375
144.5
68.249
{ "bin_edges": [ 65, 94, 123, 152, 181, 210, 239, 268, 297, 326, 354 ], "hist": [ 4, 5, 9, 5, 2, 5, 0, 0, 0, 2 ] }
false
DatPySci/tldr_comparisons_mistral7b
default
train
32
rejected
0
0
16
212
105.09375
102
55.73692
{ "bin_edges": [ 16, 36, 56, 76, 96, 116, 136, 156, 176, 196, 212 ], "hist": [ 4, 4, 5, 1, 5, 3, 4, 1, 4, 1 ] }
false
zurd46/swiss
default
train
287
assistant
0
0
36
1,999
1,009.23693
1,091
452.8824
{ "bin_edges": [ 36, 233, 430, 627, 824, 1021, 1218, 1415, 1612, 1809, 1999 ], "hist": [ 25, 19, 16, 25, 40, 63, 46, 32, 19, 2 ] }
false
zurd46/swiss
default
train
287
context
0
0
40
222
104.1324
101
28.76046
{ "bin_edges": [ 40, 59, 78, 97, 116, 135, 154, 173, 192, 211, 222 ], "hist": [ 12, 29, 77, 82, 54, 16, 10, 4, 1, 2 ] }
false
zurd46/swiss
default
train
287
user
0
0
35
150
67.61672
66
18.29477
{ "bin_edges": [ 35, 47, 59, 71, 83, 95, 107, 119, 131, 143, 150 ], "hist": [ 27, 76, 73, 53, 36, 12, 8, 1, 0, 1 ] }
false
5CD-AI/Viet-Doc-VQA
default
train
51,856
description
0
0
65
10,625
867.1406
772
408.83955
{ "bin_edges": [ 65, 1122, 2179, 3236, 4293, 5350, 6407, 7464, 8521, 9578, 10625 ], "hist": [ 40788, 10494, 554, 17, 2, 0, 0, 0, 0, 1 ] }
false
sroecker/openshift-questions-small
default
train
304
question
0
0
51
162
88.38816
86.5
19.87952
{ "bin_edges": [ 51, 63, 75, 87, 99, 111, 123, 135, 147, 159, 162 ], "hist": [ 23, 55, 74, 67, 41, 27, 10, 6, 0, 1 ] }
false
NESPED-GEN/selector_testes
default
trainSpider
8,656
query_llm
0
0
18
631
132.48001
104.5
88.77312
{ "bin_edges": [ 18, 80, 142, 204, 266, 328, 390, 452, 514, 576, 631 ], "hist": [ 3232, 2176, 1600, 899, 425, 194, 79, 38, 6, 7 ] }
false
NESPED-GEN/selector_testes
default
trainSpider
8,656
question_en
0
0
3
224
66.57763
64
24.38171
{ "bin_edges": [ 3, 26, 49, 72, 95, 118, 141, 164, 187, 210, 224 ], "hist": [ 125, 1997, 3174, 2266, 850, 187, 44, 8, 2, 3 ] }
false
NESPED-GEN/selector_testes
default
trainSpider
8,656
selector
0
0
18
266
79.30002
70
39.17476
{ "bin_edges": [ 18, 43, 68, 93, 118, 143, 168, 193, 218, 243, 266 ], "hist": [ 1461, 2626, 1809, 1279, 795, 390, 198, 75, 11, 12 ] }
false
NESPED-GEN/selector_testes
default
trainSpider
8,656
selector_correct
0
0
20
266
79.30903
70
39.16641
{ "bin_edges": [ 20, 45, 70, 95, 120, 145, 170, 195, 220, 245, 266 ], "hist": [ 1728, 2517, 1784, 1268, 711, 405, 147, 75, 9, 12 ] }
false
NESPED-GEN/selector_testes
default
devSpider
1,034
query_llm
0
0
20
486
114.82785
89
72.68497
{ "bin_edges": [ 20, 67, 114, 161, 208, 255, 302, 349, 396, 443, 486 ], "hist": [ 312, 293, 207, 113, 54, 23, 21, 6, 4, 1 ] }
false
NESPED-GEN/selector_testes
default
devSpider
1,034
question_en
0
0
18
174
68.04836
66
22.71959
{ "bin_edges": [ 18, 34, 50, 66, 82, 98, 114, 130, 146, 162, 174 ], "hist": [ 43, 179, 292, 261, 150, 69, 25, 12, 2, 1 ] }
false
NESPED-GEN/selector_testes
default
devSpider
1,034
selector
0
0
21
225
72.0058
67
30.61123
{ "bin_edges": [ 21, 42, 63, 84, 105, 126, 147, 168, 189, 210, 225 ], "hist": [ 140, 315, 272, 167, 86, 30, 16, 4, 2, 2 ] }
false
NESPED-GEN/selector_testes
default
devSpider
1,034
selector_correct
0
0
25
225
72.01934
67
30.59019
{ "bin_edges": [ 25, 46, 67, 88, 109, 130, 151, 172, 193, 214, 225 ], "hist": [ 200, 314, 246, 158, 68, 26, 14, 4, 2, 2 ] }
false
NESPED-GEN/selector_testes
default
testSpider
2,147
query_llm
0
0
22
692
122.95762
93
88.09447
{ "bin_edges": [ 22, 90, 158, 226, 294, 362, 430, 498, 566, 634, 692 ], "hist": [ 1043, 521, 332, 159, 40, 25, 15, 8, 0, 4 ] }
false
NESPED-GEN/selector_testes
default
testSpider
2,147
question_en
0
0
22
185
69.00326
67
22.6883
{ "bin_edges": [ 22, 39, 56, 73, 90, 107, 124, 141, 158, 175, 185 ], "hist": [ 152, 442, 719, 476, 227, 88, 31, 7, 4, 1 ] }
false
NESPED-GEN/selector_testes
default
testSpider
2,147
selector
0
0
21
294
75.73731
65
39.22343
{ "bin_edges": [ 21, 49, 77, 105, 133, 161, 189, 217, 245, 273, 294 ], "hist": [ 578, 703, 456, 233, 103, 39, 18, 6, 1, 10 ] }
false
NESPED-GEN/selector_testes
default
testSpider
2,147
selector_correct
0
0
21
294
75.74662
65
39.21184
{ "bin_edges": [ 21, 49, 77, 105, 133, 161, 189, 217, 245, 273, 294 ], "hist": [ 578, 703, 456, 233, 103, 39, 18, 6, 1, 10 ] }
false
Lucidest/7.21-reddit-classify
default
train
232,073
prompt
0
0
174
40,468
860.64193
486
1,156.33602
{ "bin_edges": [ 174, 4204, 8234, 12264, 16294, 20324, 24354, 28384, 32414, 36444, 40468 ], "hist": [ 228050, 3326, 476, 115, 48, 19, 9, 6, 6, 18 ] }
false
InDevOne/NeuroLEW30sec
default
train
10,815
input
221
0.02043
1
2,190
33.68595
25
45.9375
{ "bin_edges": [ 1, 220, 439, 658, 877, 1096, 1315, 1534, 1753, 1972, 2190 ], "hist": [ 10530, 44, 13, 5, 1, 0, 0, 0, 0, 1 ] }
false
InDevOne/NeuroLEW30sec
default
train
10,815
output
251
0.02321
1
3,964
37.57033
29
52.95619
{ "bin_edges": [ 1, 398, 795, 1192, 1589, 1986, 2383, 2780, 3177, 3574, 3964 ], "hist": [ 10553, 8, 1, 1, 0, 0, 0, 0, 0, 1 ] }
false
MrbBakh/Dynabench
default
test
720
text
0
0
8
785
70.6375
65
43.20325
{ "bin_edges": [ 8, 86, 164, 242, 320, 398, 476, 554, 632, 710, 785 ], "hist": [ 504, 206, 9, 0, 0, 0, 0, 0, 0, 1 ] }
false
MrbBakh/Dynabench
default
train
13,065
text
0
0
5
282
65.15966
60
34.75497
{ "bin_edges": [ 5, 33, 61, 89, 117, 145, 173, 201, 229, 257, 282 ], "hist": [ 2522, 4145, 3304, 1909, 838, 291, 42, 9, 4, 1 ] }
false
MrbBakh/Dynabench
default
validation
720
text
0
0
7
197
68.54722
64
34.36393
{ "bin_edges": [ 7, 27, 47, 67, 87, 107, 127, 147, 167, 187, 197 ], "hist": [ 59, 166, 155, 145, 85, 62, 30, 13, 4, 1 ] }
false
MrbBakh/YelpV2
default
test
3,600
text
0
0
5
523
72.44833
64
45.55601
{ "bin_edges": [ 5, 57, 109, 161, 213, 265, 317, 369, 421, 473, 523 ], "hist": [ 1532, 1455, 465, 103, 30, 7, 4, 1, 1, 2 ] }
false
MrbBakh/YelpV2
default
validation
3,600
text
0
0
5
411
72.51667
64
43.43556
{ "bin_edges": [ 5, 46, 87, 128, 169, 210, 251, 292, 333, 374, 411 ], "hist": [ 1050, 1483, 692, 260, 72, 31, 6, 3, 1, 2 ] }
false
MrbBakh/YelpV2
default
train
80,488
text
0
0
4
1,479
69.25145
61
44.90802
{ "bin_edges": [ 4, 152, 300, 448, 596, 744, 892, 1040, 1188, 1336, 1479 ], "hist": [ 76623, 3737, 111, 7, 5, 0, 2, 1, 0, 2 ] }
false
chethan-mahindrakar/high_quality_instruction_backtranslation
default
train
20
instruction
0
0
132
8,047
1,901.1
1,278
1,861.79028
{ "bin_edges": [ 132, 924, 1716, 2508, 3300, 4092, 4884, 5676, 6468, 7260, 8047 ], "hist": [ 6, 8, 2, 0, 2, 1, 0, 0, 0, 1 ] }
false
chethan-mahindrakar/high_quality_instruction_backtranslation
default
train
20
response
0
0
64
1,812
822.9
765
706.27033
{ "bin_edges": [ 64, 239, 414, 589, 764, 939, 1114, 1289, 1464, 1639, 1812 ], "hist": [ 7, 2, 0, 1, 2, 0, 1, 1, 1, 5 ] }
false
ebayes/uhura-truthfulqa
am_generation
test
807
best_answer
0
0
2
103
36.67782
36
17.13978
{ "bin_edges": [ 2, 13, 24, 35, 46, 57, 68, 79, 90, 101, 103 ], "hist": [ 81, 91, 202, 212, 123, 60, 26, 7, 3, 2 ] }
false
ebayes/uhura-truthfulqa
am_generation
test
807
question
0
0
6
219
41.96406
36
24.56161
{ "bin_edges": [ 6, 28, 50, 72, 94, 116, 138, 160, 182, 204, 219 ], "hist": [ 196, 438, 110, 25, 18, 10, 5, 1, 1, 3 ] }
false
ebayes/uhura-truthfulqa
am_generation
test
807
source
0
0
3
221
60.36307
57
30.12444
{ "bin_edges": [ 3, 25, 47, 69, 91, 113, 135, 157, 179, 201, 221 ], "hist": [ 80, 152, 324, 160, 46, 30, 5, 4, 2, 4 ] }
false
ebayes/uhura-truthfulqa
sw_multiple_choice
test
808
question
0
0
13
351
64.59035
55.5
38.21023
{ "bin_edges": [ 13, 47, 81, 115, 149, 183, 217, 251, 285, 319, 351 ], "hist": [ 264, 389, 96, 30, 11, 8, 6, 1, 1, 2 ] }
false
ebayes/uhura-truthfulqa
nso_generation
test
809
question
0
0
7
362
73.06057
62
42.25459
{ "bin_edges": [ 7, 43, 79, 115, 151, 187, 223, 259, 295, 331, 362 ], "hist": [ 129, 442, 153, 39, 19, 17, 5, 1, 2, 2 ] }
false
ebayes/uhura-truthfulqa
yo_multiple_choice
test
809
question
0
0
9
378
67.66255
58
40.86216
{ "bin_edges": [ 9, 46, 83, 120, 157, 194, 231, 268, 305, 342, 378 ], "hist": [ 228, 421, 99, 24, 19, 12, 2, 0, 2, 2 ] }
false
ebayes/uhura-truthfulqa
ha_generation
test
791
best_answer
0
0
4
179
58.61315
58
28.38349
{ "bin_edges": [ 4, 22, 40, 58, 76, 94, 112, 130, 148, 166, 179 ], "hist": [ 91, 102, 196, 199, 112, 58, 24, 6, 2, 1 ] }
false
ebayes/uhura-truthfulqa
ha_generation
test
791
question
0
0
9
350
66.02276
57
37.69192
{ "bin_edges": [ 9, 44, 79, 114, 149, 184, 219, 254, 289, 324, 350 ], "hist": [ 193, 419, 113, 29, 20, 12, 1, 2, 1, 1 ] }
false
ebayes/uhura-truthfulqa
ha_generation
test
791
source
0
0
3
221
60.39949
57
29.80417
{ "bin_edges": [ 3, 25, 47, 69, 91, 113, 135, 157, 179, 201, 221 ], "hist": [ 78, 147, 317, 159, 46, 30, 5, 4, 2, 3 ] }
false
ebayes/uhura-truthfulqa
yo_generation
test
807
best_answer
0
0
5
197
59.64312
59
27.98376
{ "bin_edges": [ 5, 25, 45, 65, 85, 105, 125, 145, 165, 185, 197 ], "hist": [ 70, 176, 229, 190, 90, 38, 10, 3, 0, 1 ] }
false
ebayes/uhura-truthfulqa
yo_generation
test
807
question
0
0
9
378
67.65923
58
40.36135
{ "bin_edges": [ 9, 46, 83, 120, 157, 194, 231, 268, 305, 342, 378 ], "hist": [ 225, 420, 102, 24, 19, 12, 1, 0, 2, 2 ] }
false
ebayes/uhura-truthfulqa
yo_generation
test
807
source
0
0
3
221
60.48079
57
30.19399
{ "bin_edges": [ 3, 25, 47, 69, 91, 113, 135, 157, 179, 201, 221 ], "hist": [ 80, 149, 326, 159, 48, 30, 5, 4, 2, 4 ] }
false
ebayes/uhura-truthfulqa
am_multiple_choice
test
808
question
0
0
6
219
41.86139
36
24.52923
{ "bin_edges": [ 6, 28, 50, 72, 94, 116, 138, 160, 182, 204, 219 ], "hist": [ 196, 441, 109, 24, 18, 10, 5, 1, 1, 3 ] }
false
ebayes/uhura-truthfulqa
ha_multiple_choice
test
808
question
0
0
9
350
67.09901
58
39.28752
{ "bin_edges": [ 9, 44, 79, 114, 149, 184, 219, 254, 289, 324, 350 ], "hist": [ 194, 427, 114, 29, 23, 15, 1, 3, 1, 1 ] }
false
ebayes/uhura-truthfulqa
sw_generation
test
805
best_answer
0
0
0
132
55.19876
56
27.46903
{ "bin_edges": [ 0, 14, 28, 42, 56, 70, 84, 98, 112, 126, 132 ], "hist": [ 80, 59, 118, 138, 170, 116, 77, 25, 17, 5 ] }
false
ebayes/uhura-truthfulqa
sw_generation
test
805
question
0
0
13
351
64.78385
56
38.01104
{ "bin_edges": [ 13, 47, 81, 115, 149, 183, 217, 251, 285, 319, 351 ], "hist": [ 259, 388, 100, 30, 11, 7, 6, 1, 1, 2 ] }
false
ebayes/uhura-truthfulqa
sw_generation
test
805
source
0
0
3
221
60.35528
57
30.12665
{ "bin_edges": [ 3, 25, 47, 69, 91, 113, 135, 157, 179, 201, 221 ], "hist": [ 80, 151, 323, 160, 47, 29, 5, 4, 2, 4 ] }
false
ebayes/uhura-truthfulqa
nso_multiple_choice
test
809
question
0
0
7
362
73.06057
62
42.25459
{ "bin_edges": [ 7, 43, 79, 115, 151, 187, 223, 259, 295, 331, 362 ], "hist": [ 129, 442, 153, 39, 19, 17, 5, 1, 2, 2 ] }
false
Egbertjing/pubmed
default
train
19,717
text
0
0
18
5,732
1,649.25191
1,669
478.3011
{ "bin_edges": [ 18, 590, 1162, 1734, 2306, 2878, 3450, 4022, 4594, 5166, 5732 ], "hist": [ 221, 2779, 8079, 7279, 1123, 179, 41, 12, 2, 2 ] }
false
Egbertjing/pubmed
default
test
1,972
text
0
0
327
4,142
1,640.37272
1,649.5
476.74826
{ "bin_edges": [ 327, 709, 1091, 1473, 1855, 2237, 2619, 3001, 3383, 3765, 4142 ], "hist": [ 47, 188, 430, 720, 427, 103, 37, 10, 6, 4 ] }
false
tejeshbhalla/func-calling-final
default
train
1,001
tools
0
0
2
4,993
569.12488
174
798.80662
{ "bin_edges": [ 2, 502, 1002, 1502, 2002, 2502, 3002, 3502, 4002, 4502, 4993 ], "hist": [ 609, 148, 114, 65, 29, 20, 10, 2, 3, 1 ] }
false
fractalego/stackoverflow_python
default
train
20,731
text
0
0
226
34,506
2,690.82611
2,068
2,369.355
{ "bin_edges": [ 226, 3655, 7084, 10513, 13942, 17371, 20800, 24229, 27658, 31087, 34506 ], "hist": [ 16748, 3116, 562, 163, 68, 28, 12, 16, 14, 4 ] }
false
danaarad/ioi_dataset
default
test
100
text
0
0
76
96
85.2
85
4.96859
{ "bin_edges": [ 76, 79, 82, 85, 88, 91, 94, 96 ], "hist": [ 14, 8, 20, 22, 26, 2, 8 ] }
false
danaarad/ioi_dataset
default
train
100
text
0
0
67
113
90.64
92
13.38741
{ "bin_edges": [ 67, 72, 77, 82, 87, 92, 97, 102, 107, 112, 113 ], "hist": [ 8, 12, 18, 4, 8, 12, 10, 16, 8, 4 ] }
false
danaarad/ioi_dataset
default
validation
100
text
0
0
67
94
79.6
79.5
7.69363
{ "bin_edges": [ 67, 70, 73, 76, 79, 82, 85, 88, 91, 94, 94 ], "hist": [ 12, 10, 14, 12, 12, 6, 16, 10, 6, 2 ] }
false
5CD-AI/Viet-Doc-VQA-II
default
train
53,340
description
0
0
42
5,552
960.89436
826
512.98071
{ "bin_edges": [ 42, 594, 1146, 1698, 2250, 2802, 3354, 3906, 4458, 5010, 5552 ], "hist": [ 14034, 24399, 9549, 3977, 1165, 172, 34, 6, 1, 3 ] }
true
acidz1/uncle_slava
default
train
49,558
text
0
0
0
4,095
186.57058
139
198.48334
{ "bin_edges": [ 0, 410, 820, 1230, 1640, 2050, 2460, 2870, 3280, 3690, 4095 ], "hist": [ 43807, 5143, 502, 71, 22, 5, 1, 5, 0, 2 ] }
false
Cheremy/processed_arguna_queries
default
train
1,406
id
0
0
23
41
31.73542
31
3.59612
{ "bin_edges": [ 23, 25, 27, 29, 31, 33, 35, 37, 39, 41, 41 ], "hist": [ 14, 75, 117, 345, 403, 106, 181, 85, 67, 13 ] }
false
Cheremy/processed_arguna_queries
default
train
1,406
question
0
0
251
5,500
1,192.72048
1,077
568.15295
{ "bin_edges": [ 251, 776, 1301, 1826, 2351, 2876, 3401, 3926, 4451, 4976, 5500 ], "hist": [ 288, 666, 309, 92, 26, 12, 7, 0, 5, 1 ] }
false
Cheremy/processed_cqadupstack_chunked
default
train
40,221
text
0
0
18
6,437
431.22033
322
376.16284
{ "bin_edges": [ 18, 660, 1302, 1944, 2586, 3228, 3870, 4512, 5154, 5796, 6437 ], "hist": [ 33240, 5623, 1016, 235, 66, 17, 10, 7, 5, 2 ] }
false
Cheremy/processed_cqadupstack_queries
default
train
1,570
question
0
0
15
149
48.32994
45
20.44463
{ "bin_edges": [ 15, 29, 43, 57, 71, 85, 99, 113, 127, 141, 149 ], "hist": [ 266, 419, 414, 252, 124, 65, 20, 6, 2, 2 ] }
false
Cheremy/processed_scidocs_queries
default
train
1,000
id
0
0
40
40
40
40
0
{ "bin_edges": [ 40, 40 ], "hist": [ 1000 ] }
false
Cheremy/processed_scidocs_queries
default
train
1,000
question
0
0
16
206
71.632
69
24.09866
{ "bin_edges": [ 16, 36, 56, 76, 96, 116, 136, 156, 176, 196, 206 ], "hist": [ 41, 222, 331, 275, 86, 31, 10, 0, 3, 1 ] }
false
Cheremy/processed_scidocs_chunked
default
train
25,658
id
0
0
40
40
40
40
0
{ "bin_edges": [ 40, 40 ], "hist": [ 25658 ] }
false
Cheremy/processed_scidocs_chunked
default
train
25,658
text
344
0.01341
1
10,000
1,147.00411
1,034
750.38179
{ "bin_edges": [ 1, 1001, 2001, 3001, 4001, 5001, 6001, 7001, 8001, 9001, 10000 ], "hist": [ 11744, 12360, 883, 108, 58, 44, 21, 12, 11, 73 ] }
false
Cheremy/processed_scifact_chunked
default
train
5,183
text
0
0
174
10,000
1,400.91588
1,329
566.79252
{ "bin_edges": [ 174, 1157, 2140, 3123, 4106, 5089, 6072, 7055, 8038, 9021, 10000 ], "hist": [ 1950, 2843, 345, 30, 8, 3, 2, 0, 0, 2 ] }
false
Cheremy/processed_scifact_queries
default
train
1,109
question
0
0
26
249
89.06402
82
33.46266
{ "bin_edges": [ 26, 49, 72, 95, 118, 141, 164, 187, 210, 233, 249 ], "hist": [ 92, 290, 312, 196, 117, 74, 21, 6, 0, 1 ] }
false
davidguzmanr/AfriInstruct-language-split
eng-xho
train
1,470
instruction
0
0
84
473
196.27687
191
49.21345
{ "bin_edges": [ 84, 123, 162, 201, 240, 279, 318, 357, 396, 435, 473 ], "hist": [ 57, 289, 522, 355, 163, 59, 15, 4, 3, 3 ] }
false
davidguzmanr/AfriInstruct-language-split
eng-xho
train
1,470
output
0
0
18
395
135.6517
130
53.3048
{ "bin_edges": [ 18, 56, 94, 132, 170, 208, 246, 284, 322, 360, 395 ], "hist": [ 61, 260, 432, 368, 215, 94, 23, 8, 5, 4 ] }
false
davidguzmanr/AfriInstruct-language-split
eng-xho
validation
253
instruction
0
0
85
384
195.32016
195
58.93819
{ "bin_edges": [ 85, 115, 145, 175, 205, 235, 265, 295, 325, 355, 384 ], "hist": [ 17, 43, 38, 44, 48, 33, 19, 6, 1, 4 ] }
false