dataset
stringlengths
5
115
config
stringlengths
1
162
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
0
77.9k
null_count
int64
0
62.9M
null_proportion
float64
0
1
min
int64
0
9.25M
max
int64
0
1.07B
mean
float64
0
90.4M
median
float64
0
80.1M
std
float64
0
130M
histogram
dict
partial
bool
2 classes
davidguzmanr/AfriInstruct-language-split
yor-eng
train
13,479
output
0
0
6
1,080
113.51183
103
69.50593
{ "bin_edges": [ 6, 114, 222, 330, 438, 546, 654, 762, 870, 978, 1080 ], "hist": [ 7668, 4956, 729, 97, 18, 6, 0, 1, 0, 4 ] }
false
davidguzmanr/AfriInstruct-language-split
tmh
train
41,210
instruction
0
0
12
8,511
215.26821
89
405.57026
{ "bin_edges": [ 12, 862, 1712, 2562, 3412, 4262, 5112, 5962, 6812, 7662, 8511 ], "hist": [ 39312, 574, 1206, 85, 21, 8, 1, 2, 0, 1 ] }
false
davidguzmanr/AfriInstruct-language-split
tmh
train
41,210
output
0
0
0
4,873
196.95278
186
180.61471
{ "bin_edges": [ 0, 488, 976, 1464, 1952, 2440, 2928, 3416, 3904, 4392, 4873 ], "hist": [ 39469, 1435, 230, 46, 11, 3, 5, 8, 0, 3 ] }
false
davidguzmanr/AfriInstruct-language-split
zul-fra
train
1,025
instruction
0
0
95
474
208.73268
203
50.86744
{ "bin_edges": [ 95, 133, 171, 209, 247, 285, 323, 361, 399, 437, 474 ], "hist": [ 37, 209, 308, 250, 146, 49, 17, 6, 2, 1 ] }
false
davidguzmanr/AfriInstruct-language-split
zul-fra
train
1,025
output
0
0
38
396
152.65561
144
53.58189
{ "bin_edges": [ 38, 74, 110, 146, 182, 218, 254, 290, 326, 362, 396 ], "hist": [ 30, 180, 312, 238, 155, 64, 25, 11, 5, 5 ] }
false
rakibulmuhajir/sample
default
train
65
description
0
0
107
6,142
1,369.96923
879
1,142.07551
{ "bin_edges": [ 107, 711, 1315, 1919, 2523, 3127, 3731, 4335, 4939, 5543, 6142 ], "hist": [ 27, 13, 5, 12, 4, 2, 1, 0, 0, 1 ] }
false
rakibulmuhajir/sample
default
train
65
product_id
0
0
12
13
12.2
12
0.40311
{ "bin_edges": [ 12, 13, 13 ], "hist": [ 52, 13 ] }
false
rakibulmuhajir/sample
default
train
65
text_for_embedding
0
0
142
6,185
1,412.89231
898
1,141.0391
{ "bin_edges": [ 142, 747, 1352, 1957, 2562, 3167, 3772, 4377, 4982, 5587, 6185 ], "hist": [ 28, 12, 5, 12, 4, 2, 1, 0, 0, 1 ] }
false
rakibulmuhajir/sample
default
train
65
title
0
0
23
88
49.87692
46
12.80687
{ "bin_edges": [ 23, 30, 37, 44, 51, 58, 65, 72, 79, 86, 88 ], "hist": [ 2, 8, 9, 20, 12, 3, 8, 1, 0, 2 ] }
false
HorizonNexusAI/Alpaca-gpt4
default
train
52,002
input
31,344
0.60275
1
2,467
57.67979
36
79.50883
{ "bin_edges": [ 1, 248, 495, 742, 989, 1236, 1483, 1730, 1977, 2224, 2467 ], "hist": [ 20018, 544, 82, 3, 0, 7, 3, 0, 0, 1 ] }
false
HorizonNexusAI/Alpaca-gpt4
default
train
52,002
instruction
0
0
9
489
59.77876
57
21.19359
{ "bin_edges": [ 9, 58, 107, 156, 205, 254, 303, 352, 401, 450, 489 ], "hist": [ 26897, 23625, 1325, 121, 20, 3, 2, 2, 4, 3 ] }
false
HorizonNexusAI/Alpaca-gpt4
default
train
52,002
output
0
0
1
4,574
677.36897
471
649.75951
{ "bin_edges": [ 1, 459, 917, 1375, 1833, 2291, 2749, 3207, 3665, 4123, 4574 ], "hist": [ 25705, 9781, 7189, 5893, 2655, 686, 91, 1, 0, 1 ] }
false
HorizonNexusAI/Alpaca-gpt4
default
train
52,002
text
0
0
166
4,929
925.91029
733
633.20557
{ "bin_edges": [ 166, 643, 1120, 1597, 2074, 2551, 3028, 3505, 3982, 4459, 4929 ], "hist": [ 23605, 11338, 7751, 6084, 2525, 636, 61, 1, 0, 1 ] }
false
ajaykarthick/wavefake-audio
default
train
17,600
audio_id
0
0
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 17600 ] }
true
adamo1139/TURTLE_v2_rated
default
train
4,136
chosen
0
0
1
2,050
378.00701
275
368.52549
{ "bin_edges": [ 1, 206, 411, 616, 821, 1026, 1231, 1436, 1641, 1846, 2050 ], "hist": [ 1731, 871, 746, 352, 188, 86, 55, 36, 37, 34 ] }
false
adamo1139/TURTLE_v2_rated
default
train
4,136
comments
0
0
46
3,357
520.85638
489
199.58991
{ "bin_edges": [ 46, 378, 710, 1042, 1374, 1706, 2038, 2370, 2702, 3034, 3357 ], "hist": [ 1021, 2522, 518, 64, 8, 1, 1, 0, 0, 1 ] }
false
adamo1139/TURTLE_v2_rated
default
train
4,136
prompt
0
0
1
1,987
563.45793
511
399.83597
{ "bin_edges": [ 1, 200, 399, 598, 797, 996, 1195, 1394, 1593, 1792, 1987 ], "hist": [ 839, 756, 913, 687, 365, 224, 148, 101, 81, 22 ] }
false
adamo1139/TURTLE_v2_rated
default
train
4,136
rejected
0
0
31
6,465
884.20358
694
570.13802
{ "bin_edges": [ 31, 675, 1319, 1963, 2607, 3251, 3895, 4539, 5183, 5827, 6465 ], "hist": [ 2002, 1253, 674, 173, 26, 3, 1, 1, 1, 2 ] }
false
bluejwu/liberal-and-conservative-news
default
train
1,661,315
text
0
0
0
31,510
114.98953
50
173.95553
{ "bin_edges": [ 0, 3152, 6304, 9456, 12608, 15760, 18912, 22064, 25216, 28368, 31510 ], "hist": [ 1661231, 35, 15, 6, 2, 6, 16, 0, 0, 4 ] }
false
adamo1139/hesoyam_03_rated1_validjson_allclassified2
default
train
87,262
comments
0
0
0
3,519
455.35836
413
187.15124
{ "bin_edges": [ 0, 352, 704, 1056, 1408, 1760, 2112, 2464, 2816, 3168, 3519 ], "hist": [ 27294, 51850, 7122, 823, 132, 27, 7, 3, 3, 1 ] }
false
DrishtiSharma/MGTD
version v0.1
Japanese_rewritten
32,764
Modified text
0
0
21
27,504
605.59709
223
826.26977
{ "bin_edges": [ 21, 2770, 5519, 8268, 11017, 13766, 16515, 19264, 22013, 24762, 27504 ], "hist": [ 31960, 733, 51, 12, 6, 1, 0, 0, 0, 1 ] }
false
DrishtiSharma/MGTD
version v0.1
Japanese_rewritten
32,764
Original text
0
0
180
27,504
612.76392
222
839.40356
{ "bin_edges": [ 180, 2913, 5646, 8379, 11112, 13845, 16578, 19311, 22044, 24777, 27504 ], "hist": [ 32049, 642, 53, 12, 6, 1, 0, 0, 0, 1 ] }
false
DrishtiSharma/MGTD
version v0.1
Japanese_partial
32,764
Modified text
2
0.00006
21
27,504
461.22602
237
656.18978
{ "bin_edges": [ 21, 2770, 5519, 8268, 11017, 13766, 16515, 19264, 22013, 24762, 27504 ], "hist": [ 32309, 414, 27, 7, 2, 2, 0, 0, 0, 1 ] }
false
DrishtiSharma/MGTD
version v0.1
Japanese_partial
32,764
Original text
0
0
180
27,504
612.76392
222
839.40356
{ "bin_edges": [ 180, 2913, 5646, 8379, 11112, 13845, 16578, 19311, 22044, 24777, 27504 ], "hist": [ 32049, 642, 53, 12, 6, 1, 0, 0, 0, 1 ] }
false
DrishtiSharma/MGTD
version v0.1
Arabic_partial
93,686
Modified text
0
0
11
47,043
1,761.90028
1,302
1,622.16922
{ "bin_edges": [ 11, 4715, 9419, 14123, 18827, 23531, 28235, 32939, 37643, 42347, 47043 ], "hist": [ 88514, 4738, 394, 32, 4, 1, 1, 1, 0, 1 ] }
false
DrishtiSharma/MGTD
version v0.1
Arabic_partial
93,686
Original text
0
0
11
47,043
1,972.76038
1,531
1,692.93706
{ "bin_edges": [ 11, 4715, 9419, 14123, 18827, 23531, 28235, 32939, 37643, 42347, 47043 ], "hist": [ 87450, 5713, 477, 39, 4, 1, 1, 0, 0, 1 ] }
false
yushengsu/PersonaHub_modified
instruction
train
50,000
input persona
0
0
7
501
151.18162
146
42.48267
{ "bin_edges": [ 7, 57, 107, 157, 207, 257, 307, 357, 407, 457, 501 ], "hist": [ 10, 5873, 24676, 14895, 3283, 940, 256, 54, 7, 6 ] }
false
yushengsu/PersonaHub_modified
instruction
train
50,000
synthesized text
0
0
15
7,772
277.05522
261
142.82274
{ "bin_edges": [ 15, 791, 1567, 2343, 3119, 3895, 4671, 5447, 6223, 6999, 7772 ], "hist": [ 49575, 363, 49, 7, 1, 2, 0, 1, 1, 1 ] }
false
yushengsu/PersonaHub_modified
knowledge
train
10,000
input persona
0
0
48
435
149.8099
145
41.53798
{ "bin_edges": [ 48, 87, 126, 165, 204, 243, 282, 321, 360, 399, 435 ], "hist": [ 271, 2648, 4064, 2108, 566, 214, 96, 29, 1, 3 ] }
false
yushengsu/PersonaHub_modified
knowledge
train
10,000
synthesized text
0
0
122
6,966
3,838.6109
3,813
446.87208
{ "bin_edges": [ 122, 807, 1492, 2177, 2862, 3547, 4232, 4917, 5602, 6287, 6966 ], "hist": [ 4, 0, 1, 59, 2554, 5601, 1628, 141, 11, 1 ] }
false
yushengsu/PersonaHub_modified
math
train
50,000
input persona
0
0
5
350
96.87264
96
23.69988
{ "bin_edges": [ 5, 40, 75, 110, 145, 180, 215, 250, 285, 320, 350 ], "hist": [ 840, 5896, 30160, 11855, 1102, 113, 21, 6, 3, 4 ] }
false
yushengsu/PersonaHub_modified
math
train
50,000
synthesized text
0
0
73
7,665
790.29064
768
245.0475
{ "bin_edges": [ 73, 833, 1593, 2353, 3113, 3873, 4633, 5393, 6153, 6913, 7665 ], "hist": [ 30429, 19368, 197, 5, 0, 0, 0, 0, 0, 1 ] }
false
yushengsu/PersonaHub_modified
npc
train
10,000
input persona
0
0
19
252
97.6905
97
20.82201
{ "bin_edges": [ 19, 43, 67, 91, 115, 139, 163, 187, 211, 235, 252 ], "hist": [ 192, 141, 3279, 4557, 1528, 263, 35, 2, 1, 2 ] }
false
yushengsu/PersonaHub_modified
npc
train
10,000
synthesized text
0
0
258
3,120
1,064.1063
604
706.2905
{ "bin_edges": [ 258, 545, 832, 1119, 1406, 1693, 1980, 2267, 2554, 2841, 3120 ], "hist": [ 3674, 2423, 113, 206, 655, 1302, 1094, 431, 93, 9 ] }
false
yushengsu/PersonaHub_modified
persona
train
200,000
persona
0
0
10
750
91.70908
92
26.90224
{ "bin_edges": [ 10, 85, 160, 235, 310, 385, 460, 535, 610, 685, 750 ], "hist": [ 72210, 125595, 2010, 132, 37, 11, 3, 1, 0, 1 ] }
false
yushengsu/PersonaHub_modified
reasoning
train
50,000
input persona
0
0
3
396
92.7893
93
26.81045
{ "bin_edges": [ 3, 43, 83, 123, 163, 203, 243, 283, 323, 363, 396 ], "hist": [ 1915, 13534, 29330, 4734, 371, 78, 19, 7, 9, 3 ] }
false
yushengsu/PersonaHub_modified
reasoning
train
50,000
synthesized text
0
0
15
5,083
545.72536
641
442.06612
{ "bin_edges": [ 15, 522, 1029, 1536, 2043, 2550, 3057, 3564, 4071, 4578, 5083 ], "hist": [ 21230, 21816, 6633, 309, 11, 0, 0, 0, 0, 1 ] }
false
yushengsu/PersonaHub_modified
tool
train
5,000
input persona
0
0
5
425
88.4174
89
30.03834
{ "bin_edges": [ 5, 48, 91, 134, 177, 220, 263, 306, 349, 392, 425 ], "hist": [ 426, 2210, 2116, 203, 27, 11, 5, 1, 0, 1 ] }
false
yushengsu/PersonaHub_modified
tool
train
5,000
synthesized text
0
0
69
1,794
545.9148
523
139.7516
{ "bin_edges": [ 69, 242, 415, 588, 761, 934, 1107, 1280, 1453, 1626, 1794 ], "hist": [ 4, 738, 2648, 1244, 290, 64, 10, 1, 0, 1 ] }
false
Vijish/scphotos
default
train
1,000
en_text
0
0
178
1,023
755.878
791
122.97186
{ "bin_edges": [ 178, 263, 348, 433, 518, 603, 688, 773, 858, 943, 1023 ], "hist": [ 4, 4, 20, 44, 57, 59, 212, 477, 113, 10 ] }
false
isavita/advent-of-code
default
train
8,153
solution
0
0
7
10,709
1,355.96001
1,091
988.05456
{ "bin_edges": [ 7, 1078, 2149, 3220, 4291, 5362, 6433, 7504, 8575, 9646, 10709 ], "hist": [ 4013, 2846, 888, 259, 85, 33, 21, 7, 0, 1 ] }
false
rakibulmuhajir/secprousa
default
train
8,987
description
0
0
0
33,053
913.13252
490
1,358.33131
{ "bin_edges": [ 0, 3306, 6612, 9918, 13224, 16530, 19836, 23142, 26448, 29754, 33053 ], "hist": [ 8527, 363, 85, 5, 5, 0, 0, 0, 0, 2 ] }
false
rakibulmuhajir/secprousa
default
train
8,987
product_id
0
0
12
13
12.62234
13
0.48483
{ "bin_edges": [ 12, 13, 13 ], "hist": [ 3394, 5593 ] }
false
rakibulmuhajir/secprousa
default
train
8,987
text_for_embedding
0
0
6
32,981
949.76166
524
1,356.62456
{ "bin_edges": [ 6, 3304, 6602, 9900, 13198, 16496, 19794, 23092, 26390, 29688, 32981 ], "hist": [ 8523, 367, 85, 5, 5, 0, 0, 0, 0, 2 ] }
false
rakibulmuhajir/secprousa
default
train
8,987
title
0
0
2
183
40.1984
37
18.58062
{ "bin_edges": [ 2, 21, 40, 59, 78, 97, 116, 135, 154, 173, 183 ], "hist": [ 1037, 3986, 2794, 779, 293, 75, 19, 3, 0, 1 ] }
false
adamo1139/TURTLE_v3
default
train
2,228
chosen
0
0
3
1,989
432.40305
346.5
366.87574
{ "bin_edges": [ 3, 202, 401, 600, 799, 998, 1197, 1396, 1595, 1794, 1989 ], "hist": [ 693, 552, 447, 240, 130, 64, 37, 20, 19, 26 ] }
false
adamo1139/TURTLE_v3
default
train
2,228
prompt
0
0
1
1,987
521.80835
445
388.00076
{ "bin_edges": [ 1, 200, 399, 598, 797, 996, 1195, 1394, 1593, 1792, 1987 ], "hist": [ 506, 498, 465, 309, 180, 104, 83, 40, 31, 12 ] }
false
adamo1139/TURTLE_v3
default
train
2,228
rejected
0
0
34
6,465
879.12792
697.5
555.54216
{ "bin_edges": [ 34, 678, 1322, 1966, 2610, 3254, 3898, 4542, 5186, 5830, 6465 ], "hist": [ 1079, 702, 342, 89, 12, 2, 0, 1, 0, 1 ] }
false
Anvithah/new-ah-lima-dataset
default
train
71
instruction
0
0
31
14,225
317.14085
101
1,676.17836
{ "bin_edges": [ 31, 1451, 2871, 4291, 5711, 7131, 8551, 9971, 11391, 12811, 14225 ], "hist": [ 70, 0, 0, 0, 0, 0, 0, 0, 0, 1 ] }
false
Anvithah/new-ah-lima-dataset
default
train
71
output
0
0
82
11,550
2,328.57746
1,584
2,254.66887
{ "bin_edges": [ 82, 1229, 2376, 3523, 4670, 5817, 6964, 8111, 9258, 10405, 11550 ], "hist": [ 26, 23, 6, 7, 4, 1, 1, 2, 0, 1 ] }
false
Avinier/ssllmv2-conversations
default
train
2,452
id
0
0
1
4
3.54853
4
0.58585
{ "bin_edges": [ 1, 2, 3, 4, 4 ], "hist": [ 9, 90, 900, 1453 ] }
false
HorizonNexusAI/Alpaca-gpt4-combined-1
default
train
52,002
input
31,344
0.60275
1
2,467
57.67979
36
79.50883
{ "bin_edges": [ 1, 248, 495, 742, 989, 1236, 1483, 1730, 1977, 2224, 2467 ], "hist": [ 20018, 544, 82, 3, 0, 7, 3, 0, 0, 1 ] }
false
HorizonNexusAI/Alpaca-gpt4-combined-1
default
train
52,002
instruction
0
0
9
489
59.77876
57
21.19359
{ "bin_edges": [ 9, 58, 107, 156, 205, 254, 303, 352, 401, 450, 489 ], "hist": [ 26897, 23625, 1325, 121, 20, 3, 2, 2, 4, 3 ] }
false
HorizonNexusAI/Alpaca-gpt4-combined-1
default
train
52,002
output
0
0
1
4,574
677.36897
471
649.75951
{ "bin_edges": [ 1, 459, 917, 1375, 1833, 2291, 2749, 3207, 3665, 4123, 4574 ], "hist": [ 25705, 9781, 7189, 5893, 2655, 686, 91, 1, 0, 1 ] }
false
HorizonNexusAI/Alpaca-gpt4-combined-1
default
train
52,002
text
0
0
166
4,929
925.91029
733
633.20557
{ "bin_edges": [ 166, 643, 1120, 1597, 2074, 2551, 3028, 3505, 3982, 4459, 4929 ], "hist": [ 23605, 11338, 7751, 6084, 2525, 636, 61, 1, 0, 1 ] }
false
Ganz00/Cleaned_ELI5_with_one_response
default
train
91,772
q_id
0
0
6
6
6
6
0
{ "bin_edges": [ 6, 6 ], "hist": [ 91772 ] }
false
Ganz00/Cleaned_ELI5_with_one_response
default
train
91,772
question
0
0
5
4,601
161.54452
107
162.00557
{ "bin_edges": [ 5, 465, 925, 1385, 1845, 2305, 2765, 3225, 3685, 4145, 4601 ], "hist": [ 87142, 4135, 375, 81, 25, 8, 0, 5, 0, 1 ] }
false
Ganz00/Cleaned_ELI5_with_one_response
default
train
91,772
reponses
0
0
28
9,829
608.29597
429
600.22112
{ "bin_edges": [ 28, 1009, 1990, 2971, 3952, 4933, 5914, 6895, 7876, 8857, 9829 ], "hist": [ 77829, 10922, 2054, 594, 224, 75, 39, 11, 14, 10 ] }
false
Ganz00/Cleaned_ELI5_with_one_response
default
test
5,411
q_id
0
0
6
6
6
6
0
{ "bin_edges": [ 6, 6 ], "hist": [ 5411 ] }
false
Ganz00/Cleaned_ELI5_with_one_response
default
test
5,411
question
0
0
11
2,698
183.65792
121
179.23543
{ "bin_edges": [ 11, 280, 549, 818, 1087, 1356, 1625, 1894, 2163, 2432, 2698 ], "hist": [ 4366, 821, 163, 34, 14, 7, 3, 2, 0, 1 ] }
false
Ganz00/Cleaned_ELI5_with_one_response
default
test
5,411
reponses
0
0
30
8,810
613.36795
439
592.08824
{ "bin_edges": [ 30, 909, 1788, 2667, 3546, 4425, 5304, 6183, 7062, 7941, 8810 ], "hist": [ 4416, 770, 148, 45, 18, 7, 3, 3, 0, 1 ] }
false
Ganz00/Cleaned_ELI5_with_one_response
default
validation
7,821
q_id
0
0
6
6
6
6
0
{ "bin_edges": [ 6, 6 ], "hist": [ 7821 ] }
false
Ganz00/Cleaned_ELI5_with_one_response
default
validation
7,821
question
0
0
5
1,834
143.70464
94
153.32106
{ "bin_edges": [ 5, 188, 371, 554, 737, 920, 1103, 1286, 1469, 1652, 1834 ], "hist": [ 6226, 1066, 303, 145, 29, 25, 13, 9, 1, 4 ] }
false
Ganz00/Cleaned_ELI5_with_one_response
default
validation
7,821
reponses
0
0
36
8,589
653.96535
448
673.84633
{ "bin_edges": [ 36, 892, 1748, 2604, 3460, 4316, 5172, 6028, 6884, 7740, 8589 ], "hist": [ 6142, 1224, 277, 106, 39, 14, 9, 6, 3, 1 ] }
false
Cheremy/alibaba_arguna_chunked
default
train
8,674
id
0
0
23
47
34.81058
35
4.47056
{ "bin_edges": [ 23, 26, 29, 32, 35, 38, 41, 44, 47, 47 ], "hist": [ 62, 458, 1713, 2075, 1945, 1506, 606, 297, 12 ] }
false
Anvithah/ah_high_quality_lima_dataset
default
train
33
instruction
0
0
31
350
131.48485
102
91.28462
{ "bin_edges": [ 31, 63, 95, 127, 159, 191, 223, 255, 287, 319, 350 ], "hist": [ 10, 3, 8, 3, 2, 2, 1, 0, 1, 3 ] }
false
Anvithah/ah_high_quality_lima_dataset
default
train
33
output
0
0
384
8,656
3,016.66667
2,296
2,003.80882
{ "bin_edges": [ 384, 1212, 2040, 2868, 3696, 4524, 5352, 6180, 7008, 7836, 8656 ], "hist": [ 4, 10, 6, 1, 5, 3, 2, 0, 0, 2 ] }
false
MrDimitri/georgian-quora-faq
default
train
5,605
answer
0
0
1
6,985
498.99411
332
476.0096
{ "bin_edges": [ 1, 700, 1399, 2098, 2797, 3496, 4195, 4894, 5593, 6292, 6985 ], "hist": [ 4049, 1163, 390, 2, 0, 0, 0, 0, 0, 1 ] }
false
MrDimitri/georgian-quora-faq
default
train
5,605
question
0
0
2
405
83.01195
65
58.75403
{ "bin_edges": [ 2, 43, 84, 125, 166, 207, 248, 289, 330, 371, 405 ], "hist": [ 1132, 2625, 842, 476, 204, 196, 50, 52, 26, 2 ] }
false
MrDimitri/georgian-wikipedia
default
train
28,065
summary
0
0
0
5,659
293.54142
233
234.69643
{ "bin_edges": [ 0, 566, 1132, 1698, 2264, 2830, 3396, 3962, 4528, 5094, 5659 ], "hist": [ 25147, 2596, 271, 42, 6, 2, 0, 0, 0, 1 ] }
false
MrDimitri/georgian-wikipedia
default
train
28,065
text
0
0
0
157,322
1,363.37481
307
4,386.43975
{ "bin_edges": [ 0, 15733, 31466, 47199, 62932, 78665, 94398, 110131, 125864, 141597, 157322 ], "hist": [ 27760, 180, 75, 29, 10, 4, 2, 2, 2, 1 ] }
false
MrDimitri/georgian-wikipedia
default
train
28,065
title
0
0
1
89
15.90572
14
8.99274
{ "bin_edges": [ 1, 10, 19, 28, 37, 46, 55, 64, 73, 82, 89 ], "hist": [ 7183, 12775, 4911, 2322, 619, 188, 43, 18, 4, 2 ] }
false
OpenDatasets/dalle-3-dataset
default
train
3,124
caption
0
0
0
1,183
210.36908
178
154.5877
{ "bin_edges": [ 0, 119, 238, 357, 476, 595, 714, 833, 952, 1071, 1183 ], "hist": [ 1113, 817, 662, 339, 148, 27, 5, 3, 9, 1 ] }
true
OpenDatasets/dalle-3-dataset
default
train
3,124
image_hash
0
0
32
32
32
32
0
{ "bin_edges": [ 32, 32 ], "hist": [ 3124 ] }
true
OpenDatasets/dalle-3-dataset
default
train
3,124
link
0
0
177
355
232.53873
212
59.30407
{ "bin_edges": [ 177, 195, 213, 231, 249, 267, 285, 303, 321, 339, 355 ], "hist": [ 634, 1754, 109, 20, 1, 1, 2, 8, 5, 590 ] }
true
OpenDatasets/dalle-3-dataset
default
train
3,124
message_id
0
0
19
19
19
19
0
{ "bin_edges": [ 19, 19 ], "hist": [ 3124 ] }
true
OpenDatasets/dalle-3-dataset
default
train
3,124
synthetic_caption
0
0
331
2,291
1,083.41645
1,086
327.54334
{ "bin_edges": [ 331, 528, 725, 922, 1119, 1316, 1513, 1710, 1907, 2104, 2291 ], "hist": [ 42, 473, 603, 542, 701, 465, 201, 65, 24, 8 ] }
true
OpenDatasets/dalle-3-dataset
default
train
3,124
timestamp
0
0
25
32
31.99104
32
0.25036
{ "bin_edges": [ 25, 26, 27, 28, 29, 30, 31, 32, 32 ], "hist": [ 4, 0, 0, 0, 0, 0, 0, 3120 ] }
true
Cheremy/alibaba_scidocs_chunked
default
train
25,314
id
0
0
40
40
40
40
0
{ "bin_edges": [ 40, 40 ], "hist": [ 25314 ] }
false
yushengsu/fineweb_edu_cleaned_modified
default
train
1,057,694
file_path
0
0
125
138
129.42177
126
5.44177
{ "bin_edges": [ 125, 127, 129, 131, 133, 135, 137, 138 ], "hist": [ 745317, 0, 0, 0, 0, 0, 312377 ] }
true
yushengsu/fineweb_edu_cleaned_modified
default
train
1,057,694
id
0
0
47
47
47
47
0
{ "bin_edges": [ 47, 47 ], "hist": [ 1057694 ] }
true
yushengsu/fineweb_edu_cleaned_modified
default
train
1,057,694
text
0
0
195
587,236
4,756.49678
2,904
8,309.7548
{ "bin_edges": [ 195, 58900, 117605, 176310, 235015, 293720, 352425, 411130, 469835, 528540, 587236 ], "hist": [ 1054013, 3082, 382, 107, 50, 24, 14, 12, 6, 4 ] }
true
yushengsu/fineweb_edu_cleaned_modified
default
train
1,057,694
url
0
0
14
2,283
72.11426
67
29.17888
{ "bin_edges": [ 14, 241, 468, 695, 922, 1149, 1376, 1603, 1830, 2057, 2283 ], "hist": [ 1056506, 1054, 83, 26, 10, 4, 6, 1, 2, 2 ] }
true
yushengsu/stack-v2-python-with-content-chunk1-modified
default
train
113,619
directory_id
0
0
40
40
40
40
0
{ "bin_edges": [ 40, 40 ], "hist": [ 113619 ] }
true
yushengsu/stack-v2-python-with-content-chunk1-modified
default
train
113,619
repo_name
0
0
5
114
24.73491
23
8.93652
{ "bin_edges": [ 5, 16, 27, 38, 49, 60, 71, 82, 93, 104, 114 ], "hist": [ 12881, 61126, 30639, 6890, 1455, 403, 129, 64, 19, 13 ] }
true
yushengsu/stack-v2-python-with-content-chunk1-modified
default
train
113,619
repo_url
0
0
24
133
43.73493
42
8.93653
{ "bin_edges": [ 24, 35, 46, 57, 68, 79, 90, 101, 112, 123, 133 ], "hist": [ 12881, 61126, 30639, 6890, 1455, 403, 129, 64, 19, 13 ] }
true
yushengsu/stack-v2-python-with-content-chunk1-modified
default
train
113,619
revision_id
0
0
40
40
40
40
0
{ "bin_edges": [ 40, 40 ], "hist": [ 113619 ] }
true
yushengsu/stack-v2-python-with-content-chunk1-modified
default
train
113,619
snapshot_id
0
0
40
40
40
40
0
{ "bin_edges": [ 40, 40 ], "hist": [ 113619 ] }
true
usamaamjad23/Tested-1k-Python-Alpaca
default
train
1,000
instruction
0
0
37
5,175
617.389
342.5
654.08189
{ "bin_edges": [ 37, 551, 1065, 1579, 2093, 2607, 3121, 3635, 4149, 4663, 5175 ], "hist": [ 597, 184, 142, 55, 9, 4, 3, 2, 1, 3 ] }
false
usamaamjad23/Tested-1k-Python-Alpaca
default
train
1,000
output
0
0
140
5,495
1,248.373
1,108.5
690.32818
{ "bin_edges": [ 140, 676, 1212, 1748, 2284, 2820, 3356, 3892, 4428, 4964, 5495 ], "hist": [ 156, 417, 274, 86, 34, 18, 5, 3, 3, 4 ] }
false
Moodyspider266/Curated_service_projects
default
train
104
assistant
0
0
96
3,716
985.90385
855
929.953
{ "bin_edges": [ 96, 459, 822, 1185, 1548, 1911, 2274, 2637, 3000, 3363, 3716 ], "hist": [ 44, 8, 8, 16, 8, 10, 3, 4, 2, 1 ] }
false
Moodyspider266/Curated_service_projects
default
train
104
human
0
0
16
250
80.29808
60.5
50.38066
{ "bin_edges": [ 16, 40, 64, 88, 112, 136, 160, 184, 208, 232, 250 ], "hist": [ 9, 45, 21, 9, 6, 3, 3, 5, 1, 2 ] }
false
Moodyspider266/Curated_service_projects
default
train
104
text
0
0
823
4,485
1,760.20192
1,629
964.43524
{ "bin_edges": [ 823, 1190, 1557, 1924, 2291, 2658, 3025, 3392, 3759, 4126, 4485 ], "hist": [ 43, 6, 9, 17, 9, 8, 4, 4, 3, 1 ] }
false
jpwahle/WizardLMTeam-WizardLM_evol_instruct_70k_train_gpt-4o-mini_discussion_sample_chosen_rejected
default
train
47
chosen
0
0
795
4,564
2,303.08511
2,398
869.48485
{ "bin_edges": [ 795, 1172, 1549, 1926, 2303, 2680, 3057, 3434, 3811, 4188, 4564 ], "hist": [ 8, 3, 3, 8, 9, 7, 5, 2, 1, 1 ] }
false
jpwahle/WizardLMTeam-WizardLM_evol_instruct_70k_train_gpt-4o-mini_discussion_sample_chosen_rejected
default
train
47
prompt
0
0
64
2,119
506.21277
444
368.69228
{ "bin_edges": [ 64, 270, 476, 682, 888, 1094, 1300, 1506, 1712, 1918, 2119 ], "hist": [ 12, 15, 9, 6, 3, 0, 1, 0, 0, 1 ] }
false
jpwahle/WizardLMTeam-WizardLM_evol_instruct_70k_train_gpt-4o-mini_discussion_sample_chosen_rejected
default
train
47
rejected
0
0
802
4,263
2,213.53191
2,290
912.41754
{ "bin_edges": [ 802, 1149, 1496, 1843, 2190, 2537, 2884, 3231, 3578, 3925, 4263 ], "hist": [ 6, 10, 2, 2, 10, 7, 5, 1, 2, 2 ] }
false
de-Rodrigo/merit
en-digital-seq
validation
1,831
ground_truth
0
0
295
1,157
598.21518
503
270.10544
{ "bin_edges": [ 295, 382, 469, 556, 643, 730, 817, 904, 991, 1078, 1157 ], "hist": [ 445, 305, 552, 21, 0, 0, 25, 150, 209, 124 ] }
false
de-Rodrigo/merit
en-digital-seq
test
4,349
ground_truth
0
0
418
1,182
633.37848
533
226.38714
{ "bin_edges": [ 418, 495, 572, 649, 726, 803, 880, 957, 1034, 1111, 1182 ], "hist": [ 936, 2120, 395, 1, 0, 0, 0, 85, 736, 76 ] }
false
de-Rodrigo/merit
en-digital-seq
train
7,324
ground_truth
0
0
299
1,187
596.47037
501
267.39361
{ "bin_edges": [ 299, 388, 477, 566, 655, 744, 833, 922, 1011, 1100, 1187 ], "hist": [ 1981, 1104, 2212, 18, 0, 0, 218, 686, 869, 236 ] }
false
de-Rodrigo/merit
es-digital-seq
test
4,426
ground_truth
0
0
511
1,345
802.24514
699
221.52877
{ "bin_edges": [ 511, 595, 679, 763, 847, 931, 1015, 1099, 1183, 1267, 1345 ], "hist": [ 503, 1312, 831, 287, 567, 50, 5, 288, 521, 62 ] }
false