dataset
stringlengths
5
115
config
stringlengths
1
162
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
0
77.9k
null_count
int64
0
62.9M
null_proportion
float64
0
1
min
int64
0
9.25M
max
int64
0
1.07B
mean
float64
0
90.4M
median
float64
0
80.1M
std
float64
0
130M
histogram
dict
partial
bool
2 classes
open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-7B-32k-details
cognitivecomputations__dolphin-2.9.3-mistral-7B-32k__leaderboard_musr_object_placements
latest
256
target_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 256 ] }
false
open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-7B-32k-details
cognitivecomputations__dolphin-2.9.3-mistral-7B-32k__leaderboard_musr_team_allocation
2024_07_22T09_19_12.455984
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-7B-32k-details
cognitivecomputations__dolphin-2.9.3-mistral-7B-32k__leaderboard_musr_team_allocation
2024_07_22T09_19_12.455984
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-7B-32k-details
cognitivecomputations__dolphin-2.9.3-mistral-7B-32k__leaderboard_musr_team_allocation
2024_07_22T09_19_12.455984
250
target
0
0
37
89
58.508
58
10.91377
{ "bin_edges": [ 37, 43, 49, 55, 61, 67, 73, 79, 85, 89 ], "hist": [ 17, 30, 43, 54, 52, 29, 12, 9, 4 ] }
false
open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-7B-32k-details
cognitivecomputations__dolphin-2.9.3-mistral-7B-32k__leaderboard_musr_team_allocation
2024_07_22T09_19_12.455984
250
target_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-7B-32k-details
cognitivecomputations__dolphin-2.9.3-mistral-7B-32k__leaderboard_musr_team_allocation
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-7B-32k-details
cognitivecomputations__dolphin-2.9.3-mistral-7B-32k__leaderboard_musr_team_allocation
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-7B-32k-details
cognitivecomputations__dolphin-2.9.3-mistral-7B-32k__leaderboard_musr_team_allocation
latest
250
target
0
0
37
89
58.508
58
10.91377
{ "bin_edges": [ 37, 43, 49, 55, 61, 67, 73, 79, 85, 89 ], "hist": [ 17, 30, 43, 54, 52, 29, 12, 9, 4 ] }
false
open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-7B-32k-details
cognitivecomputations__dolphin-2.9.3-mistral-7B-32k__leaderboard_musr_team_allocation
latest
250
target_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-7B-32k-details
cognitivecomputations__dolphin-2.9.3-mistral-7B-32k__leaderboard_bbh_navigate
2024_07_22T09_19_12.455984
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-7B-32k-details
cognitivecomputations__dolphin-2.9.3-mistral-7B-32k__leaderboard_bbh_navigate
2024_07_22T09_19_12.455984
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-7B-32k-details
cognitivecomputations__dolphin-2.9.3-mistral-7B-32k__leaderboard_bbh_navigate
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/cognitivecomputations__dolphin-2.9.3-mistral-7B-32k-details
cognitivecomputations__dolphin-2.9.3-mistral-7B-32k__leaderboard_bbh_navigate
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
cicixu/add_s_all_robust
default
test
636
text
0
0
176
2,017
1,026.3239
811.5
539.40435
{ "bin_edges": [ 176, 361, 546, 731, 916, 1101, 1286, 1471, 1656, 1841, 2017 ], "hist": [ 22, 107, 146, 80, 43, 35, 28, 8, 139, 28 ] }
false
cicixu/add_s_all_robust
default
train
5,724
text
0
0
169
2,331
1,044.02568
830
546.04755
{ "bin_edges": [ 169, 386, 603, 820, 1037, 1254, 1471, 1688, 1905, 2122, 2331 ], "hist": [ 311, 1226, 1287, 511, 426, 284, 111, 1540, 22, 6 ] }
false
hysts-bot-data/daily-papers-stats
default
train
3,201
arxiv_id
0
0
10
10
10
10
0
{ "bin_edges": [ 10, 10 ], "hist": [ 3201 ] }
false
kshitizgajurel/Multilingual-Nepali-Customer-Care-Services-Dataset
Devanagari
train
17,760
query
0
0
10
1,070
42.67111
40
44.03
{ "bin_edges": [ 10, 117, 224, 331, 438, 545, 652, 759, 866, 973, 1070 ], "hist": [ 17725, 0, 0, 0, 0, 0, 0, 0, 10, 25 ] }
false
kshitizgajurel/Multilingual-Nepali-Customer-Care-Services-Dataset
Devanagari
train
17,760
response
0
0
11
1,648
56.63435
53
59.09411
{ "bin_edges": [ 11, 175, 339, 503, 667, 831, 995, 1159, 1323, 1487, 1648 ], "hist": [ 17725, 0, 0, 0, 0, 0, 10, 10, 10, 5 ] }
false
kshitizgajurel/Multilingual-Nepali-Customer-Care-Services-Dataset
Romanized
train
17,760
query
0
0
12
1,232
46.72466
44
48.93027
{ "bin_edges": [ 12, 135, 258, 381, 504, 627, 750, 873, 996, 1119, 1232 ], "hist": [ 17725, 0, 0, 0, 0, 0, 0, 0, 25, 10 ] }
false
kshitizgajurel/Multilingual-Nepali-Customer-Care-Services-Dataset
Romanized
train
17,760
response
0
0
14
1,822
61.78328
57
65.29943
{ "bin_edges": [ 14, 195, 376, 557, 738, 919, 1100, 1281, 1462, 1643, 1822 ], "hist": [ 17725, 0, 0, 0, 0, 0, 5, 15, 10, 5 ] }
false
AyseEe501/llama3_deneme
default
train
15,001
Answer
0
0
478
540
488.31978
489
3.3174
{ "bin_edges": [ 478, 485, 492, 499, 506, 513, 520, 527, 534, 540 ], "hist": [ 693, 13218, 1031, 40, 0, 0, 0, 18, 1 ] }
false
AyseEe501/llama3_deneme
default
train
15,001
Question
0
0
480
535
490.65829
492
2.96669
{ "bin_edges": [ 480, 486, 492, 498, 504, 510, 516, 522, 528, 534, 535 ], "hist": [ 822, 6054, 8106, 0, 0, 0, 0, 0, 7, 12 ] }
false
danaaubakirova/docmatix-subset
default
test
384
answers
0
0
42
1,789
190.11198
160.5
143.13322
{ "bin_edges": [ 42, 217, 392, 567, 742, 917, 1092, 1267, 1442, 1617, 1789 ], "hist": [ 262, 105, 13, 2, 0, 0, 0, 1, 0, 1 ] }
false
danaaubakirova/docmatix-subset
default
test
384
image
0
0
5,864
1,919,716
236,361.90625
200,794
198,775.9579
{ "bin_edges": [ 5864, 197250, 388636, 580022, 771408, 962794, 1154180, 1345566, 1536952, 1728338, 1919716 ], "hist": [ 185, 149, 34, 10, 2, 0, 1, 2, 0, 1 ] }
false
danaaubakirova/docmatix-subset
default
test
384
query
0
0
3
167
71.63802
70
26.29193
{ "bin_edges": [ 3, 20, 37, 54, 71, 88, 105, 122, 139, 156, 167 ], "hist": [ 6, 18, 72, 97, 105, 47, 21, 11, 5, 2 ] }
false
danaaubakirova/docmatix-subset
default
train
1,750
answers
0
0
1
992
184.57543
157
114.00527
{ "bin_edges": [ 1, 101, 201, 301, 401, 501, 601, 701, 801, 901, 992 ], "hist": [ 429, 702, 386, 143, 56, 20, 9, 2, 1, 2 ] }
false
danaaubakirova/docmatix-subset
default
train
1,750
image
0
0
7,264
4,761,940
245,343.952
188,946
299,335.17599
{ "bin_edges": [ 7264, 482732, 958200, 1433668, 1909136, 2384604, 2860072, 3335540, 3811008, 4286476, 4761940 ], "hist": [ 1604, 124, 11, 3, 2, 1, 1, 0, 0, 4 ] }
false
danaaubakirova/docmatix-subset
default
train
1,750
query
0
0
3
199
70.736
68
27.39108
{ "bin_edges": [ 3, 23, 43, 63, 83, 103, 123, 143, 163, 183, 199 ], "hist": [ 26, 238, 472, 482, 328, 124, 56, 17, 5, 2 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_boolean_expressions
2024_07_22T09_32_37.006648
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_boolean_expressions
2024_07_22T09_32_37.006648
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_boolean_expressions
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_boolean_expressions
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_disambiguation_qa
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_disambiguation_qa
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_disambiguation_qa
2024_07_22T09_32_37.006648
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_disambiguation_qa
2024_07_22T09_32_37.006648
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_hyperbaton
2024_07_22T09_32_37.006648
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_hyperbaton
2024_07_22T09_32_37.006648
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_hyperbaton
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_hyperbaton
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_object_counting
2024_07_22T09_32_37.006648
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_object_counting
2024_07_22T09_32_37.006648
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_object_counting
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_object_counting
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_formal_fallacies
2024_07_22T09_32_37.006648
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_formal_fallacies
2024_07_22T09_32_37.006648
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_formal_fallacies
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_formal_fallacies
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_navigate
2024_07_22T09_32_37.006648
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_navigate
2024_07_22T09_32_37.006648
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_navigate
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_navigate
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_penguins_in_a_table
2024_07_22T09_32_37.006648
146
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 146 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_penguins_in_a_table
2024_07_22T09_32_37.006648
146
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 146 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_penguins_in_a_table
latest
146
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 146 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_penguins_in_a_table
latest
146
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 146 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_ruin_names
2024_07_22T09_32_37.006648
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_ruin_names
2024_07_22T09_32_37.006648
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_ruin_names
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_ruin_names
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_salient_translation_error_detection
2024_07_22T09_32_37.006648
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_salient_translation_error_detection
2024_07_22T09_32_37.006648
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_salient_translation_error_detection
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_salient_translation_error_detection
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_sports_understanding
2024_07_22T09_32_37.006648
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_sports_understanding
2024_07_22T09_32_37.006648
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_sports_understanding
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_sports_understanding
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_temporal_sequences
2024_07_22T09_32_37.006648
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_temporal_sequences
2024_07_22T09_32_37.006648
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_temporal_sequences
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_temporal_sequences
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_geometric_shapes
2024_07_22T09_32_37.006648
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_geometric_shapes
2024_07_22T09_32_37.006648
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_geometric_shapes
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_geometric_shapes
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
piyushsalunke/tuner2
default
train
214
requirement
0
0
28
544
128.69626
104
79.23156
{ "bin_edges": [ 28, 80, 132, 184, 236, 288, 340, 392, 444, 496, 544 ], "hist": [ 61, 78, 37, 14, 17, 3, 1, 1, 0, 2 ] }
false
piyushsalunke/tuner2
default
train
214
testcase
0
0
341
22,666
2,724.8972
1,219.5
3,799.36229
{ "bin_edges": [ 341, 2574, 4807, 7040, 9273, 11506, 13739, 15972, 18205, 20438, 22666 ], "hist": [ 161, 22, 8, 5, 7, 7, 0, 1, 2, 1 ] }
false
piyushsalunke/tuner2
default
test
20
requirement
0
0
64
341
143.05
128.5
77.55302
{ "bin_edges": [ 64, 92, 120, 148, 176, 204, 232, 260, 288, 316, 341 ], "hist": [ 6, 3, 4, 1, 2, 1, 1, 1, 0, 1 ] }
false
piyushsalunke/tuner2
default
test
20
testcase
0
0
527
17,857
4,923.2
3,396
4,792.40127
{ "bin_edges": [ 527, 2261, 3995, 5729, 7463, 9197, 10931, 12665, 14399, 16133, 17857 ], "hist": [ 7, 4, 3, 1, 1, 2, 0, 1, 0, 1 ] }
false
piyushsalunke/tuner2
default
validation
12
requirement
0
0
29
261
150.16667
125.5
66.57441
{ "bin_edges": [ 29, 53, 77, 101, 125, 149, 173, 197, 221, 245, 261 ], "hist": [ 1, 0, 0, 5, 1, 0, 2, 1, 1, 1 ] }
false
piyushsalunke/tuner2
default
validation
12
testcase
0
0
519
2,032
1,076.58333
1,004.5
409.44651
{ "bin_edges": [ 519, 671, 823, 975, 1127, 1279, 1431, 1583, 1735, 1887, 2032 ], "hist": [ 2, 1, 3, 2, 2, 0, 0, 1, 0, 1 ] }
false
NeuroSpaceX/spam_dataset_v4Plus
default
train
661,296
message
0
0
20
4,096
261.6547
137
384.51337
{ "bin_edges": [ 20, 428, 836, 1244, 1652, 2060, 2468, 2876, 3284, 3692, 4096 ], "hist": [ 555235, 76435, 14841, 4615, 3182, 2115, 1555, 1181, 898, 1239 ] }
false
howkewlisthat/axsy_t5_function_dataset
default
test
37
expected_function_call
0
0
14
63
35.27027
33
11.85947
{ "bin_edges": [ 14, 19, 24, 29, 34, 39, 44, 49, 54, 59, 63 ], "hist": [ 2, 1, 10, 6, 7, 2, 2, 2, 4, 1 ] }
false
howkewlisthat/axsy_t5_function_dataset
default
test
37
user_input
0
0
16
65
36.02703
35
12.16664
{ "bin_edges": [ 16, 21, 26, 31, 36, 41, 46, 51, 56, 61, 65 ], "hist": [ 1, 8, 4, 6, 7, 4, 2, 1, 3, 1 ] }
false
howkewlisthat/axsy_t5_function_dataset
default
train
146
expected_function_call
0
0
14
63
34.30137
32
11.70403
{ "bin_edges": [ 14, 19, 24, 29, 34, 39, 44, 49, 54, 59, 63 ], "hist": [ 4, 20, 30, 30, 23, 8, 10, 4, 7, 10 ] }
false
howkewlisthat/axsy_t5_function_dataset
default
train
146
user_input
0
0
10
65
35.54795
33.5
12.63333
{ "bin_edges": [ 10, 16, 22, 28, 34, 40, 46, 52, 58, 64, 65 ], "hist": [ 3, 19, 12, 39, 26, 15, 12, 10, 7, 3 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_causal_judgement
2024_07_22T09_32_37.006648
187
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 187 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_causal_judgement
2024_07_22T09_32_37.006648
187
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 187 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_causal_judgement
latest
187
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 187 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_causal_judgement
latest
187
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 187 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_reasoning_about_colored_objects
2024_07_22T09_32_37.006648
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_reasoning_about_colored_objects
2024_07_22T09_32_37.006648
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_reasoning_about_colored_objects
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_reasoning_about_colored_objects
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_movie_recommendation
2024_07_22T09_32_37.006648
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_movie_recommendation
2024_07_22T09_32_37.006648
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_movie_recommendation
latest
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_movie_recommendation
latest
250
prompt_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false
open-llm-leaderboard/maldv__badger-mu-llama-3-8b-details
maldv__badger-mu-llama-3-8b__leaderboard_bbh_date_understanding
2024_07_22T09_32_37.006648
250
doc_hash
0
0
64
64
64
64
0
{ "bin_edges": [ 64, 64 ], "hist": [ 250 ] }
false