dataset
stringlengths
4
115
config
stringlengths
1
121
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
1
22.7k
min
int64
0
1.81M
max
int64
0
981M
mean
float64
0
42.2M
median
float64
0
24M
std
float64
0
84.2M
histogram
dict
partial
bool
2 classes
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=0.95, 'top_k'=10000, 'top_p'=0.2}
train
3,270
prediction
5
499
39.18563
6
60.99415
{ "bin_edges": [ 5, 55, 105, 155, 205, 255, 305, 355, 405, 455, 499 ], "hist": [ 2204, 476, 542, 9, 4, 4, 5, 5, 13, 8 ] }
false
wbxlala/har_str_100
default
train
5,881
sample
2,588
4,066
3,365.04132
3,286
426.11789
{ "bin_edges": [ 2588, 2736, 2884, 3032, 3180, 3328, 3476, 3624, 3772, 3920, 4066 ], "hist": [ 143, 781, 995, 694, 407, 167, 168, 856, 1318, 352 ] }
false
wbxlala/har_str_100
default
test
1,471
sample
3,224
5,200
4,297.16723
4,230
604.11869
{ "bin_edges": [ 3224, 3422, 3620, 3818, 4016, 4214, 4412, 4610, 4808, 5006, 5200 ], "hist": [ 52, 200, 231, 149, 93, 64, 19, 135, 374, 154 ] }
false
wiserifle/data-grabber-1k
default
train
1,000
text
246
894
556.624
554
224.40291
{ "bin_edges": [ 246, 311, 376, 441, 506, 571, 636, 701, 766, 831, 894 ], "hist": [ 316, 20, 0, 2, 237, 84, 0, 4, 187, 150 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=100, 'top_p'=0.1}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=100, 'top_p'=0.1}
train
3,270
prediction
5
499
39.18471
6
61.00867
{ "bin_edges": [ 5, 55, 105, 155, 205, 255, 305, 355, 405, 455, 499 ], "hist": [ 2204, 476, 542, 9, 4, 4, 5, 5, 13, 8 ] }
false
PL-MTEB/polemo2_out
default
train
5,783
text
1
5,391
780.56839
698
433.13954
{ "bin_edges": [ 1, 541, 1081, 1621, 2161, 2701, 3241, 3781, 4321, 4861, 5391 ], "hist": [ 1782, 3018, 691, 193, 81, 12, 2, 2, 1, 1 ] }
false
PL-MTEB/polemo2_out
default
validation
494
text
20
2,035
589.26518
565.5
317.58414
{ "bin_edges": [ 20, 222, 424, 626, 828, 1030, 1232, 1434, 1636, 1838, 2035 ], "hist": [ 60, 96, 134, 110, 53, 21, 12, 5, 2, 1 ] }
false
PL-MTEB/polemo2_out
default
test
494
text
8
1,831
587.04251
570
300.595
{ "bin_edges": [ 8, 191, 374, 557, 740, 923, 1106, 1289, 1472, 1655, 1831 ], "hist": [ 49, 73, 110, 139, 61, 35, 13, 11, 2, 1 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=100, 'top_p'=0.05}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=100, 'top_p'=0.05}
train
3,270
prediction
5
499
39.18471
6
61.00867
{ "bin_edges": [ 5, 55, 105, 155, 205, 255, 305, 355, 405, 455, 499 ], "hist": [ 2204, 476, 542, 9, 4, 4, 5, 5, 13, 8 ] }
false
minnesotanlp/LLM-Artifacts
freeform_deepfake_human
train
93,318
metaphor
572
52,106
19,807.99537
14,939.5
15,258.16375
{ "bin_edges": [ 572, 5726, 10880, 16034, 21188, 26342, 31496, 36650, 41804, 46958, 52106 ], "hist": [ 15973, 16877, 16927, 11368, 7585, 5107, 3585, 2667, 1962, 11267 ] }
false
minnesotanlp/LLM-Artifacts
freeform_deepfake_human
train
93,318
text
21
59,509
1,111.84973
706
1,419.88169
{ "bin_edges": [ 21, 5970, 11919, 17868, 23817, 29766, 35715, 41664, 47613, 53562, 59509 ], "hist": [ 91973, 1218, 83, 24, 9, 3, 4, 3, 0, 1 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=100, 'top_p'=0.2}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=100, 'top_p'=0.2}
train
3,270
prediction
5
499
39.18135
6
60.99621
{ "bin_edges": [ 5, 55, 105, 155, 205, 255, 305, 355, 405, 455, 499 ], "hist": [ 2204, 476, 542, 9, 4, 4, 5, 5, 13, 8 ] }
false
kyleeasterly/purple-aerospace-mix-v2-300-8
default
train
2,700
text
314
13,893
4,288.47667
3,911
2,023.17564
{ "bin_edges": [ 314, 1672, 3030, 4388, 5746, 7104, 8462, 9820, 11178, 12536, 13893 ], "hist": [ 103, 668, 879, 536, 276, 108, 48, 69, 10, 3 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=1000, 'top_p'=0.1}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=1000, 'top_p'=0.1}
train
3,270
prediction
5
499
39.18471
6
61.00867
{ "bin_edges": [ 5, 55, 105, 155, 205, 255, 305, 355, 405, 455, 499 ], "hist": [ 2204, 476, 542, 9, 4, 4, 5, 5, 13, 8 ] }
false
niting3c/Malicious_packets
default
train
1,256,906
text
37
3,541
1,713.14733
1,742
332.162
{ "bin_edges": [ 37, 388, 739, 1090, 1441, 1792, 2143, 2494, 2845, 3196, 3541 ], "hist": [ 10762, 6145, 6956, 33968, 860124, 309148, 3359, 2874, 2898, 20672 ] }
false
nus-yam/ex-repair
selfapr-corpus-defects4j-Chart-in
train
120,157
ground_truth
19
133
63.09499
65
15.88855
{ "bin_edges": [ 19, 31, 43, 55, 67, 79, 91, 103, 115, 127, 133 ], "hist": [ 2245, 14132, 18505, 28351, 31963, 24793, 116, 44, 0, 8 ] }
true
nus-yam/ex-repair
selfapr-corpus-defects4j-Chart-in
train
120,157
path
41
96
60.08304
58
10.4199
{ "bin_edges": [ 41, 47, 53, 59, 65, 71, 77, 83, 89, 95, 96 ], "hist": [ 2998, 31880, 28896, 18451, 15893, 15001, 3365, 2181, 0, 1492 ] }
true
nus-yam/ex-repair
selfapr-corpus-defects4j-Chart-in
train
120,157
source
2,047
196,813
43,367.11287
26,504
42,562.94986
{ "bin_edges": [ 2047, 21524, 41001, 60478, 79955, 99432, 118909, 138386, 157863, 177340, 196813 ], "hist": [ 49612, 26506, 18647, 10205, 3502, 3401, 955, 2908, 0, 4421 ] }
true
nus-yam/ex-repair
selfapr-corpus-defects4j-Chart-in
train
120,157
source_diag
155
2,942
1,008.04751
877
588.64169
{ "bin_edges": [ 155, 434, 713, 992, 1271, 1550, 1829, 2108, 2387, 2666, 2942 ], "hist": [ 15822, 32488, 20243, 21868, 8952, 6404, 4355, 6047, 3657, 321 ] }
true
vvud/eb-data
default
train
13
instruction
14
44
25.76923
25
8.47697
{ "bin_edges": [ 14, 18, 22, 26, 30, 34, 38, 42, 44 ], "hist": [ 2, 2, 3, 2, 2, 1, 0, 1 ] }
false
vvud/eb-data
default
train
13
text
61
289
165.76923
77
108.78814
{ "bin_edges": [ 61, 84, 107, 130, 153, 176, 199, 222, 245, 268, 289 ], "hist": [ 7, 0, 0, 0, 0, 0, 0, 0, 0, 6 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=1000, 'top_p'=0.05}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=1000, 'top_p'=0.05}
train
3,270
prediction
5
499
39.18471
6
61.00867
{ "bin_edges": [ 5, 55, 105, 155, 205, 255, 305, 355, 405, 455, 499 ], "hist": [ 2204, 476, 542, 9, 4, 4, 5, 5, 13, 8 ] }
false
HydraLM/goat_standardized
default
train
5,238,900
message
3
503
53.93334
28
73.34676
{ "bin_edges": [ 3, 54, 105, 156, 207, 258, 309, 360, 411, 462, 503 ], "hist": [ 4038526, 587114, 149639, 128946, 124380, 80161, 77347, 20112, 29604, 3071 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=10000, 'top_p'=0.05}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=10000, 'top_p'=0.05}
train
3,270
prediction
5
499
39.18471
6
61.00867
{ "bin_edges": [ 5, 55, 105, 155, 205, 255, 305, 355, 405, 455, 499 ], "hist": [ 2204, 476, 542, 9, 4, 4, 5, 5, 13, 8 ] }
false
sam-mosaic/dolly_hhrlhf_yashgoenka-gorilla-16k
default
train
60,310
prompt
142
24,057
311.58388
194
558.0579
{ "bin_edges": [ 142, 2534, 4926, 7318, 9710, 12102, 14494, 16886, 19278, 21670, 24057 ], "hist": [ 59721, 420, 119, 29, 5, 8, 3, 1, 3, 1 ] }
false
sam-mosaic/dolly_hhrlhf_yashgoenka-gorilla-16k
default
train
60,310
response
0
26,097
423.56395
325
418.69481
{ "bin_edges": [ 0, 2610, 5220, 7830, 10440, 13050, 15660, 18270, 20880, 23490, 26097 ], "hist": [ 60125, 156, 19, 8, 1, 0, 0, 0, 0, 1 ] }
false
sam-mosaic/dolly_hhrlhf_yashgoenka-gorilla-16k
default
test
15,129
prompt
140
5,951
258.04455
243
172.67941
{ "bin_edges": [ 140, 722, 1304, 1886, 2468, 3050, 3632, 4214, 4796, 5378, 5951 ], "hist": [ 14990, 79, 29, 13, 5, 4, 2, 2, 2, 3 ] }
false
sam-mosaic/dolly_hhrlhf_yashgoenka-gorilla-16k
default
test
15,129
response
0
3,485
791.48298
780
447.19793
{ "bin_edges": [ 0, 349, 698, 1047, 1396, 1745, 2094, 2443, 2792, 3141, 3485 ], "hist": [ 2701, 3708, 4531, 2738, 1140, 245, 42, 17, 5, 2 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=1000, 'top_p'=0.2}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=1000, 'top_p'=0.2}
train
3,270
prediction
5
499
39.18471
6
61.00867
{ "bin_edges": [ 5, 55, 105, 155, 205, 255, 305, 355, 405, 455, 499 ], "hist": [ 2204, 476, 542, 9, 4, 4, 5, 5, 13, 8 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=10000, 'top_p'=0.1}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=10000, 'top_p'=0.1}
train
3,270
prediction
5
499
39.18471
6
61.00867
{ "bin_edges": [ 5, 55, 105, 155, 205, 255, 305, 355, 405, 455, 499 ], "hist": [ 2204, 476, 542, 9, 4, 4, 5, 5, 13, 8 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=10000, 'top_p'=0.2}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.0, 'top_k'=10000, 'top_p'=0.2}
train
3,270
prediction
5
499
39.1841
6
60.99492
{ "bin_edges": [ 5, 55, 105, 155, 205, 255, 305, 355, 405, 455, 499 ], "hist": [ 2204, 476, 542, 9, 4, 4, 5, 5, 13, 8 ] }
false
Fin-Eval/Fin-Eval
default
test
287
答案
1
463
50.51163
10
90.31935
{ "bin_edges": [ 1, 48, 95, 142, 189, 236, 283, 330, 377, 424, 463 ], "hist": [ 120, 24, 11, 6, 3, 0, 3, 0, 2, 3 ] }
false
Fin-Eval/Fin-Eval
default
test
287
问题
3
1,322
126.00697
46
220.45295
{ "bin_edges": [ 3, 135, 267, 399, 531, 663, 795, 927, 1059, 1191, 1322 ], "hist": [ 209, 49, 11, 1, 3, 0, 5, 7, 1, 1 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=100, 'top_p'=0.5}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=100, 'top_p'=0.5}
train
3,270
prediction
5
519
39.13242
6
60.99337
{ "bin_edges": [ 5, 57, 109, 161, 213, 265, 317, 369, 421, 473, 519 ], "hist": [ 2636, 47, 536, 10, 8, 6, 4, 7, 10, 6 ] }
false
atmallen/quirky_capitals_alice
default
train
512
statement
54
108
71.16016
70
7.16248
{ "bin_edges": [ 54, 60, 66, 72, 78, 84, 90, 96, 102, 108, 108 ], "hist": [ 7, 111, 166, 144, 55, 17, 10, 1, 0, 1 ] }
false
pykeio/vtuber-chats-reprocessed
default
train
2,235,416
message
0
460
14.18688
11
12.71161
{ "bin_edges": [ 0, 47, 94, 141, 188, 235, 282, 329, 376, 423, 460 ], "hist": [ 2175863, 53161, 4811, 1050, 421, 92, 9, 6, 1, 2 ] }
false
Falah/desert_fashion_SDXL_refiner_prompts
default
train
2,000,000
prompts
821
928
875.3353
875
14.90166
{ "bin_edges": [ 821, 832, 843, 854, 865, 876, 887, 898, 909, 920, 928 ], "hist": [ 1465, 25328, 123903, 326413, 523174, 531903, 329848, 118797, 18611, 558 ] }
false
niting3c/Malicious_packets
default
test
894,651
text
37
3,541
1,882.34861
1,746
813.30224
{ "bin_edges": [ 37, 388, 739, 1090, 1441, 1792, 2143, 2494, 2845, 3196, 3541 ], "hist": [ 29029, 34648, 35712, 47005, 400930, 164788, 21296, 17726, 17577, 125940 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=1000, 'top_p'=0.5}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=1000, 'top_p'=0.5}
train
3,270
prediction
5
519
38.76728
6
60.46844
{ "bin_edges": [ 5, 57, 109, 161, 213, 265, 317, 369, 421, 473, 519 ], "hist": [ 2646, 48, 525, 10, 9, 5, 5, 7, 11, 4 ] }
false
dhiruHF/small-occupation-classifier
default
train
300
text
84
196
123.48667
121
20.03082
{ "bin_edges": [ 84, 96, 108, 120, 132, 144, 156, 168, 180, 192, 196 ], "hist": [ 17, 54, 67, 62, 55, 21, 17, 5, 1, 1 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=1000, 'top_p'=1.0}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=1000, 'top_p'=1.0}
train
3,270
prediction
4
518
44.34465
6
74.6261
{ "bin_edges": [ 4, 56, 108, 160, 212, 264, 316, 368, 420, 472, 518 ], "hist": [ 2487, 191, 473, 15, 12, 23, 14, 30, 22, 3 ] }
false
Tamazight-NLP/AmaWar
examples
train
460
source_string
8
69
22.4
22
7.7057
{ "bin_edges": [ 8, 15, 22, 29, 36, 43, 50, 57, 64, 69 ], "hist": [ 75, 151, 143, 69, 17, 3, 1, 0, 1 ] }
false
Tamazight-NLP/AmaWar
examples
train
460
target_string
7
66
20.81304
20
7.49407
{ "bin_edges": [ 7, 13, 19, 25, 31, 37, 43, 49, 55, 61, 66 ], "hist": [ 58, 132, 154, 78, 22, 10, 3, 2, 0, 1 ] }
false
nlplabtdtu/people_qa
default
train
19,579
answer
20
14,777
875.56264
797
579.84933
{ "bin_edges": [ 20, 1496, 2972, 4448, 5924, 7400, 8876, 10352, 11828, 13304, 14777 ], "hist": [ 16395, 3171, 10, 2, 0, 0, 0, 0, 0, 1 ] }
false
nlplabtdtu/people_qa
default
train
19,579
question
11
193
47.98841
46
19.23707
{ "bin_edges": [ 11, 30, 49, 68, 87, 106, 125, 144, 163, 182, 193 ], "hist": [ 3409, 7332, 5872, 2281, 550, 108, 19, 6, 1, 1 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=10000, 'top_p'=1.0}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=10000, 'top_p'=1.0}
train
3,270
prediction
4
529
43.22049
6
72.20388
{ "bin_edges": [ 4, 57, 110, 163, 216, 269, 322, 375, 428, 481, 529 ], "hist": [ 2570, 153, 437, 10, 20, 13, 21, 28, 15, 3 ] }
false
FreedomIntelligence/sharegpt-spanish
default
train
5,704
id
1
5
4.82889
5
0.43376
{ "bin_edges": [ 1, 2, 3, 4, 5, 5 ], "hist": [ 6, 7, 75, 781, 4835 ] }
false
Tamazight-NLP/AmaWar
expressions
train
406
source_string
4
39
15.44089
15
4.36826
{ "bin_edges": [ 4, 8, 12, 16, 20, 24, 28, 32, 36, 39 ], "hist": [ 9, 59, 146, 142, 33, 10, 4, 2, 1 ] }
false
Tamazight-NLP/AmaWar
expressions
train
406
target_string
5
37
13.02463
12
4.3307
{ "bin_edges": [ 5, 9, 13, 17, 21, 25, 29, 33, 37, 37 ], "hist": [ 23, 188, 133, 31, 19, 9, 1, 1, 1 ] }
false
judy93536/benz-peri-52k-torchdata
default
train
26,079
text
51
35,475
2,306.20599
1,920
1,867.33364
{ "bin_edges": [ 51, 3594, 7137, 10680, 14223, 17766, 21309, 24852, 28395, 31938, 35475 ], "hist": [ 22527, 2851, 546, 95, 29, 17, 3, 7, 3, 1 ] }
false
mattjackson1989/github-issues
default
train
500
body
9
19,442
1,437.88147
633.5
2,329.23909
{ "bin_edges": [ 9, 1953, 3897, 5841, 7785, 9729, 11673, 13617, 15561, 17505, 19442 ], "hist": [ 380, 40, 21, 8, 8, 1, 3, 2, 0, 1 ] }
false
mattjackson1989/github-issues
default
train
500
comments_url
70
70
70
70
0
{ "bin_edges": [ 70, 70 ], "hist": [ 500 ] }
false
mattjackson1989/github-issues
default
train
500
events_url
68
68
68
68
0
{ "bin_edges": [ 68, 68 ], "hist": [ 500 ] }
false
mattjackson1989/github-issues
default
train
500
html_url
49
51
50.112
51
0.9947
{ "bin_edges": [ 49, 50, 51, 51 ], "hist": [ 222, 0, 278 ] }
false
mattjackson1989/github-issues
default
train
500
labels_url
75
75
75
75
0
{ "bin_edges": [ 75, 75 ], "hist": [ 500 ] }
false
mattjackson1989/github-issues
default
train
500
node_id
18
19
18.444
18
0.49735
{ "bin_edges": [ 18, 19, 19 ], "hist": [ 278, 222 ] }
false
mattjackson1989/github-issues
default
train
500
timeline_url
70
70
70
70
0
{ "bin_edges": [ 70, 70 ], "hist": [ 500 ] }
false
mattjackson1989/github-issues
default
train
500
title
9
150
49.468
46
25.96315
{ "bin_edges": [ 9, 24, 39, 54, 69, 84, 99, 114, 129, 144, 150 ], "hist": [ 63, 116, 145, 91, 41, 13, 12, 13, 2, 4 ] }
false
mattjackson1989/github-issues
default
train
500
url
61
61
61
61
0
{ "bin_edges": [ 61, 61 ], "hist": [ 500 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=10000, 'top_p'=0.5}
train
3,270
id
17
20
19.40061
19
0.61718
{ "bin_edges": [ 17, 18, 19, 20, 20 ], "hist": [ 21, 167, 1563, 1519 ] }
false
automated-research-group/llama2_7b_chat-boolq-results
{'do_sample'=True, 'beams'=5, 'temperature'=1.05, 'top_k'=10000, 'top_p'=0.5}
train
3,270
prediction
5
510
39.18838
6
61.47645
{ "bin_edges": [ 5, 56, 107, 158, 209, 260, 311, 362, 413, 464, 510 ], "hist": [ 2589, 88, 542, 11, 7, 1, 5, 9, 12, 6 ] }
false
Patt/copa_th
default
train
400
choice1
10
62
27.405
26
9.15891
{ "bin_edges": [ 10, 16, 22, 28, 34, 40, 46, 52, 58, 62 ], "hist": [ 32, 86, 103, 87, 50, 25, 11, 5, 1 ] }
false
Patt/copa_th
default
train
400
choice1_th
5
61
21.2975
20
9.12673
{ "bin_edges": [ 5, 11, 17, 23, 29, 35, 41, 47, 53, 59, 61 ], "hist": [ 34, 99, 114, 76, 43, 21, 7, 3, 2, 1 ] }
false
Patt/copa_th
default
train
400
choice2
11
64
27.5375
27
9.19674
{ "bin_edges": [ 11, 17, 23, 29, 35, 41, 47, 53, 59, 64 ], "hist": [ 44, 88, 97, 85, 56, 16, 9, 2, 3 ] }
false
Patt/copa_th
default
train
400
choice2_th
5
63
21.34
20
8.62876
{ "bin_edges": [ 5, 11, 17, 23, 29, 35, 41, 47, 53, 59, 63 ], "hist": [ 26, 104, 112, 80, 54, 11, 8, 4, 0, 1 ] }
false
Patt/copa_th
default
train
400
premise
14
67
34.63
34
10.23866
{ "bin_edges": [ 14, 20, 26, 32, 38, 44, 50, 56, 62, 67 ], "hist": [ 22, 57, 79, 91, 74, 45, 21, 9, 2 ] }
false
Patt/copa_th
default
train
400
premise_th
4
64
26.7775
25
10.33077
{ "bin_edges": [ 4, 11, 18, 25, 32, 39, 46, 53, 60, 64 ], "hist": [ 15, 62, 115, 76, 71, 47, 7, 6, 1 ] }
false
Tamazight-NLP/AmaWar
poems
train
532
source_string
5
69
31.39286
32
7.48871
{ "bin_edges": [ 5, 12, 19, 26, 33, 40, 47, 54, 61, 68, 69 ], "hist": [ 4, 39, 33, 234, 177, 34, 5, 3, 2, 1 ] }
false
Tamazight-NLP/AmaWar
poems
train
532
target_string
6
69
28.34586
28
8.80703
{ "bin_edges": [ 6, 13, 20, 27, 34, 41, 48, 55, 62, 69, 69 ], "hist": [ 34, 36, 151, 170, 104, 24, 6, 6, 0, 1 ] }
false
judy93536/benz-peri-52k-torchdata
default
test
26,079
text
44
138,037
2,321.22673
1,918
2,111.67188
{ "bin_edges": [ 44, 13844, 27644, 41444, 55244, 69044, 82844, 96644, 110444, 124244, 138037 ], "hist": [ 26008, 59, 11, 0, 0, 0, 0, 0, 0, 1 ] }
false
Patt/copa_th
default
validation
100
choice1
12
63
28.84
28.5
9.58821
{ "bin_edges": [ 12, 18, 24, 30, 36, 42, 48, 54, 60, 63 ], "hist": [ 12, 17, 29, 19, 13, 7, 1, 1, 1 ] }
false
Patt/copa_th
default
validation
100
choice1_th
6
48
22.95
22
9.18208
{ "bin_edges": [ 6, 11, 16, 21, 26, 31, 36, 41, 46, 48 ], "hist": [ 8, 16, 18, 19, 21, 11, 3, 2, 2 ] }
false
Patt/copa_th
default
validation
100
choice2
12
56
28.46
27.5
9.51704
{ "bin_edges": [ 12, 17, 22, 27, 32, 37, 42, 47, 52, 56 ], "hist": [ 9, 18, 18, 23, 12, 11, 5, 3, 1 ] }
false
Patt/copa_th
default
validation
100
choice2_th
6
48
22.58
21
8.4509
{ "bin_edges": [ 6, 11, 16, 21, 26, 31, 36, 41, 46, 48 ], "hist": [ 4, 19, 19, 24, 15, 10, 6, 1, 2 ] }
false
Patt/copa_th
default
validation
100
premise
10
66
34.01
34
10.90315
{ "bin_edges": [ 10, 16, 22, 28, 34, 40, 46, 52, 58, 64, 66 ], "hist": [ 2, 13, 11, 23, 19, 21, 6, 2, 2, 1 ] }
false
Patt/copa_th
default
validation
100
premise_th
5
80
27.34
26.5
11.72171
{ "bin_edges": [ 5, 13, 21, 29, 37, 45, 53, 61, 69, 77, 80 ], "hist": [ 8, 23, 22, 28, 14, 3, 0, 1, 0, 1 ] }
false
Svetlana0303/sveta_dataset
default
train
340
Текст
3
571
94.09706
72.5
76.94402
{ "bin_edges": [ 3, 60, 117, 174, 231, 288, 345, 402, 459, 516, 571 ], "hist": [ 132, 106, 63, 23, 6, 7, 0, 0, 1, 2 ] }
false
Patt/copa_th
default
test
500
choice1
8
62
26.62
26
8.6439
{ "bin_edges": [ 8, 14, 20, 26, 32, 38, 44, 50, 56, 62, 62 ], "hist": [ 17, 87, 144, 111, 89, 29, 20, 2, 0, 1 ] }
false
Patt/copa_th
default
test
500
choice1_th
6
53
20.344
19
8.52073
{ "bin_edges": [ 6, 11, 16, 21, 26, 31, 36, 41, 46, 51, 53 ], "hist": [ 46, 123, 106, 94, 70, 36, 11, 8, 5, 1 ] }
false
Patt/copa_th
default
test
500
choice2
8
61
26.558
26
8.48355
{ "bin_edges": [ 8, 14, 20, 26, 32, 38, 44, 50, 56, 61 ], "hist": [ 15, 92, 140, 123, 79, 32, 11, 7, 1 ] }
false
Patt/copa_th
default
test
500
choice2_th
5
54
20.528
19
8.2012
{ "bin_edges": [ 5, 10, 15, 20, 25, 30, 35, 40, 45, 50, 54 ], "hist": [ 30, 94, 131, 99, 81, 38, 10, 11, 5, 1 ] }
false
Patt/copa_th
default
test
500
premise
9
67
33.01
32
9.93294
{ "bin_edges": [ 9, 15, 21, 27, 33, 39, 45, 51, 57, 63, 67 ], "hist": [ 5, 36, 86, 140, 111, 64, 25, 21, 9, 3 ] }
false
Patt/copa_th
default
test
500
premise_th
5
57
25.494
24
9.51478
{ "bin_edges": [ 5, 11, 17, 23, 29, 35, 41, 47, 53, 57 ], "hist": [ 17, 70, 129, 113, 74, 64, 20, 7, 6 ] }
false
michaelmallari/mlb-statcast-pitchers
default
train
878
last_name, first_name
9
20
13.54214
13
2.0903
{ "bin_edges": [ 9, 11, 13, 15, 17, 19, 20 ], "hist": [ 46, 246, 307, 205, 66, 8 ] }
false
Tamazight-NLP/AmaWar
proverbs
train
293
source_string
13
72
30.01706
29
9.92073
{ "bin_edges": [ 13, 19, 25, 31, 37, 43, 49, 55, 61, 67, 72 ], "hist": [ 29, 63, 77, 54, 40, 17, 8, 3, 1, 1 ] }
false
Tamazight-NLP/AmaWar
proverbs
train
293
target_string
12
103
28.73379
28
9.89313
{ "bin_edges": [ 12, 22, 32, 42, 52, 62, 72, 82, 92, 102, 103 ], "hist": [ 67, 125, 77, 20, 2, 0, 1, 0, 0, 1 ] }
false
Tamazight-NLP/AmaWar
riddles
train
255
source_string
21
142
48.58431
46
16.35747
{ "bin_edges": [ 21, 34, 47, 60, 73, 86, 99, 112, 125, 138, 142 ], "hist": [ 36, 97, 64, 37, 15, 3, 2, 0, 0, 1 ] }
false