| dataset
				 stringlengths 4 115 | config
				 stringlengths 1 121 | split
				 stringlengths 1 228 | num_examples
				 int64 3 341M | column_name
				 stringlengths 1 22.7k | min
				 int64 0 1.81M | max
				 int64 0 981M | mean
				 float64 0 42.2M | median
				 float64 0 24M | std
				 float64 0 84.2M | histogram
				 dict | partial
				 bool 2
				classes | 
|---|---|---|---|---|---|---|---|---|---|---|---|
| 
	dair-ai/emotion | 
	split | 
	train | 16,000 | 
	text | 7 | 300 | 96.84581 | 86 | 55.90495 | 
	{
  "bin_edges": [
    7,
    37,
    67,
    97,
    127,
    157,
    187,
    217,
    247,
    277,
    300
  ],
  "hist": [
    1833,
    3789,
    3611,
    2623,
    1804,
    1048,
    635,
    353,
    213,
    91
  ]
} | false | 
| 
	dair-ai/emotion | 
	split | 
	validation | 2,000 | 
	text | 11 | 295 | 95.3475 | 85 | 54.82376 | 
	{
  "bin_edges": [
    11,
    40,
    69,
    98,
    127,
    156,
    185,
    214,
    243,
    272,
    295
  ],
  "hist": [
    293,
    473,
    415,
    309,
    228,
    140,
    66,
    35,
    29,
    12
  ]
} | false | 
| 
	dair-ai/emotion | 
	split | 
	test | 2,000 | 
	text | 14 | 296 | 96.5865 | 86 | 55.71599 | 
	{
  "bin_edges": [
    14,
    43,
    72,
    101,
    130,
    159,
    188,
    217,
    246,
    275,
    296
  ],
  "hist": [
    346,
    466,
    377,
    308,
    214,
    144,
    69,
    45,
    24,
    7
  ]
} | false | 
| 
	dair-ai/emotion | 
	unsplit | 
	train | 416,809 | 
	text | 2 | 830 | 97.0284 | 86 | 56.19823 | 
	{
  "bin_edges": [
    2,
    85,
    168,
    251,
    334,
    417,
    500,
    583,
    666,
    749,
    830
  ],
  "hist": [
    204631,
    162639,
    42259,
    7256,
    17,
    4,
    2,
    0,
    0,
    1
  ]
} | false | 
| 
	allenai/qasper | 
	qasper | 
	train | 888 | 
	abstract | 279 | 2,022 | 982.02703 | 959.5 | 294.62389 | 
	{
  "bin_edges": [
    279,
    454,
    629,
    804,
    979,
    1154,
    1329,
    1504,
    1679,
    1854,
    2022
  ],
  "hist": [
    16,
    79,
    151,
    219,
    193,
    127,
    57,
    25,
    17,
    4
  ]
} | false | 
| 
	allenai/qasper | 
	qasper | 
	train | 888 | 
	id | 10 | 10 | 10 | 10 | 0 | 
	{
  "bin_edges": [
    10,
    10
  ],
  "hist": [
    888
  ]
} | false | 
| 
	allenai/qasper | 
	qasper | 
	train | 888 | 
	title | 12 | 156 | 71.95495 | 70 | 20.95424 | 
	{
  "bin_edges": [
    12,
    27,
    42,
    57,
    72,
    87,
    102,
    117,
    132,
    147,
    156
  ],
  "hist": [
    6,
    43,
    166,
    247,
    217,
    136,
    47,
    19,
    6,
    1
  ]
} | false | 
| 
	allenai/qasper | 
	qasper | 
	validation | 281 | 
	abstract | 338 | 1,868 | 953.51957 | 909 | 299.11687 | 
	{
  "bin_edges": [
    338,
    492,
    646,
    800,
    954,
    1108,
    1262,
    1416,
    1570,
    1724,
    1868
  ],
  "hist": [
    9,
    31,
    47,
    76,
    40,
    30,
    28,
    9,
    6,
    5
  ]
} | false | 
| 
	allenai/qasper | 
	qasper | 
	validation | 281 | 
	id | 10 | 10 | 10 | 10 | 0 | 
	{
  "bin_edges": [
    10,
    10
  ],
  "hist": [
    281
  ]
} | false | 
| 
	allenai/qasper | 
	qasper | 
	validation | 281 | 
	title | 19 | 142 | 71.29181 | 68 | 21.77187 | 
	{
  "bin_edges": [
    19,
    32,
    45,
    58,
    71,
    84,
    97,
    110,
    123,
    136,
    142
  ],
  "hist": [
    7,
    18,
    47,
    78,
    63,
    31,
    16,
    16,
    4,
    1
  ]
} | false | 
| 
	allenai/qasper | 
	qasper | 
	test | 416 | 
	abstract | 252 | 1,909 | 930.93029 | 901.5 | 295.71791 | 
	{
  "bin_edges": [
    252,
    418,
    584,
    750,
    916,
    1082,
    1248,
    1414,
    1580,
    1746,
    1909
  ],
  "hist": [
    16,
    26,
    67,
    108,
    77,
    63,
    33,
    15,
    8,
    3
  ]
} | false | 
| 
	allenai/qasper | 
	qasper | 
	test | 416 | 
	id | 10 | 10 | 10 | 10 | 0 | 
	{
  "bin_edges": [
    10,
    10
  ],
  "hist": [
    416
  ]
} | false | 
| 
	allenai/qasper | 
	qasper | 
	test | 416 | 
	title | 12 | 153 | 71.10337 | 69 | 22.801 | 
	{
  "bin_edges": [
    12,
    27,
    42,
    57,
    72,
    87,
    102,
    117,
    132,
    147,
    153
  ],
  "hist": [
    8,
    28,
    76,
    111,
    88,
    65,
    28,
    9,
    2,
    1
  ]
} | false | 
| 
	llm-book/wrime-sentiment | 
	default | 
	test | 1,781 | 
	datetime | 13 | 16 | 14.76811 | 15 | 0.77721 | 
	{
  "bin_edges": [
    13,
    14,
    15,
    16,
    16
  ],
  "hist": [
    82,
    546,
    856,
    297
  ]
} | false | 
| 
	llm-book/wrime-sentiment | 
	default | 
	test | 1,781 | 
	sentence | 3 | 154 | 47.29927 | 39 | 33.27888 | 
	{
  "bin_edges": [
    3,
    19,
    35,
    51,
    67,
    83,
    99,
    115,
    131,
    147,
    154
  ],
  "hist": [
    354,
    435,
    341,
    215,
    160,
    104,
    73,
    45,
    52,
    2
  ]
} | false | 
| 
	llm-book/wrime-sentiment | 
	default | 
	train | 20,149 | 
	datetime | 13 | 16 | 14.68773 | 15 | 0.76921 | 
	{
  "bin_edges": [
    13,
    14,
    15,
    16,
    16
  ],
  "hist": [
    1201,
    6486,
    9866,
    2596
  ]
} | false | 
| 
	llm-book/wrime-sentiment | 
	default | 
	train | 20,149 | 
	sentence | 1 | 173 | 39.52747 | 30 | 30.78759 | 
	{
  "bin_edges": [
    1,
    19,
    37,
    55,
    73,
    91,
    109,
    127,
    145,
    163,
    173
  ],
  "hist": [
    5468,
    6483,
    3644,
    1882,
    1005,
    623,
    444,
    541,
    56,
    3
  ]
} | false | 
| 
	llm-book/wrime-sentiment | 
	default | 
	validation | 1,608 | 
	datetime | 13 | 16 | 14.60261 | 15 | 0.80411 | 
	{
  "bin_edges": [
    13,
    14,
    15,
    16,
    16
  ],
  "hist": [
    140,
    546,
    735,
    187
  ]
} | false | 
| 
	llm-book/wrime-sentiment | 
	default | 
	validation | 1,608 | 
	sentence | 3 | 163 | 53.86754 | 46 | 34.72008 | 
	{
  "bin_edges": [
    3,
    20,
    37,
    54,
    71,
    88,
    105,
    122,
    139,
    156,
    163
  ],
  "hist": [
    257,
    374,
    306,
    227,
    142,
    113,
    102,
    54,
    32,
    1
  ]
} | false | 
| 
	ttxy/resume_ner | 
	default | 
	test | 477 | 
	label | 9 | 1,035 | 160.87631 | 115 | 155.29666 | 
	{
  "bin_edges": [
    9,
    112,
    215,
    318,
    421,
    524,
    627,
    730,
    833,
    936,
    1035
  ],
  "hist": [
    222,
    156,
    44,
    23,
    13,
    6,
    8,
    1,
    2,
    2
  ]
} | false | 
| 
	ttxy/resume_ner | 
	default | 
	test | 477 | 
	text | 7 | 333 | 62.31237 | 49 | 52.13344 | 
	{
  "bin_edges": [
    7,
    40,
    73,
    106,
    139,
    172,
    205,
    238,
    271,
    304,
    333
  ],
  "hist": [
    179,
    174,
    61,
    22,
    19,
    5,
    7,
    6,
    1,
    3
  ]
} | false | 
| 
	ttxy/resume_ner | 
	default | 
	train | 3,821 | 
	label | 7 | 1,149 | 162.03612 | 121 | 149.21149 | 
	{
  "bin_edges": [
    7,
    122,
    237,
    352,
    467,
    582,
    697,
    812,
    927,
    1042,
    1149
  ],
  "hist": [
    1940,
    1173,
    329,
    175,
    89,
    55,
    35,
    17,
    7,
    1
  ]
} | false | 
| 
	ttxy/resume_ner | 
	default | 
	train | 3,821 | 
	text | 5 | 355 | 63.95629 | 53 | 50.26945 | 
	{
  "bin_edges": [
    5,
    41,
    77,
    113,
    149,
    185,
    221,
    257,
    293,
    329,
    355
  ],
  "hist": [
    1312,
    1526,
    480,
    244,
    109,
    71,
    37,
    23,
    15,
    4
  ]
} | false | 
| 
	ttxy/resume_ner | 
	default | 
	validation | 463 | 
	label | 9 | 859 | 145.27214 | 109 | 129.88859 | 
	{
  "bin_edges": [
    9,
    95,
    181,
    267,
    353,
    439,
    525,
    611,
    697,
    783,
    859
  ],
  "hist": [
    180,
    184,
    48,
    14,
    11,
    12,
    6,
    6,
    0,
    2
  ]
} | false | 
| 
	ttxy/resume_ner | 
	default | 
	validation | 463 | 
	text | 7 | 355 | 59 | 49 | 45.65881 | 
	{
  "bin_edges": [
    7,
    42,
    77,
    112,
    147,
    182,
    217,
    252,
    287,
    322,
    355
  ],
  "hist": [
    184,
    186,
    49,
    17,
    15,
    3,
    5,
    3,
    0,
    1
  ]
} | false | 
| 
	fedryanto/UnibQuADV2 | 
	plain_text | 
	train | 4,010 | 
	context | 65 | 3,845 | 1,107.15387 | 827 | 816.68357 | 
	{
  "bin_edges": [
    65,
    444,
    823,
    1202,
    1581,
    1960,
    2339,
    2718,
    3097,
    3476,
    3845
  ],
  "hist": [
    797,
    1184,
    599,
    594,
    212,
    148,
    198,
    162,
    71,
    45
  ]
} | false | 
| 
	fedryanto/UnibQuADV2 | 
	plain_text | 
	train | 4,010 | 
	id | 7 | 7 | 7 | 7 | 0 | 
	{
  "bin_edges": [
    7,
    7
  ],
  "hist": [
    4010
  ]
} | false | 
| 
	fedryanto/UnibQuADV2 | 
	plain_text | 
	train | 4,010 | 
	question | 11 | 147 | 50.14988 | 47 | 17.43961 | 
	{
  "bin_edges": [
    11,
    25,
    39,
    53,
    67,
    81,
    95,
    109,
    123,
    137,
    147
  ],
  "hist": [
    102,
    1035,
    1360,
    906,
    332,
    174,
    82,
    16,
    2,
    1
  ]
} | false | 
| 
	fedryanto/UnibQuADV2 | 
	plain_text | 
	validation | 1,036 | 
	context | 73 | 4,188 | 1,119.86873 | 620 | 1,219.10519 | 
	{
  "bin_edges": [
    73,
    485,
    897,
    1309,
    1721,
    2133,
    2545,
    2957,
    3369,
    3781,
    4188
  ],
  "hist": [
    382,
    322,
    79,
    63,
    2,
    68,
    0,
    1,
    0,
    119
  ]
} | false | 
| 
	fedryanto/UnibQuADV2 | 
	plain_text | 
	validation | 1,036 | 
	id | 7 | 7 | 7 | 7 | 0 | 
	{
  "bin_edges": [
    7,
    7
  ],
  "hist": [
    1036
  ]
} | false | 
| 
	fedryanto/UnibQuADV2 | 
	plain_text | 
	validation | 1,036 | 
	question | 13 | 120 | 50.65251 | 47 | 16.24953 | 
	{
  "bin_edges": [
    13,
    24,
    35,
    46,
    57,
    68,
    79,
    90,
    101,
    112,
    120
  ],
  "hist": [
    12,
    67,
    397,
    256,
    168,
    64,
    41,
    18,
    10,
    3
  ]
} | false | 
| 
	redwoodresearch/generated_stories | 
	default | 
	train | 3,825 | 
	text | 3,130 | 6,715 | 4,849.41752 | 4,919 | 512.67936 | 
	{
  "bin_edges": [
    3130,
    3489,
    3848,
    4207,
    4566,
    4925,
    5284,
    5643,
    6002,
    6361,
    6715
  ],
  "hist": [
    32,
    162,
    301,
    411,
    1031,
    1228,
    508,
    123,
    25,
    4
  ]
} | false | 
| 
	redwoodresearch/generated_stories | 
	default | 
	validation | 675 | 
	text | 2,838 | 6,655 | 4,844.32 | 4,915 | 514.73498 | 
	{
  "bin_edges": [
    2838,
    3220,
    3602,
    3984,
    4366,
    4748,
    5130,
    5512,
    5894,
    6276,
    6655
  ],
  "hist": [
    1,
    8,
    47,
    54,
    132,
    231,
    155,
    40,
    6,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	accountant | 
	test | 443 | 
	A | 1 | 110 | 17.36795 | 13 | 15.60652 | 
	{
  "bin_edges": [
    1,
    12,
    23,
    34,
    45,
    56,
    67,
    78,
    89,
    100,
    110
  ],
  "hist": [
    212,
    96,
    77,
    32,
    11,
    9,
    3,
    2,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	accountant | 
	test | 443 | 
	B | 1 | 104 | 17.89391 | 13 | 15.71883 | 
	{
  "bin_edges": [
    1,
    12,
    23,
    34,
    45,
    56,
    67,
    78,
    89,
    100,
    104
  ],
  "hist": [
    207,
    95,
    71,
    41,
    16,
    7,
    4,
    1,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	accountant | 
	test | 443 | 
	C | 1 | 114 | 18.18284 | 14 | 16.41089 | 
	{
  "bin_edges": [
    1,
    13,
    25,
    37,
    49,
    61,
    73,
    85,
    97,
    109,
    114
  ],
  "hist": [
    212,
    108,
    70,
    27,
    14,
    8,
    2,
    0,
    1,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	accountant | 
	test | 443 | 
	D | 1 | 94 | 17.85553 | 13 | 15.44108 | 
	{
  "bin_edges": [
    1,
    11,
    21,
    31,
    41,
    51,
    61,
    71,
    81,
    91,
    94
  ],
  "hist": [
    194,
    97,
    69,
    42,
    24,
    9,
    4,
    3,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	accountant | 
	test | 443 | 
	question | 14 | 280 | 58.45372 | 39 | 42.39338 | 
	{
  "bin_edges": [
    14,
    41,
    68,
    95,
    122,
    149,
    176,
    203,
    230,
    257,
    280
  ],
  "hist": [
    227,
    82,
    46,
    42,
    27,
    14,
    1,
    2,
    1,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	accountant | 
	val | 49 | 
	A | 1 | 78 | 16.63265 | 8 | 17.51986 | 
	{
  "bin_edges": [
    1,
    9,
    17,
    25,
    33,
    41,
    49,
    57,
    65,
    73,
    78
  ],
  "hist": [
    26,
    5,
    6,
    5,
    1,
    2,
    2,
    1,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	accountant | 
	val | 49 | 
	B | 2 | 94 | 17.79592 | 9 | 20.10077 | 
	{
  "bin_edges": [
    2,
    12,
    22,
    32,
    42,
    52,
    62,
    72,
    82,
    92,
    94
  ],
  "hist": [
    28,
    9,
    5,
    0,
    3,
    2,
    0,
    0,
    1,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	accountant | 
	val | 49 | 
	C | 1 | 64 | 18.53061 | 10 | 17.97927 | 
	{
  "bin_edges": [
    1,
    8,
    15,
    22,
    29,
    36,
    43,
    50,
    57,
    64,
    64
  ],
  "hist": [
    20,
    7,
    8,
    2,
    3,
    3,
    0,
    4,
    1,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	accountant | 
	val | 49 | 
	D | 1 | 75 | 17.46939 | 11 | 16.60585 | 
	{
  "bin_edges": [
    1,
    9,
    17,
    25,
    33,
    41,
    49,
    57,
    65,
    73,
    75
  ],
  "hist": [
    22,
    9,
    5,
    4,
    3,
    3,
    2,
    0,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	accountant | 
	val | 49 | 
	question | 17 | 206 | 59.4898 | 38 | 48.18589 | 
	{
  "bin_edges": [
    17,
    36,
    55,
    74,
    93,
    112,
    131,
    150,
    169,
    188,
    206
  ],
  "hist": [
    23,
    8,
    5,
    3,
    4,
    2,
    0,
    1,
    1,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	advanced_mathematics | 
	test | 173 | 
	A | 1 | 215 | 31.00578 | 18 | 35.88637 | 
	{
  "bin_edges": [
    1,
    23,
    45,
    67,
    89,
    111,
    133,
    155,
    177,
    199,
    215
  ],
  "hist": [
    109,
    23,
    17,
    14,
    2,
    4,
    1,
    1,
    0,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	advanced_mathematics | 
	test | 173 | 
	B | 1 | 215 | 30.72254 | 18 | 36.1361 | 
	{
  "bin_edges": [
    1,
    23,
    45,
    67,
    89,
    111,
    133,
    155,
    177,
    199,
    215
  ],
  "hist": [
    110,
    22,
    17,
    14,
    2,
    4,
    1,
    1,
    0,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	advanced_mathematics | 
	test | 173 | 
	C | 1 | 215 | 30.93642 | 18 | 35.80635 | 
	{
  "bin_edges": [
    1,
    23,
    45,
    67,
    89,
    111,
    133,
    155,
    177,
    199,
    215
  ],
  "hist": [
    110,
    22,
    17,
    14,
    2,
    4,
    2,
    0,
    0,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	advanced_mathematics | 
	test | 173 | 
	D | 1 | 215 | 31.77457 | 18 | 35.47552 | 
	{
  "bin_edges": [
    1,
    23,
    45,
    67,
    89,
    111,
    133,
    155,
    177,
    199,
    215
  ],
  "hist": [
    107,
    26,
    16,
    14,
    2,
    4,
    1,
    1,
    0,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	advanced_mathematics | 
	test | 173 | 
	question | 39 | 445 | 113.3237 | 96 | 61.42863 | 
	{
  "bin_edges": [
    39,
    80,
    121,
    162,
    203,
    244,
    285,
    326,
    367,
    408,
    445
  ],
  "hist": [
    55,
    56,
    36,
    11,
    6,
    5,
    2,
    1,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	art_studies | 
	test | 298 | 
	A | 1 | 35 | 3.9396 | 3 | 2.61626 | 
	{
  "bin_edges": [
    1,
    5,
    9,
    13,
    17,
    21,
    25,
    29,
    33,
    35
  ],
  "hist": [
    204,
    86,
    6,
    1,
    0,
    0,
    0,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	art_studies | 
	test | 298 | 
	B | 1 | 15 | 3.82886 | 3 | 2.01699 | 
	{
  "bin_edges": [
    1,
    3,
    5,
    7,
    9,
    11,
    13,
    15,
    15
  ],
  "hist": [
    88,
    113,
    73,
    14,
    6,
    1,
    2,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	art_studies | 
	test | 298 | 
	C | 1 | 22 | 3.97651 | 3 | 2.28804 | 
	{
  "bin_edges": [
    1,
    4,
    7,
    10,
    13,
    16,
    19,
    22,
    22
  ],
  "hist": [
    156,
    109,
    25,
    6,
    1,
    0,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	art_studies | 
	test | 298 | 
	D | 1 | 18 | 4.04027 | 3 | 2.23495 | 
	{
  "bin_edges": [
    1,
    3,
    5,
    7,
    9,
    11,
    13,
    15,
    17,
    18
  ],
  "hist": [
    74,
    123,
    61,
    28,
    8,
    1,
    2,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	art_studies | 
	test | 298 | 
	question | 10 | 79 | 22.69128 | 20 | 9.73687 | 
	{
  "bin_edges": [
    10,
    17,
    24,
    31,
    38,
    45,
    52,
    59,
    66,
    73,
    79
  ],
  "hist": [
    86,
    108,
    52,
    29,
    12,
    5,
    5,
    0,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	art_studies | 
	val | 33 | 
	A | 1 | 8 | 3.81818 | 3 | 1.81064 | 
	{
  "bin_edges": [
    1,
    2,
    3,
    4,
    5,
    6,
    7,
    8,
    8
  ],
  "hist": [
    1,
    9,
    7,
    5,
    4,
    4,
    2,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	art_studies | 
	val | 33 | 
	B | 1 | 8 | 3.84848 | 3 | 1.83918 | 
	{
  "bin_edges": [
    1,
    2,
    3,
    4,
    5,
    6,
    7,
    8,
    8
  ],
  "hist": [
    1,
    9,
    8,
    2,
    6,
    4,
    2,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	art_studies | 
	val | 33 | 
	C | 1 | 10 | 3.81818 | 3 | 1.9757 | 
	{
  "bin_edges": [
    1,
    2,
    3,
    4,
    5,
    6,
    7,
    8,
    9,
    10,
    10
  ],
  "hist": [
    1,
    9,
    8,
    5,
    3,
    4,
    2,
    0,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	art_studies | 
	val | 33 | 
	D | 2 | 7 | 3.48485 | 3 | 1.62252 | 
	{
  "bin_edges": [
    2,
    3,
    4,
    5,
    6,
    7,
    7
  ],
  "hist": [
    12,
    9,
    3,
    5,
    1,
    3
  ]
} | false | 
| 
	ceval/ceval-exam | 
	art_studies | 
	val | 33 | 
	question | 12 | 65 | 23.33333 | 21 | 11.14862 | 
	{
  "bin_edges": [
    12,
    18,
    24,
    30,
    36,
    42,
    48,
    54,
    60,
    65
  ],
  "hist": [
    12,
    9,
    6,
    3,
    1,
    0,
    1,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	basic_medicine | 
	test | 175 | 
	A | 1 | 29 | 6.75429 | 5 | 4.64265 | 
	{
  "bin_edges": [
    1,
    4,
    7,
    10,
    13,
    16,
    19,
    22,
    25,
    28,
    29
  ],
  "hist": [
    36,
    79,
    23,
    13,
    16,
    4,
    2,
    0,
    0,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	basic_medicine | 
	test | 175 | 
	B | 1 | 28 | 6.80571 | 5 | 4.29608 | 
	{
  "bin_edges": [
    1,
    4,
    7,
    10,
    13,
    16,
    19,
    22,
    25,
    28,
    28
  ],
  "hist": [
    38,
    66,
    26,
    27,
    11,
    6,
    0,
    0,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	basic_medicine | 
	test | 175 | 
	C | 1 | 28 | 7.28571 | 6 | 4.89294 | 
	{
  "bin_edges": [
    1,
    4,
    7,
    10,
    13,
    16,
    19,
    22,
    25,
    28,
    28
  ],
  "hist": [
    32,
    64,
    45,
    8,
    13,
    7,
    1,
    4,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	basic_medicine | 
	test | 175 | 
	D | 2 | 37 | 7.68571 | 6 | 5.57174 | 
	{
  "bin_edges": [
    2,
    6,
    10,
    14,
    18,
    22,
    26,
    30,
    34,
    37
  ],
  "hist": [
    83,
    49,
    21,
    11,
    6,
    2,
    1,
    1,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	basic_medicine | 
	test | 175 | 
	question | 7 | 66 | 21.19429 | 20 | 7.77167 | 
	{
  "bin_edges": [
    7,
    13,
    19,
    25,
    31,
    37,
    43,
    49,
    55,
    61,
    66
  ],
  "hist": [
    10,
    60,
    66,
    24,
    10,
    2,
    0,
    1,
    1,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	business_administration | 
	test | 301 | 
	A | 1 | 44 | 7.15282 | 5 | 7.09342 | 
	{
  "bin_edges": [
    1,
    6,
    11,
    16,
    21,
    26,
    31,
    36,
    41,
    44
  ],
  "hist": [
    180,
    66,
    21,
    17,
    6,
    4,
    4,
    0,
    3
  ]
} | false | 
| 
	ceval/ceval-exam | 
	business_administration | 
	test | 301 | 
	B | 1 | 74 | 7.37209 | 5 | 8.2487 | 
	{
  "bin_edges": [
    1,
    9,
    17,
    25,
    33,
    41,
    49,
    57,
    65,
    73,
    74
  ],
  "hist": [
    237,
    32,
    22,
    6,
    1,
    1,
    0,
    0,
    1,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	business_administration | 
	test | 301 | 
	C | 1 | 82 | 7.55482 | 4 | 9.21454 | 
	{
  "bin_edges": [
    1,
    10,
    19,
    28,
    37,
    46,
    55,
    64,
    73,
    82,
    82
  ],
  "hist": [
    243,
    32,
    13,
    9,
    1,
    0,
    0,
    2,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	business_administration | 
	test | 301 | 
	D | 1 | 71 | 7.49169 | 4 | 8.67779 | 
	{
  "bin_edges": [
    1,
    9,
    17,
    25,
    33,
    41,
    49,
    57,
    65,
    71
  ],
  "hist": [
    232,
    41,
    15,
    7,
    2,
    0,
    2,
    1,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	business_administration | 
	test | 301 | 
	question | 10 | 281 | 55.6711 | 37 | 47.00718 | 
	{
  "bin_edges": [
    10,
    38,
    66,
    94,
    122,
    150,
    178,
    206,
    234,
    262,
    281
  ],
  "hist": [
    153,
    72,
    21,
    20,
    15,
    11,
    5,
    3,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	business_administration | 
	val | 33 | 
	A | 1 | 46 | 9.33333 | 6 | 9.84463 | 
	{
  "bin_edges": [
    1,
    6,
    11,
    16,
    21,
    26,
    31,
    36,
    41,
    46,
    46
  ],
  "hist": [
    16,
    9,
    3,
    2,
    0,
    1,
    1,
    0,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	business_administration | 
	val | 33 | 
	B | 1 | 45 | 10.60606 | 6 | 10.99699 | 
	{
  "bin_edges": [
    1,
    6,
    11,
    16,
    21,
    26,
    31,
    36,
    41,
    45
  ],
  "hist": [
    16,
    7,
    2,
    3,
    2,
    0,
    1,
    0,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	business_administration | 
	val | 33 | 
	C | 2 | 62 | 12.54545 | 6 | 15.13706 | 
	{
  "bin_edges": [
    2,
    9,
    16,
    23,
    30,
    37,
    44,
    51,
    58,
    62
  ],
  "hist": [
    19,
    7,
    2,
    2,
    0,
    0,
    1,
    1,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	business_administration | 
	val | 33 | 
	D | 2 | 55 | 11.36364 | 6 | 12.36886 | 
	{
  "bin_edges": [
    2,
    8,
    14,
    20,
    26,
    32,
    38,
    44,
    50,
    55
  ],
  "hist": [
    22,
    2,
    3,
    2,
    1,
    1,
    1,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	business_administration | 
	val | 33 | 
	question | 13 | 171 | 45.66667 | 32 | 39.182 | 
	{
  "bin_edges": [
    13,
    29,
    45,
    61,
    77,
    93,
    109,
    125,
    141,
    157,
    171
  ],
  "hist": [
    14,
    10,
    2,
    1,
    2,
    1,
    1,
    0,
    1,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	chinese_language_and_literature | 
	test | 209 | 
	A | 1 | 24 | 5.35885 | 3 | 4.94785 | 
	{
  "bin_edges": [
    1,
    4,
    7,
    10,
    13,
    16,
    19,
    22,
    24
  ],
  "hist": [
    105,
    55,
    15,
    11,
    10,
    8,
    1,
    4
  ]
} | false | 
| 
	ceval/ceval-exam | 
	chinese_language_and_literature | 
	test | 209 | 
	B | 1 | 31 | 5.51196 | 4 | 5.37812 | 
	{
  "bin_edges": [
    1,
    5,
    9,
    13,
    17,
    21,
    25,
    29,
    31
  ],
  "hist": [
    133,
    35,
    20,
    10,
    6,
    2,
    1,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	chinese_language_and_literature | 
	test | 209 | 
	C | 1 | 37 | 5.65072 | 4 | 5.93373 | 
	{
  "bin_edges": [
    1,
    5,
    9,
    13,
    17,
    21,
    25,
    29,
    33,
    37,
    37
  ],
  "hist": [
    142,
    28,
    15,
    8,
    7,
    5,
    1,
    2,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	chinese_language_and_literature | 
	test | 209 | 
	D | 1 | 41 | 5.69378 | 4 | 5.80633 | 
	{
  "bin_edges": [
    1,
    6,
    11,
    16,
    21,
    26,
    31,
    36,
    41,
    41
  ],
  "hist": [
    152,
    27,
    13,
    10,
    4,
    1,
    1,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	chinese_language_and_literature | 
	test | 209 | 
	question | 8 | 86 | 23.27273 | 21 | 10.34767 | 
	{
  "bin_edges": [
    8,
    16,
    24,
    32,
    40,
    48,
    56,
    64,
    72,
    80,
    86
  ],
  "hist": [
    34,
    105,
    34,
    20,
    10,
    4,
    0,
    1,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	civil_servant | 
	test | 429 | 
	A | 1 | 85 | 13.5711 | 9 | 12.72876 | 
	{
  "bin_edges": [
    1,
    10,
    19,
    28,
    37,
    46,
    55,
    64,
    73,
    82,
    85
  ],
  "hist": [
    217,
    102,
    59,
    26,
    8,
    10,
    4,
    2,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	civil_servant | 
	test | 429 | 
	B | 1 | 78 | 14.18182 | 10 | 13.41831 | 
	{
  "bin_edges": [
    1,
    9,
    17,
    25,
    33,
    41,
    49,
    57,
    65,
    73,
    78
  ],
  "hist": [
    193,
    104,
    58,
    35,
    16,
    11,
    4,
    3,
    3,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	civil_servant | 
	test | 429 | 
	C | 1 | 78 | 13.89977 | 10 | 12.57377 | 
	{
  "bin_edges": [
    1,
    9,
    17,
    25,
    33,
    41,
    49,
    57,
    65,
    73,
    78
  ],
  "hist": [
    186,
    114,
    55,
    41,
    14,
    7,
    5,
    4,
    2,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	civil_servant | 
	test | 429 | 
	D | 1 | 69 | 14.71329 | 10 | 13.27581 | 
	{
  "bin_edges": [
    1,
    8,
    15,
    22,
    29,
    36,
    43,
    50,
    57,
    64,
    69
  ],
  "hist": [
    167,
    94,
    65,
    50,
    19,
    13,
    10,
    3,
    3,
    5
  ]
} | false | 
| 
	ceval/ceval-exam | 
	civil_servant | 
	test | 429 | 
	question | 9 | 315 | 81.8951 | 64 | 67.62804 | 
	{
  "bin_edges": [
    9,
    40,
    71,
    102,
    133,
    164,
    195,
    226,
    257,
    288,
    315
  ],
  "hist": [
    175,
    49,
    56,
    45,
    42,
    34,
    13,
    10,
    4,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	civil_servant | 
	val | 47 | 
	A | 2 | 32 | 11.42553 | 11 | 7.52025 | 
	{
  "bin_edges": [
    2,
    6,
    10,
    14,
    18,
    22,
    26,
    30,
    32
  ],
  "hist": [
    11,
    10,
    10,
    6,
    5,
    2,
    2,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	civil_servant | 
	val | 47 | 
	B | 2 | 50 | 11.76596 | 11 | 8.84704 | 
	{
  "bin_edges": [
    2,
    7,
    12,
    17,
    22,
    27,
    32,
    37,
    42,
    47,
    50
  ],
  "hist": [
    14,
    11,
    11,
    7,
    2,
    1,
    0,
    0,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	civil_servant | 
	val | 47 | 
	C | 2 | 30 | 11.59574 | 11 | 7.25816 | 
	{
  "bin_edges": [
    2,
    5,
    8,
    11,
    14,
    17,
    20,
    23,
    26,
    29,
    30
  ],
  "hist": [
    10,
    6,
    6,
    6,
    10,
    2,
    3,
    2,
    0,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	civil_servant | 
	val | 47 | 
	D | 2 | 32 | 12.23404 | 11 | 8.15932 | 
	{
  "bin_edges": [
    2,
    6,
    10,
    14,
    18,
    22,
    26,
    30,
    32
  ],
  "hist": [
    12,
    11,
    6,
    6,
    6,
    2,
    2,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	civil_servant | 
	val | 47 | 
	question | 13 | 283 | 100.97872 | 110 | 72.37508 | 
	{
  "bin_edges": [
    13,
    41,
    69,
    97,
    125,
    153,
    181,
    209,
    237,
    265,
    283
  ],
  "hist": [
    16,
    0,
    5,
    10,
    6,
    4,
    2,
    2,
    0,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	clinical_medicine | 
	test | 200 | 
	A | 1 | 32 | 7.91 | 6 | 5.97498 | 
	{
  "bin_edges": [
    1,
    5,
    9,
    13,
    17,
    21,
    25,
    29,
    32
  ],
  "hist": [
    62,
    77,
    28,
    13,
    10,
    3,
    5,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	clinical_medicine | 
	test | 200 | 
	B | 1 | 32 | 7.97 | 6 | 5.65589 | 
	{
  "bin_edges": [
    1,
    5,
    9,
    13,
    17,
    21,
    25,
    29,
    32
  ],
  "hist": [
    58,
    71,
    31,
    22,
    8,
    8,
    0,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	clinical_medicine | 
	test | 200 | 
	C | 1 | 30 | 8.505 | 7 | 5.32369 | 
	{
  "bin_edges": [
    1,
    4,
    7,
    10,
    13,
    16,
    19,
    22,
    25,
    28,
    30
  ],
  "hist": [
    28,
    59,
    42,
    30,
    17,
    14,
    6,
    2,
    1,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	clinical_medicine | 
	test | 200 | 
	D | 1 | 26 | 8.63 | 7 | 5.40641 | 
	{
  "bin_edges": [
    1,
    4,
    7,
    10,
    13,
    16,
    19,
    22,
    25,
    26
  ],
  "hist": [
    31,
    57,
    41,
    28,
    14,
    19,
    4,
    5,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	clinical_medicine | 
	test | 200 | 
	question | 8 | 180 | 34.11 | 21 | 32.06664 | 
	{
  "bin_edges": [
    8,
    26,
    44,
    62,
    80,
    98,
    116,
    134,
    152,
    170,
    180
  ],
  "hist": [
    132,
    31,
    9,
    9,
    3,
    5,
    6,
    3,
    1,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	college_economics | 
	test | 497 | 
	A | 1 | 49 | 10.09256 | 8 | 7.23368 | 
	{
  "bin_edges": [
    1,
    6,
    11,
    16,
    21,
    26,
    31,
    36,
    41,
    46,
    49
  ],
  "hist": [
    156,
    149,
    99,
    47,
    29,
    10,
    2,
    3,
    1,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	college_economics | 
	test | 497 | 
	B | 1 | 42 | 10.38028 | 8 | 7.30487 | 
	{
  "bin_edges": [
    1,
    6,
    11,
    16,
    21,
    26,
    31,
    36,
    41,
    42
  ],
  "hist": [
    147,
    154,
    103,
    45,
    26,
    10,
    8,
    3,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	college_economics | 
	test | 497 | 
	C | 1 | 41 | 10.68813 | 9 | 7.68596 | 
	{
  "bin_edges": [
    1,
    6,
    11,
    16,
    21,
    26,
    31,
    36,
    41,
    41
  ],
  "hist": [
    147,
    140,
    100,
    52,
    30,
    14,
    8,
    5,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	college_economics | 
	test | 497 | 
	D | 1 | 51 | 10.23742 | 8 | 7.48385 | 
	{
  "bin_edges": [
    1,
    7,
    13,
    19,
    25,
    31,
    37,
    43,
    49,
    51
  ],
  "hist": [
    212,
    141,
    74,
    40,
    19,
    7,
    3,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	college_economics | 
	test | 497 | 
	question | 8 | 235 | 36.10664 | 27 | 27.96453 | 
	{
  "bin_edges": [
    8,
    31,
    54,
    77,
    100,
    123,
    146,
    169,
    192,
    215,
    235
  ],
  "hist": [
    293,
    122,
    39,
    19,
    17,
    4,
    1,
    0,
    0,
    2
  ]
} | false | 
| 
	ceval/ceval-exam | 
	college_economics | 
	val | 55 | 
	A | 1 | 44 | 11.03636 | 8 | 8.91724 | 
	{
  "bin_edges": [
    1,
    6,
    11,
    16,
    21,
    26,
    31,
    36,
    41,
    44
  ],
  "hist": [
    15,
    20,
    8,
    4,
    2,
    4,
    1,
    0,
    1
  ]
} | false | 
| 
	ceval/ceval-exam | 
	college_economics | 
	val | 55 | 
	B | 1 | 47 | 11.12727 | 9 | 8.15374 | 
	{
  "bin_edges": [
    1,
    6,
    11,
    16,
    21,
    26,
    31,
    36,
    41,
    46,
    47
  ],
  "hist": [
    16,
    15,
    11,
    7,
    5,
    0,
    0,
    0,
    0,
    1
  ]
} | false | 
			Subsets and Splits
				
	
				
			
				
No community queries yet
The top public SQL queries from the community will appear here once available.
