Dataset Preview
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code: DatasetGenerationError
Exception: TypeError
Message: Couldn't cast array of type string to null
Traceback: Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1872, in _prepare_split_single
for key, table in generator:
^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/json/json.py", line 289, in _generate_tables
self._cast_table(pa_table, json_field_paths=json_field_paths),
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/json/json.py", line 124, in _cast_table
pa_table = table_cast(pa_table, self.info.features.arrow_schema)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2272, in table_cast
return cast_table_to_schema(table, schema)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2224, in cast_table_to_schema
cast_array_to_feature(
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 1795, in wrapper
return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2002, in cast_array_to_feature
_c(array.field(name) if name in array_fields else null_array, subfeature)
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 1797, in wrapper
return func(array, *args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2086, in cast_array_to_feature
return array_cast(
^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 1797, in wrapper
return func(array, *args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 1948, in array_cast
raise TypeError(f"Couldn't cast array of type {_short_str(array.type)} to {_short_str(pa_type)}")
TypeError: Couldn't cast array of type string to null
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1347, in compute_config_parquet_and_info_response
parquet_operations = convert_to_parquet(builder)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 980, in convert_to_parquet
builder.download_and_prepare(
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 884, in download_and_prepare
self._download_and_prepare(
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 947, in _download_and_prepare
self._prepare_split(split_generator, **prepare_split_kwargs)
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1739, in _prepare_split
for job_id, done, content in self._prepare_split_single(
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1922, in _prepare_split_single
raise DatasetGenerationError("An error occurred while generating the dataset") from e
datasets.exceptions.DatasetGenerationError: An error occurred while generating the datasetNeed help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
benchmark_config_hash string | config dict | measurements dict | metadata dict |
|---|---|---|---|
a6ae95a4ff9612efe2180b1f76599e1105627d9226e6b752fbc58778bc9b037d | {
"attn_implementation": "flex_attention",
"batch_size": 1,
"compile_mode": "default",
"compile_options": {},
"gpu_monitoring": true,
"kernelize": false,
"measurement_iterations": 1,
"name": "w1_i1-monitored-b1_s32_n5-flex_attention-compiled_default-unkernelized",
"num_tokens_to_generate": 5,
"sdpa_... | {
"e2e_latency": [
0.16807457
],
"gpu_metrics": [
{
"memory_used": [
0.0000182875
],
"monitoring_status": "success",
"timestamp_0": 1760645527.3762913,
"timestamps": [
0
],
"utilization": [
0
]
}
],
"shape_and_decoded_outputs"... | {
"branch_name": "feat/benchmark_v2_ci",
"commit_id": "ed0ce751ce264343d65432bd98c77983203fa592",
"commit_message": "wip",
"hardware_info": {
"gpu_memory_total_gb": 79.4371337891,
"gpu_name": "NVIDIA H100 80GB HBM3",
"python_version": "3.13.5",
"torch_version": "2.7.1+cu126"
},
"model_id": "... |
c9dbfb6b2b27ac313972a368dcbc99c91035eeda27063ff87e7d84d267d60577 | {
"attn_implementation": "eager",
"batch_size": 1,
"compile_mode": "default",
"compile_options": {},
"gpu_monitoring": true,
"kernelize": false,
"measurement_iterations": 1,
"name": "w1_i1-monitored-b1_s32_n5-eager-compiled_default-unkernelized",
"num_tokens_to_generate": 5,
"sdpa_backend": null,
... | {
"e2e_latency": [
0.0749783169
],
"gpu_metrics": [
{
"memory_used": [
0.0000182856
],
"monitoring_status": "success",
"timestamp_0": 1760645557.6526833,
"timestamps": [
0
],
"utilization": [
0
]
}
],
"shape_and_decoded_output... | {
"branch_name": "feat/benchmark_v2_ci",
"commit_id": "ed0ce751ce264343d65432bd98c77983203fa592",
"commit_message": "wip",
"hardware_info": {
"gpu_memory_total_gb": 79.4371337891,
"gpu_name": "NVIDIA H100 80GB HBM3",
"python_version": "3.13.5",
"torch_version": "2.7.1+cu126"
},
"model_id": "... |
287b04d98eb051e4e2310293a8561a23cd754afb7d8b6fb8be6c3fc199556d9a | {
"attn_implementation": "flash_attention_2",
"batch_size": 1,
"compile_mode": null,
"compile_options": {},
"gpu_monitoring": true,
"kernelize": false,
"measurement_iterations": 1,
"name": "w1_i1-monitored-b1_s32_n5-flash_attention_2-uncompiled-unkernelized",
"num_tokens_to_generate": 5,
"sdpa_backe... | {
"e2e_latency": [
0.1649869829
],
"gpu_metrics": [
{
"memory_used": [
0.0000182837
],
"monitoring_status": "success",
"timestamp_0": 1760645566.5239754,
"timestamps": [
0
],
"utilization": [
3
]
}
],
"shape_and_decoded_output... | {
"branch_name": "feat/benchmark_v2_ci",
"commit_id": "ed0ce751ce264343d65432bd98c77983203fa592",
"commit_message": "wip",
"hardware_info": {
"gpu_memory_total_gb": 79.4371337891,
"gpu_name": "NVIDIA H100 80GB HBM3",
"python_version": "3.13.5",
"torch_version": "2.7.1+cu126"
},
"model_id": "... |
a6ae95a4ff9612efe2180b1f76599e1105627d9226e6b752fbc58778bc9b037d | {
"attn_implementation": "flex_attention",
"batch_size": 1,
"compile_mode": "default",
"compile_options": {},
"gpu_monitoring": true,
"kernelize": false,
"measurement_iterations": 1,
"name": "w1_i1-monitored-b1_s32_n5-flex_attention-compiled_default-unkernelized",
"num_tokens_to_generate": 5,
"sdpa_... | {
"e2e_latency": [
0.16584115
],
"gpu_metrics": [
{
"memory_used": [
0.0000182875
],
"monitoring_status": "success",
"timestamp_0": 1760647170.1845338,
"timestamps": [
0
],
"utilization": [
0
]
}
],
"shape_and_decoded_outputs"... | {
"branch_name": "feat/benchmark_v2_ci",
"commit_id": "172461638063bd19075f5737c25a4877623d6053",
"commit_message": "wip",
"hardware_info": {
"gpu_memory_total_gb": 79.4371337891,
"gpu_name": "NVIDIA H100 80GB HBM3",
"python_version": "3.13.5",
"torch_version": "2.7.1+cu126"
},
"model_id": "... |
c9dbfb6b2b27ac313972a368dcbc99c91035eeda27063ff87e7d84d267d60577 | {
"attn_implementation": "eager",
"batch_size": 1,
"compile_mode": "default",
"compile_options": {},
"gpu_monitoring": true,
"kernelize": false,
"measurement_iterations": 1,
"name": "w1_i1-monitored-b1_s32_n5-eager-compiled_default-unkernelized",
"num_tokens_to_generate": 5,
"sdpa_backend": null,
... | {
"e2e_latency": [
0.075524445
],
"gpu_metrics": [
{
"memory_used": [
0.0000182856
],
"monitoring_status": "success",
"timestamp_0": 1760647199.7972832,
"timestamps": [
0
],
"utilization": [
0
]
}
],
"shape_and_decoded_outputs... | {
"branch_name": "feat/benchmark_v2_ci",
"commit_id": "172461638063bd19075f5737c25a4877623d6053",
"commit_message": "wip",
"hardware_info": {
"gpu_memory_total_gb": 79.4371337891,
"gpu_name": "NVIDIA H100 80GB HBM3",
"python_version": "3.13.5",
"torch_version": "2.7.1+cu126"
},
"model_id": "... |
287b04d98eb051e4e2310293a8561a23cd754afb7d8b6fb8be6c3fc199556d9a | {
"attn_implementation": "flash_attention_2",
"batch_size": 1,
"compile_mode": null,
"compile_options": {},
"gpu_monitoring": true,
"kernelize": false,
"measurement_iterations": 1,
"name": "w1_i1-monitored-b1_s32_n5-flash_attention_2-uncompiled-unkernelized",
"num_tokens_to_generate": 5,
"sdpa_backe... | {
"e2e_latency": [
0.1624490441
],
"gpu_metrics": [
{
"memory_used": [
0.0000182837
],
"monitoring_status": "success",
"timestamp_0": 1760647208.4838204,
"timestamps": [
0
],
"utilization": [
8
]
}
],
"shape_and_decoded_output... | {
"branch_name": "feat/benchmark_v2_ci",
"commit_id": "172461638063bd19075f5737c25a4877623d6053",
"commit_message": "wip",
"hardware_info": {
"gpu_memory_total_gb": 79.4371337891,
"gpu_name": "NVIDIA H100 80GB HBM3",
"python_version": "3.13.5",
"torch_version": "2.7.1+cu126"
},
"model_id": "... |
a6ae95a4ff9612efe2180b1f76599e1105627d9226e6b752fbc58778bc9b037d | {
"attn_implementation": "flex_attention",
"batch_size": 1,
"compile_mode": "default",
"compile_options": {},
"gpu_monitoring": true,
"kernelize": false,
"measurement_iterations": 1,
"name": "w1_i1-monitored-b1_s32_n5-flex_attention-compiled_default-unkernelized",
"num_tokens_to_generate": 5,
"sdpa_... | {
"e2e_latency": [
0.301411524
],
"gpu_metrics": [
{
"memory_used": [
0.0000149896,
0.000015012
],
"monitoring_status": "success",
"timestamp_0": 1760711175.5919824,
"timestamps": [
0,
0.2011740208
],
"utilization": [
36,
... | {
"branch_name": "feat/benchmark_v2_ci",
"commit_id": "17792fa5339a6eb3accfd08ed587b39438c3bbbb",
"commit_message": "feat: add benchmark v2 ci with results pushed to dataset",
"hardware_info": {
"gpu_memory_total_gb": 22.3012695312,
"gpu_name": "NVIDIA A10G",
"python_version": "3.10.12",
"torch_... |
c9dbfb6b2b27ac313972a368dcbc99c91035eeda27063ff87e7d84d267d60577 | {
"attn_implementation": "eager",
"batch_size": 1,
"compile_mode": "default",
"compile_options": {},
"gpu_monitoring": true,
"kernelize": false,
"measurement_iterations": 1,
"name": "w1_i1-monitored-b1_s32_n5-eager-compiled_default-unkernelized",
"num_tokens_to_generate": 5,
"sdpa_backend": null,
... | {
"e2e_latency": [
0.195601331
],
"gpu_metrics": [
{
"memory_used": [
0.0000150064
],
"monitoring_status": "success",
"timestamp_0": 1760711207.4833257,
"timestamps": [
0
],
"utilization": [
32
]
}
],
"shape_and_decoded_output... | {
"branch_name": "feat/benchmark_v2_ci",
"commit_id": "17792fa5339a6eb3accfd08ed587b39438c3bbbb",
"commit_message": "feat: add benchmark v2 ci with results pushed to dataset",
"hardware_info": {
"gpu_memory_total_gb": 22.3012695312,
"gpu_name": "NVIDIA A10G",
"python_version": "3.10.12",
"torch_... |
287b04d98eb051e4e2310293a8561a23cd754afb7d8b6fb8be6c3fc199556d9a | {
"attn_implementation": "flash_attention_2",
"batch_size": 1,
"compile_mode": null,
"compile_options": {},
"gpu_monitoring": true,
"kernelize": false,
"measurement_iterations": 1,
"name": "w1_i1-monitored-b1_s32_n5-flash_attention_2-uncompiled-unkernelized",
"num_tokens_to_generate": 5,
"sdpa_backe... | {
"e2e_latency": [
0.229721673
],
"gpu_metrics": [
{
"memory_used": [
0.0000149859,
0.0000149859
],
"monitoring_status": "success",
"timestamp_0": 1760711216.9743993,
"timestamps": [
0,
0.2014379501
],
"utilization": [
83,
... | {
"branch_name": "feat/benchmark_v2_ci",
"commit_id": "17792fa5339a6eb3accfd08ed587b39438c3bbbb",
"commit_message": "feat: add benchmark v2 ci with results pushed to dataset",
"hardware_info": {
"gpu_memory_total_gb": 22.3012695312,
"gpu_name": "NVIDIA A10G",
"python_version": "3.10.12",
"torch_... |
7270d2d09828b77f58bf99e97af10c4e2c7e76e077cfe9f2df4e7a9745446502 | {
"attn_implementation": "flex_attention",
"batch_size": 1,
"compile_mode": "default",
"compile_options": {},
"gpu_monitoring": false,
"kernelize": false,
"measurement_iterations": 1,
"name": "w1_i1-unmonitored-b1_s32_n5-flex_attention-compiled_default-unkernelized",
"num_tokens_to_generate": 5,
"sd... | {
"e2e_latency": [
0.164014553
],
"gpu_metrics": null,
"shape_and_decoded_outputs": [
"(1, 37) | 18 Brumaire in"
],
"token_generation_times": [
[
0.1293741519,
0.138163317,
0.146759828,
0.15533767,
0.16386757,
0.163985402
]
]
} | {
"branch_name": "feat/benchmark_v2_ci",
"commit_id": "172461638063bd19075f5737c25a4877623d6053",
"commit_message": "wip",
"hardware_info": {
"gpu_memory_total_gb": 79.4371337891,
"gpu_name": "NVIDIA H100 80GB HBM3",
"python_version": "3.13.5",
"torch_version": "2.7.1+cu126"
},
"model_id": "... |
b48c4a4b4391aab327d4192d9cc4b767758bae3e1387f1bdeb8b80388f3f269e | {
"attn_implementation": "eager",
"batch_size": 1,
"compile_mode": "default",
"compile_options": {},
"gpu_monitoring": false,
"kernelize": false,
"measurement_iterations": 1,
"name": "w1_i1-unmonitored-b1_s32_n5-eager-compiled_default-unkernelized",
"num_tokens_to_generate": 5,
"sdpa_backend": null,... | {
"e2e_latency": [
0.0730237979
],
"gpu_metrics": null,
"shape_and_decoded_outputs": [
"(1, 37) | 18 Brumaire in"
],
"token_generation_times": [
[
0.0357038829,
0.045202032,
0.054447157,
0.063692251,
0.072889665,
0.073005287
]
]
} | {
"branch_name": "feat/benchmark_v2_ci",
"commit_id": "172461638063bd19075f5737c25a4877623d6053",
"commit_message": "wip",
"hardware_info": {
"gpu_memory_total_gb": 79.4371337891,
"gpu_name": "NVIDIA H100 80GB HBM3",
"python_version": "3.13.5",
"torch_version": "2.7.1+cu126"
},
"model_id": "... |
13d589b91f3120dd28f7dbe6fd35924a4ea3631ace3653805107fc833b7a6a3a | {
"attn_implementation": "flash_attention_2",
"batch_size": 1,
"compile_mode": null,
"compile_options": {},
"gpu_monitoring": false,
"kernelize": false,
"measurement_iterations": 1,
"name": "w1_i1-unmonitored-b1_s32_n5-flash_attention_2-uncompiled-unkernelized",
"num_tokens_to_generate": 5,
"sdpa_ba... | {
"e2e_latency": [
0.159017964
],
"gpu_metrics": null,
"shape_and_decoded_outputs": [
"(1, 37) | 18 Brumaire in"
],
"token_generation_times": [
[
0.0329530181,
0.064417906,
0.095871574,
0.127355222,
0.15882957,
0.1589581331
]
]
} | {
"branch_name": "feat/benchmark_v2_ci",
"commit_id": "172461638063bd19075f5737c25a4877623d6053",
"commit_message": "wip",
"hardware_info": {
"gpu_memory_total_gb": 79.4371337891,
"gpu_name": "NVIDIA H100 80GB HBM3",
"python_version": "3.13.5",
"torch_version": "2.7.1+cu126"
},
"model_id": "... |
559718c572a3306e03673f1ceaea0502f26fa89dcf56c6d322b088ce890c3572 | {
"attn_implementation": "flex_attention",
"batch_size": 16,
"compile_mode": "default",
"compile_options": {},
"gpu_monitoring": false,
"kernelize": false,
"measurement_iterations": 7,
"name": "w3_i7-unmonitored-b16_s64_n512-flex_attention-compiled_default-unkernelized",
"num_tokens_to_generate": 512,... | {
"e2e_latency": [
4.8839595449,
4.876068968,
4.8807041399,
4.887873122,
4.881046177,
4.877157859,
4.87416037
],
"gpu_metrics": null,
"shape_and_decoded_outputs": [
"(16, 576) | modern French politics.\nThe French Revolution was a complex and multifaceted event, and its causes a... | {
"branch_name": "feat/benchmark_v2_ci",
"commit_id": "172461638063bd19075f5737c25a4877623d6053",
"commit_message": "wip",
"hardware_info": {
"gpu_memory_total_gb": 79.4371337891,
"gpu_name": "NVIDIA H100 80GB HBM3",
"python_version": "3.13.5",
"torch_version": "2.7.1+cu126"
},
"model_id": "... |
9b17330710998ec756daac57d9a57125fcf457de3592ee5bb9a87aa01ec6ad4a | {
"attn_implementation": "eager",
"batch_size": 16,
"compile_mode": "default",
"compile_options": {},
"gpu_monitoring": false,
"kernelize": false,
"measurement_iterations": 7,
"name": "w3_i7-unmonitored-b16_s64_n512-eager-compiled_default-unkernelized",
"num_tokens_to_generate": 512,
"sdpa_backend":... | {
"e2e_latency": [
7.389178365,
7.393108413,
7.3910297221,
7.398342408,
7.385578134,
7.383540092,
7.3863331081
],
"gpu_metrics": null,
"shape_and_decoded_outputs": [
"(16, 576) | modern French politics.\nThe French Revolution was a complex and multifaceted event that involved a ... | {
"branch_name": "feat/benchmark_v2_ci",
"commit_id": "172461638063bd19075f5737c25a4877623d6053",
"commit_message": "wip",
"hardware_info": {
"gpu_memory_total_gb": 79.4371337891,
"gpu_name": "NVIDIA H100 80GB HBM3",
"python_version": "3.13.5",
"torch_version": "2.7.1+cu126"
},
"model_id": "... |
End of preview.
README.md exists but content is empty.
- Downloads last month
- 2,371