The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code: DatasetGenerationError Exception: ArrowNotImplementedError Message: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. Traceback: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2013, in _prepare_split_single writer.write_table(table) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 583, in write_table self._build_writer(inferred_schema=pa_table.schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 404, in _build_writer self.pa_writer = self._WRITER_CLASS(self.stream, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__ self.writer = _parquet.ParquetWriter( File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__ File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2029, in _prepare_split_single num_examples, num_bytes = writer.finalize() File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 602, in finalize self._build_writer(self.schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 404, in _build_writer self.pa_writer = self._WRITER_CLASS(self.stream, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__ self.writer = _parquet.ParquetWriter( File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__ File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1396, in compute_config_parquet_and_info_response parquet_operations = convert_to_parquet(builder) File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1045, in convert_to_parquet builder.download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1029, in download_and_prepare self._download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1124, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1884, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2040, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
config
dict | report
dict | name
string | backend
dict | scenario
dict | launcher
dict | environment
dict | overall
dict | warmup
dict | train
dict |
---|---|---|---|---|---|---|---|---|---|
{
"name": "cuda_training_transformers_fill-mask_google-bert/bert-base-uncased",
"backend": {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"library": "transformers",
"model_type": "bert",
"model": "google-bert/bert-base-uncased",
"processor": "google-bert/bert-base-uncased",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.12.0",
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1290.661888,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.5065605163574219,
"mean": 0.10131210327148439,
"stdev": 0.11552794326120178,
"p50": 0.04341964721679688,
"p90": 0.2171164688110352,
"p95": 0.27474166030883784,
"p99": 0.3208418135070801,
"values": [
0.3323668518066406,
0.044240894317626955,
0.04330086517333984,
0.043232257843017576,
0.04341964721679688
]
},
"throughput": {
"unit": "samples/s",
"value": 98.70488991037095
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1290.661888,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.37660774612426756,
"mean": 0.18830387306213378,
"stdev": 0.14406297874450683,
"p50": 0.18830387306213378,
"p90": 0.3035542560577392,
"p95": 0.3179605539321899,
"p99": 0.32948559223175045,
"values": [
0.3323668518066406,
0.044240894317626955
]
},
"throughput": {
"unit": "samples/s",
"value": 21.242260899647764
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1290.661888,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.1299527702331543,
"mean": 0.0433175900777181,
"stdev": 0.00007741010343702389,
"p50": 0.04330086517333984,
"p90": 0.04339589080810547,
"p95": 0.04340776901245117,
"p99": 0.043417271575927736,
"values": [
0.04330086517333984,
0.043232257843017576,
0.04341964721679688
]
},
"throughput": {
"unit": "samples/s",
"value": 138.51186063756367
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_fill-mask_google-bert/bert-base-uncased | {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"library": "transformers",
"model_type": "bert",
"model": "google-bert/bert-base-uncased",
"processor": "google-bert/bert-base-uncased",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.12.0",
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1290.661888,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.5065605163574219,
"mean": 0.10131210327148439,
"stdev": 0.11552794326120178,
"p50": 0.04341964721679688,
"p90": 0.2171164688110352,
"p95": 0.27474166030883784,
"p99": 0.3208418135070801,
"values": [
0.3323668518066406,
0.044240894317626955,
0.04330086517333984,
0.043232257843017576,
0.04341964721679688
]
},
"throughput": {
"unit": "samples/s",
"value": 98.70488991037095
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1290.661888,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.37660774612426756,
"mean": 0.18830387306213378,
"stdev": 0.14406297874450683,
"p50": 0.18830387306213378,
"p90": 0.3035542560577392,
"p95": 0.3179605539321899,
"p99": 0.32948559223175045,
"values": [
0.3323668518066406,
0.044240894317626955
]
},
"throughput": {
"unit": "samples/s",
"value": 21.242260899647764
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1290.661888,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.1299527702331543,
"mean": 0.0433175900777181,
"stdev": 0.00007741010343702389,
"p50": 0.04330086517333984,
"p90": 0.04339589080810547,
"p95": 0.04340776901245117,
"p99": 0.043417271575927736,
"values": [
0.04330086517333984,
0.043232257843017576,
0.04341964721679688
]
},
"throughput": {
"unit": "samples/s",
"value": 138.51186063756367
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_fill-mask_google-bert/bert-base-uncased",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"model": "google-bert/bert-base-uncased",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1063.8336,
"max_global_vram": 3169.32096,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.7448790740966797,
"mean": 0.14897581481933594,
"stdev": 0.2054173633207176,
"p50": 0.04632883071899414,
"p90": 0.35471870422363283,
"p95": 0.4572641067504882,
"p99": 0.5393004287719726,
"values": [
0.5598095092773437,
0.04708249664306641,
0.04632883071899414,
0.04576665496826172,
0.04589158248901367
]
},
"throughput": {
"unit": "samples/s",
"value": 67.1249894630687
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1063.8336,
"max_global_vram": 3169.32096,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.6068920059204101,
"mean": 0.3034460029602051,
"stdev": 0.25636350631713867,
"p50": 0.3034460029602051,
"p90": 0.508536808013916,
"p95": 0.5341731586456299,
"p99": 0.554682239151001,
"values": [
0.5598095092773437,
0.04708249664306641
]
},
"throughput": {
"unit": "samples/s",
"value": 13.181916917602548
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1063.8336,
"max_global_vram": 3169.32096,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.13798706817626955,
"mean": 0.045995689392089846,
"stdev": 0.00024102431292157434,
"p50": 0.04589158248901367,
"p90": 0.04624138107299805,
"p95": 0.0462851058959961,
"p99": 0.04632008575439454,
"values": [
0.04632883071899414,
0.04576665496826172,
0.04589158248901367
]
},
"throughput": {
"unit": "samples/s",
"value": 130.4470066499722
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_image-classification_google/vit-base-patch16-224",
"backend": {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"library": "transformers",
"model_type": "vit",
"model": "google/vit-base-patch16-224",
"processor": "google/vit-base-patch16-224",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.12.0",
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1620.91008,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.484606990814209,
"mean": 0.0969213981628418,
"stdev": 0.11383927924646801,
"p50": 0.03995340728759766,
"p90": 0.21085471649169923,
"p95": 0.267727269744873,
"p99": 0.3132253123474121,
"values": [
0.32459982299804685,
0.039929855346679685,
0.03988684844970703,
0.040237056732177735,
0.03995340728759766
]
},
"throughput": {
"unit": "samples/s",
"value": 103.17639024561501
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1620.91008,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.36452967834472655,
"mean": 0.18226483917236327,
"stdev": 0.1423349838256836,
"p50": 0.18226483917236327,
"p90": 0.29613282623291015,
"p95": 0.3103663246154785,
"p99": 0.32175312332153316,
"values": [
0.32459982299804685,
0.039929855346679685
]
},
"throughput": {
"unit": "samples/s",
"value": 21.9460869038888
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1620.91008,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.12007731246948242,
"mean": 0.040025770823160806,
"stdev": 0.00015185260668856861,
"p50": 0.03995340728759766,
"p90": 0.04018032684326172,
"p95": 0.04020869178771973,
"p99": 0.04023138374328614,
"values": [
0.03988684844970703,
0.040237056732177735,
0.03995340728759766
]
},
"throughput": {
"unit": "samples/s",
"value": 149.90342163574564
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_image-classification_google/vit-base-patch16-224 | {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"library": "transformers",
"model_type": "vit",
"model": "google/vit-base-patch16-224",
"processor": "google/vit-base-patch16-224",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.12.0",
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1620.91008,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.484606990814209,
"mean": 0.0969213981628418,
"stdev": 0.11383927924646801,
"p50": 0.03995340728759766,
"p90": 0.21085471649169923,
"p95": 0.267727269744873,
"p99": 0.3132253123474121,
"values": [
0.32459982299804685,
0.039929855346679685,
0.03988684844970703,
0.040237056732177735,
0.03995340728759766
]
},
"throughput": {
"unit": "samples/s",
"value": 103.17639024561501
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1620.91008,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.36452967834472655,
"mean": 0.18226483917236327,
"stdev": 0.1423349838256836,
"p50": 0.18226483917236327,
"p90": 0.29613282623291015,
"p95": 0.3103663246154785,
"p99": 0.32175312332153316,
"values": [
0.32459982299804685,
0.039929855346679685
]
},
"throughput": {
"unit": "samples/s",
"value": 21.9460869038888
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1620.91008,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.12007731246948242,
"mean": 0.040025770823160806,
"stdev": 0.00015185260668856861,
"p50": 0.03995340728759766,
"p90": 0.04018032684326172,
"p95": 0.04020869178771973,
"p99": 0.04023138374328614,
"values": [
0.03988684844970703,
0.040237056732177735,
0.03995340728759766
]
},
"throughput": {
"unit": "samples/s",
"value": 149.90342163574564
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_image-classification_google/vit-base-patch16-224",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"model": "google/vit-base-patch16-224",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1446.354944,
"max_global_vram": 2628.255744,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1756.126208
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.48406525039672854,
"mean": 0.09681305007934571,
"stdev": 0.1110534407118009,
"p50": 0.04146995162963867,
"p90": 0.20794796142578126,
"p95": 0.26343380432128904,
"p99": 0.3078224786376953,
"values": [
0.3189196472167969,
0.04103168106079102,
0.041490432739257815,
0.04146995162963867,
0.04115353775024414
]
},
"throughput": {
"unit": "samples/s",
"value": 103.29185984538483
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1446.354944,
"max_global_vram": 2628.255744,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1756.126208
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.3599513282775879,
"mean": 0.17997566413879396,
"stdev": 0.13894398307800293,
"p50": 0.17997566413879396,
"p90": 0.2911308506011963,
"p95": 0.30502524890899657,
"p99": 0.31614076755523685,
"values": [
0.3189196472167969,
0.04103168106079102
]
},
"throughput": {
"unit": "samples/s",
"value": 22.225227055782792
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1446.354944,
"max_global_vram": 2628.255744,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1756.126208
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.12411392211914063,
"mean": 0.041371307373046874,
"stdev": 0.00015421321911462263,
"p50": 0.04146995162963867,
"p90": 0.04148633651733399,
"p95": 0.0414883846282959,
"p99": 0.041490023117065435,
"values": [
0.041490432739257815,
0.04146995162963867,
0.04115353775024414
]
},
"throughput": {
"unit": "samples/s",
"value": 145.02804917180256
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_multiple-choice_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.12.0",
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1309.253632,
"max_global_vram": 3386.376192,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.5473637313842773,
"mean": 0.10947274627685546,
"stdev": 0.11701307238000899,
"p50": 0.050726913452148435,
"p90": 0.22695501251220704,
"p95": 0.28522533950805656,
"p99": 0.3318416011047363,
"values": [
0.3434956665039062,
0.052144031524658206,
0.05045043182373047,
0.050726913452148435,
0.050546688079833986
]
},
"throughput": {
"unit": "samples/s",
"value": 91.34693647595267
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1309.253632,
"max_global_vram": 3386.376192,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.3956396980285644,
"mean": 0.1978198490142822,
"stdev": 0.14567581748962402,
"p50": 0.1978198490142822,
"p90": 0.31436050300598145,
"p95": 0.32892808475494384,
"p99": 0.34058215015411375,
"values": [
0.3434956665039062,
0.052144031524658206
]
},
"throughput": {
"unit": "samples/s",
"value": 20.22041781920078
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1309.253632,
"max_global_vram": 3386.376192,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.15172403335571288,
"mean": 0.05057467778523763,
"stdev": 0.00011459520181810869,
"p50": 0.050546688079833986,
"p90": 0.05069086837768554,
"p95": 0.05070889091491699,
"p99": 0.050723308944702145,
"values": [
0.05045043182373047,
0.050726913452148435,
0.050546688079833986
]
},
"throughput": {
"unit": "samples/s",
"value": 118.63644540611102
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_multiple-choice_FacebookAI/roberta-base | {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.12.0",
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1309.253632,
"max_global_vram": 3386.376192,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.5473637313842773,
"mean": 0.10947274627685546,
"stdev": 0.11701307238000899,
"p50": 0.050726913452148435,
"p90": 0.22695501251220704,
"p95": 0.28522533950805656,
"p99": 0.3318416011047363,
"values": [
0.3434956665039062,
0.052144031524658206,
0.05045043182373047,
0.050726913452148435,
0.050546688079833986
]
},
"throughput": {
"unit": "samples/s",
"value": 91.34693647595267
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1309.253632,
"max_global_vram": 3386.376192,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.3956396980285644,
"mean": 0.1978198490142822,
"stdev": 0.14567581748962402,
"p50": 0.1978198490142822,
"p90": 0.31436050300598145,
"p95": 0.32892808475494384,
"p99": 0.34058215015411375,
"values": [
0.3434956665039062,
0.052144031524658206
]
},
"throughput": {
"unit": "samples/s",
"value": 20.22041781920078
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1309.253632,
"max_global_vram": 3386.376192,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.15172403335571288,
"mean": 0.05057467778523763,
"stdev": 0.00011459520181810869,
"p50": 0.050546688079833986,
"p90": 0.05069086837768554,
"p95": 0.05070889091491699,
"p99": 0.050723308944702145,
"values": [
0.05045043182373047,
0.050726913452148435,
0.050546688079833986
]
},
"throughput": {
"unit": "samples/s",
"value": 118.63644540611102
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_multiple-choice_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"model": "FacebookAI/roberta-base",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1093.496832,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.8026234703063965,
"mean": 0.16052469406127928,
"stdev": 0.22240891148008993,
"p50": 0.04907724761962891,
"p90": 0.38326721343994147,
"p95": 0.49430444412231433,
"p99": 0.5831342286682129,
"values": [
0.6053416748046875,
0.05015552139282226,
0.04897484970092773,
0.04907417678833008,
0.04907724761962891
]
},
"throughput": {
"unit": "samples/s",
"value": 62.295711313940046
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1093.496832,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.6554971961975098,
"mean": 0.3277485980987549,
"stdev": 0.27759307670593264,
"p50": 0.3277485980987549,
"p90": 0.549823059463501,
"p95": 0.5775823671340942,
"p99": 0.5997898132705688,
"values": [
0.6053416748046875,
0.05015552139282226
]
},
"throughput": {
"unit": "samples/s",
"value": 12.204476306546239
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1093.496832,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.14712627410888673,
"mean": 0.049042091369628914,
"stdev": 0.000047563564546161045,
"p50": 0.04907417678833008,
"p90": 0.04907663345336914,
"p95": 0.049076940536499025,
"p99": 0.04907718620300293,
"values": [
0.04897484970092773,
0.04907417678833008,
0.04907724761962891
]
},
"throughput": {
"unit": "samples/s",
"value": 122.34388527149387
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_text-classification_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.12.0",
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1295.552512,
"max_global_vram": 3386.376192,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.5399429092407227,
"mean": 0.10798858184814455,
"stdev": 0.11411113151864324,
"p50": 0.05082419204711914,
"p90": 0.22245437316894534,
"p95": 0.27933163299560543,
"p99": 0.3248334408569336,
"values": [
0.3362088928222656,
0.051822593688964844,
0.05060403060913086,
0.05048320007324219,
0.05082419204711914
]
},
"throughput": {
"unit": "samples/s",
"value": 92.60238285249616
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1295.552512,
"max_global_vram": 3386.376192,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.3880314865112305,
"mean": 0.19401574325561524,
"stdev": 0.1421931495666504,
"p50": 0.19401574325561524,
"p90": 0.30777026290893555,
"p95": 0.32198957786560056,
"p99": 0.33336502983093264,
"values": [
0.3362088928222656,
0.051822593688964844
]
},
"throughput": {
"unit": "samples/s",
"value": 20.616883624387174
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1295.552512,
"max_global_vram": 3386.376192,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.15191142272949218,
"mean": 0.05063714090983073,
"stdev": 0.00014116444446448864,
"p50": 0.05060403060913086,
"p90": 0.050780159759521486,
"p95": 0.05080217590332031,
"p99": 0.05081978881835937,
"values": [
0.05060403060913086,
0.05048320007324219,
0.05082419204711914
]
},
"throughput": {
"unit": "samples/s",
"value": 118.49010216995003
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_text-classification_FacebookAI/roberta-base | {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.12.0",
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1295.552512,
"max_global_vram": 3386.376192,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.5399429092407227,
"mean": 0.10798858184814455,
"stdev": 0.11411113151864324,
"p50": 0.05082419204711914,
"p90": 0.22245437316894534,
"p95": 0.27933163299560543,
"p99": 0.3248334408569336,
"values": [
0.3362088928222656,
0.051822593688964844,
0.05060403060913086,
0.05048320007324219,
0.05082419204711914
]
},
"throughput": {
"unit": "samples/s",
"value": 92.60238285249616
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1295.552512,
"max_global_vram": 3386.376192,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.3880314865112305,
"mean": 0.19401574325561524,
"stdev": 0.1421931495666504,
"p50": 0.19401574325561524,
"p90": 0.30777026290893555,
"p95": 0.32198957786560056,
"p99": 0.33336502983093264,
"values": [
0.3362088928222656,
0.051822593688964844
]
},
"throughput": {
"unit": "samples/s",
"value": 20.616883624387174
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1295.552512,
"max_global_vram": 3386.376192,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.15191142272949218,
"mean": 0.05063714090983073,
"stdev": 0.00014116444446448864,
"p50": 0.05060403060913086,
"p90": 0.050780159759521486,
"p95": 0.05080217590332031,
"p99": 0.05081978881835937,
"values": [
0.05060403060913086,
0.05048320007324219,
0.05082419204711914
]
},
"throughput": {
"unit": "samples/s",
"value": 118.49010216995003
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_text-classification_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"model": "FacebookAI/roberta-base",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1080.1152,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.778464241027832,
"mean": 0.1556928482055664,
"stdev": 0.21129125718980005,
"p50": 0.05020159912109375,
"p90": 0.36730080566406254,
"p95": 0.4727875488281249,
"p99": 0.557176943359375,
"values": [
0.5782742919921875,
0.050840576171875,
0.04969574356079102,
0.049452030181884765,
0.05020159912109375
]
},
"throughput": {
"unit": "samples/s",
"value": 64.22902602948513
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1080.1152,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.6291148681640625,
"mean": 0.31455743408203124,
"stdev": 0.26371685791015625,
"p50": 0.31455743408203124,
"p90": 0.5255309204101563,
"p95": 0.5519026062011718,
"p99": 0.5729999548339844,
"values": [
0.5782742919921875,
0.050840576171875
]
},
"throughput": {
"unit": "samples/s",
"value": 12.716278703357135
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1080.1152,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.14934937286376954,
"mean": 0.049783124287923176,
"stdev": 0.0003121857804388603,
"p50": 0.04969574356079102,
"p90": 0.05010042800903321,
"p95": 0.050151013565063476,
"p99": 0.0501914820098877,
"values": [
0.04969574356079102,
0.049452030181884765,
0.05020159912109375
]
},
"throughput": {
"unit": "samples/s",
"value": 120.52276922795568
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_text-generation_openai-community/gpt2",
"backend": {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "gpt2",
"model": "openai-community/gpt2",
"processor": "openai-community/gpt2",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.12.0",
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1339.682816,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.5305875473022461,
"mean": 0.10611750946044922,
"stdev": 0.12328177218832713,
"p50": 0.04456441497802734,
"p90": 0.22963568420410158,
"p95": 0.29115780792236323,
"p99": 0.34037550689697266,
"values": [
0.352679931640625,
0.045069313049316405,
0.04391113662719726,
0.04436275100708008,
0.04456441497802734
]
},
"throughput": {
"unit": "samples/s",
"value": 94.23515545026125
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1339.682816,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.3977492446899414,
"mean": 0.1988746223449707,
"stdev": 0.15380530929565428,
"p50": 0.1988746223449707,
"p90": 0.32191886978149414,
"p95": 0.33729940071105957,
"p99": 0.3496038254547119,
"values": [
0.352679931640625,
0.045069313049316405
]
},
"throughput": {
"unit": "samples/s",
"value": 20.113174586255376
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1339.682816,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.1328383026123047,
"mean": 0.04427943420410157,
"stdev": 0.0002731292955845127,
"p50": 0.04436275100708008,
"p90": 0.04452408218383789,
"p95": 0.044544248580932616,
"p99": 0.044560381698608396,
"values": [
0.04391113662719726,
0.04436275100708008,
0.04456441497802734
]
},
"throughput": {
"unit": "samples/s",
"value": 135.50308642932535
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_text-generation_openai-community/gpt2 | {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "gpt2",
"model": "openai-community/gpt2",
"processor": "openai-community/gpt2",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.12.0",
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1339.682816,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.5305875473022461,
"mean": 0.10611750946044922,
"stdev": 0.12328177218832713,
"p50": 0.04456441497802734,
"p90": 0.22963568420410158,
"p95": 0.29115780792236323,
"p99": 0.34037550689697266,
"values": [
0.352679931640625,
0.045069313049316405,
0.04391113662719726,
0.04436275100708008,
0.04456441497802734
]
},
"throughput": {
"unit": "samples/s",
"value": 94.23515545026125
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1339.682816,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.3977492446899414,
"mean": 0.1988746223449707,
"stdev": 0.15380530929565428,
"p50": 0.1988746223449707,
"p90": 0.32191886978149414,
"p95": 0.33729940071105957,
"p99": 0.3496038254547119,
"values": [
0.352679931640625,
0.045069313049316405
]
},
"throughput": {
"unit": "samples/s",
"value": 20.113174586255376
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1339.682816,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.1328383026123047,
"mean": 0.04427943420410157,
"stdev": 0.0002731292955845127,
"p50": 0.04436275100708008,
"p90": 0.04452408218383789,
"p95": 0.044544248580932616,
"p99": 0.044560381698608396,
"values": [
0.04391113662719726,
0.04436275100708008,
0.04456441497802734
]
},
"throughput": {
"unit": "samples/s",
"value": 135.50308642932535
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_text-generation_openai-community/gpt2",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"model": "openai-community/gpt2",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1107.173376,
"max_global_vram": 3563.585536,
"max_process_vram": 0,
"max_reserved": 2915.04128,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.8139294586181639,
"mean": 0.16278589172363278,
"stdev": 0.2273662362653359,
"p50": 0.04927385711669922,
"p90": 0.3902293930053711,
"p95": 0.5038737297058105,
"p99": 0.594789199066162,
"values": [
0.61751806640625,
0.049296382904052735,
0.04860006332397461,
0.0492410888671875,
0.04927385711669922
]
},
"throughput": {
"unit": "samples/s",
"value": 61.4303849929289
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1107.173376,
"max_global_vram": 3563.585536,
"max_process_vram": 0,
"max_reserved": 2915.04128,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.6668144493103026,
"mean": 0.3334072246551513,
"stdev": 0.2841108417510986,
"p50": 0.3334072246551513,
"p90": 0.5606958980560303,
"p95": 0.5891069822311401,
"p99": 0.6118358495712279,
"values": [
0.61751806640625,
0.049296382904052735
]
},
"throughput": {
"unit": "samples/s",
"value": 11.99734050195603
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1107.173376,
"max_global_vram": 3563.585536,
"max_process_vram": 0,
"max_reserved": 2915.04128,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.1471150093078613,
"mean": 0.04903833643595377,
"stdev": 0.00031019448743967263,
"p50": 0.0492410888671875,
"p90": 0.04926730346679687,
"p95": 0.049270580291748044,
"p99": 0.04927320175170898,
"values": [
0.04860006332397461,
0.0492410888671875,
0.04927385711669922
]
},
"throughput": {
"unit": "samples/s",
"value": 122.35325331307405
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_token-classification_microsoft/deberta-v3-base",
"backend": {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"library": "transformers",
"model_type": "deberta-v2",
"model": "microsoft/deberta-v3-base",
"processor": "microsoft/deberta-v3-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.12.0",
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1376.518144,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.7574855728149414,
"mean": 0.15149711456298828,
"stdev": 0.12952168445762077,
"p50": 0.08688435363769531,
"p90": 0.2814818298339844,
"p95": 0.3460093933105468,
"p99": 0.39763144409179685,
"values": [
0.4105369567871094,
0.08688435363769531,
0.08789913940429687,
0.08591667175292969,
0.08624845123291015
]
},
"throughput": {
"unit": "samples/s",
"value": 66.00785783178912
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1376.518144,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.4974213104248047,
"mean": 0.24871065521240235,
"stdev": 0.16182630157470704,
"p50": 0.24871065521240235,
"p90": 0.378171696472168,
"p95": 0.39435432662963865,
"p99": 0.40730043075561523,
"values": [
0.4105369567871094,
0.08688435363769531
]
},
"throughput": {
"unit": "samples/s",
"value": 16.082945849601597
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1376.518144,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.2600642623901367,
"mean": 0.0866880874633789,
"stdev": 0.0008669888536518835,
"p50": 0.08624845123291015,
"p90": 0.08756900177001953,
"p95": 0.08773407058715821,
"p99": 0.08786612564086914,
"values": [
0.08789913940429687,
0.08591667175292969,
0.08624845123291015
]
},
"throughput": {
"unit": "samples/s",
"value": 69.21366217168743
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_token-classification_microsoft/deberta-v3-base | {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"library": "transformers",
"model_type": "deberta-v2",
"model": "microsoft/deberta-v3-base",
"processor": "microsoft/deberta-v3-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.12.0",
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1376.518144,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.7574855728149414,
"mean": 0.15149711456298828,
"stdev": 0.12952168445762077,
"p50": 0.08688435363769531,
"p90": 0.2814818298339844,
"p95": 0.3460093933105468,
"p99": 0.39763144409179685,
"values": [
0.4105369567871094,
0.08688435363769531,
0.08789913940429687,
0.08591667175292969,
0.08624845123291015
]
},
"throughput": {
"unit": "samples/s",
"value": 66.00785783178912
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1376.518144,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.4974213104248047,
"mean": 0.24871065521240235,
"stdev": 0.16182630157470704,
"p50": 0.24871065521240235,
"p90": 0.378171696472168,
"p95": 0.39435432662963865,
"p99": 0.40730043075561523,
"values": [
0.4105369567871094,
0.08688435363769531
]
},
"throughput": {
"unit": "samples/s",
"value": 16.082945849601597
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1376.518144,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.2600642623901367,
"mean": 0.0866880874633789,
"stdev": 0.0008669888536518835,
"p50": 0.08624845123291015,
"p90": 0.08756900177001953,
"p95": 0.08773407058715821,
"p99": 0.08786612564086914,
"values": [
0.08789913940429687,
0.08591667175292969,
0.08624845123291015
]
},
"throughput": {
"unit": "samples/s",
"value": 69.21366217168743
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_token-classification_microsoft/deberta-v3-base",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"model": "microsoft/deberta-v3-base",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1140.764672,
"max_global_vram": 4597.481472,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 5,
"total": 1.0522736740112304,
"mean": 0.21045473480224608,
"stdev": 0.2521626361507063,
"p50": 0.08427519989013672,
"p90": 0.46303764648437507,
"p95": 0.5889081359863281,
"p99": 0.6896045275878907,
"values": [
0.7147786254882813,
0.08542617797851562,
0.08360550689697266,
0.08418816375732421,
0.08427519989013672
]
},
"throughput": {
"unit": "samples/s",
"value": 47.51615595342393
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1140.764672,
"max_global_vram": 4597.481472,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.8002048034667969,
"mean": 0.40010240173339845,
"stdev": 0.3146762237548828,
"p50": 0.40010240173339845,
"p90": 0.6518433807373047,
"p95": 0.683311003112793,
"p99": 0.7084851010131836,
"values": [
0.7147786254882813,
0.08542617797851562
]
},
"throughput": {
"unit": "samples/s",
"value": 9.997440611879489
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1140.764672,
"max_global_vram": 4597.481472,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.2520688705444336,
"mean": 0.08402295684814454,
"stdev": 0.0002973125946472148,
"p50": 0.08418816375732421,
"p90": 0.08425779266357422,
"p95": 0.08426649627685547,
"p99": 0.08427345916748047,
"values": [
0.08360550689697266,
0.08418816375732421,
0.08427519989013672
]
},
"throughput": {
"unit": "samples/s",
"value": 71.40905563278207
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |