Dataset Preview
Full Screen Viewer
Full Screen
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code: DatasetGenerationError Exception: ArrowNotImplementedError Message: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. Traceback: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2013, in _prepare_split_single writer.write_table(table) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 583, in write_table self._build_writer(inferred_schema=pa_table.schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 404, in _build_writer self.pa_writer = self._WRITER_CLASS(self.stream, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__ self.writer = _parquet.ParquetWriter( File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__ File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2029, in _prepare_split_single num_examples, num_bytes = writer.finalize() File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 602, in finalize self._build_writer(self.schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 404, in _build_writer self.pa_writer = self._WRITER_CLASS(self.stream, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__ self.writer = _parquet.ParquetWriter( File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__ File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1396, in compute_config_parquet_and_info_response parquet_operations = convert_to_parquet(builder) File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1045, in convert_to_parquet builder.download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1029, in download_and_prepare self._download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1124, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1884, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2040, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
config
dict | report
dict | name
string | backend
dict | scenario
dict | launcher
dict | environment
dict | overall
dict | warmup
dict | train
dict |
---|---|---|---|---|---|---|---|---|---|
{
"name": "cpu_training_transformers_fill-mask_google-bert/bert-base-uncased",
"backend": {
"name": "pytorch",
"version": "2.4.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"library": "transformers",
"model_type": "bert",
"model": "google-bert/bert-base-uncased",
"processor": "google-bert/bert-base-uncased",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16766.7712,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.8.0-1014-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": "81ee8771280a17916e9c5b93baa1d4a38cffa99d",
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2496.827392,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.3493527759999893,
"mean": 0.6698705551999978,
"stdev": 0.06414548103097348,
"p50": 0.6484932549999485,
"p90": 0.7377179964000107,
"p95": 0.7672214882000162,
"p99": 0.7908242816400207,
"values": [
0.7967249800000218,
0.6484932549999485,
0.6492075209999939,
0.6279653470000426,
0.6269616729999825
]
},
"throughput": {
"unit": "samples/s",
"value": 14.928257291461902
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2496.827392,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.4452182349999703,
"mean": 0.7226091174999851,
"stdev": 0.07411586250003666,
"p50": 0.7226091174999851,
"p90": 0.7819018075000145,
"p95": 0.7893133937500181,
"p99": 0.7952426627500211,
"values": [
0.7967249800000218,
0.6484932549999485
]
},
"throughput": {
"unit": "samples/s",
"value": 5.535496166777999
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2496.827392,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.904134541000019,
"mean": 0.634711513666673,
"stdev": 0.010258411567436393,
"p50": 0.6279653470000426,
"p90": 0.6449590862000036,
"p95": 0.6470833035999988,
"p99": 0.6487826775199949,
"values": [
0.6492075209999939,
0.6279653470000426,
0.6269616729999825
]
},
"throughput": {
"unit": "samples/s",
"value": 9.453113533955802
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_fill-mask_google-bert/bert-base-uncased | {
"name": "pytorch",
"version": "2.4.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"library": "transformers",
"model_type": "bert",
"model": "google-bert/bert-base-uncased",
"processor": "google-bert/bert-base-uncased",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16766.7712,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.8.0-1014-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": "81ee8771280a17916e9c5b93baa1d4a38cffa99d",
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2496.827392,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.3493527759999893,
"mean": 0.6698705551999978,
"stdev": 0.06414548103097348,
"p50": 0.6484932549999485,
"p90": 0.7377179964000107,
"p95": 0.7672214882000162,
"p99": 0.7908242816400207,
"values": [
0.7967249800000218,
0.6484932549999485,
0.6492075209999939,
0.6279653470000426,
0.6269616729999825
]
},
"throughput": {
"unit": "samples/s",
"value": 14.928257291461902
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2496.827392,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.4452182349999703,
"mean": 0.7226091174999851,
"stdev": 0.07411586250003666,
"p50": 0.7226091174999851,
"p90": 0.7819018075000145,
"p95": 0.7893133937500181,
"p99": 0.7952426627500211,
"values": [
0.7967249800000218,
0.6484932549999485
]
},
"throughput": {
"unit": "samples/s",
"value": 5.535496166777999
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2496.827392,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.904134541000019,
"mean": 0.634711513666673,
"stdev": 0.010258411567436393,
"p50": 0.6279653470000426,
"p90": 0.6449590862000036,
"p95": 0.6470833035999988,
"p99": 0.6487826775199949,
"values": [
0.6492075209999939,
0.6279653470000426,
0.6269616729999825
]
},
"throughput": {
"unit": "samples/s",
"value": 9.453113533955802
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_fill-mask_google-bert/bert-base-uncased",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"model": "google-bert/bert-base-uncased",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2488.782848,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 2.738775005999969,
"mean": 0.5477550011999938,
"stdev": 0.03693447784258994,
"p50": 0.5307143729999666,
"p90": 0.5856752317999963,
"p95": 0.6036043034000045,
"p99": 0.6179475606800111,
"values": [
0.6215333750000127,
0.5307143729999666,
0.5270229210000252,
0.5276163199999928,
0.5318880169999716
]
},
"throughput": {
"unit": "samples/s",
"value": 18.25633719106628
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2488.782848,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.1522477479999793,
"mean": 0.5761238739999897,
"stdev": 0.045409501000023056,
"p50": 0.5761238739999897,
"p90": 0.612451474800008,
"p95": 0.6169924249000104,
"p99": 0.6206251849800123,
"values": [
0.6215333750000127,
0.5307143729999666
]
},
"throughput": {
"unit": "samples/s",
"value": 6.942951300088064
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2488.782848,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.5865272579999896,
"mean": 0.5288424193333299,
"stdev": 0.0021671455040491463,
"p50": 0.5276163199999928,
"p90": 0.5310336775999758,
"p95": 0.5314608472999737,
"p99": 0.531802583059972,
"values": [
0.5270229210000252,
0.5276163199999928,
0.5318880169999716
]
},
"throughput": {
"unit": "samples/s",
"value": 11.345534663357242
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_image-classification_google/vit-base-patch16-224",
"backend": {
"name": "pytorch",
"version": "2.4.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"library": "transformers",
"model_type": "vit",
"model": "google/vit-base-patch16-224",
"processor": "google/vit-base-patch16-224",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16766.7712,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.8.0-1014-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": "81ee8771280a17916e9c5b93baa1d4a38cffa99d",
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2469.924864,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 7.68747861199995,
"mean": 1.5374957223999899,
"stdev": 0.049981819692801334,
"p50": 1.513895032999983,
"p90": 1.5939417259999913,
"p95": 1.6145628339999802,
"p99": 1.6310597203999715,
"values": [
1.6351839419999692,
1.513895032999983,
1.5017850280000289,
1.5320784020000247,
1.5045362069999442
]
},
"throughput": {
"unit": "samples/s",
"value": 6.504083136173066
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2469.924864,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 3.149078974999952,
"mean": 1.574539487499976,
"stdev": 0.06064445449999312,
"p50": 1.574539487499976,
"p90": 1.6230550510999706,
"p95": 1.6291194965499698,
"p99": 1.6339710529099694,
"values": [
1.6351839419999692,
1.513895032999983
]
},
"throughput": {
"unit": "samples/s",
"value": 2.540425331822655
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2469.924864,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 4.538399636999998,
"mean": 1.5127998789999992,
"stdev": 0.013678165892381806,
"p50": 1.5045362069999442,
"p90": 1.5265699630000085,
"p95": 1.5293241825000166,
"p99": 1.531527558100023,
"values": [
1.5017850280000289,
1.5320784020000247,
1.5045362069999442
]
},
"throughput": {
"unit": "samples/s",
"value": 3.966155790524097
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_image-classification_google/vit-base-patch16-224 | {
"name": "pytorch",
"version": "2.4.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"library": "transformers",
"model_type": "vit",
"model": "google/vit-base-patch16-224",
"processor": "google/vit-base-patch16-224",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16766.7712,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.8.0-1014-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": "81ee8771280a17916e9c5b93baa1d4a38cffa99d",
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2469.924864,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 7.68747861199995,
"mean": 1.5374957223999899,
"stdev": 0.049981819692801334,
"p50": 1.513895032999983,
"p90": 1.5939417259999913,
"p95": 1.6145628339999802,
"p99": 1.6310597203999715,
"values": [
1.6351839419999692,
1.513895032999983,
1.5017850280000289,
1.5320784020000247,
1.5045362069999442
]
},
"throughput": {
"unit": "samples/s",
"value": 6.504083136173066
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2469.924864,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 3.149078974999952,
"mean": 1.574539487499976,
"stdev": 0.06064445449999312,
"p50": 1.574539487499976,
"p90": 1.6230550510999706,
"p95": 1.6291194965499698,
"p99": 1.6339710529099694,
"values": [
1.6351839419999692,
1.513895032999983
]
},
"throughput": {
"unit": "samples/s",
"value": 2.540425331822655
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2469.924864,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 4.538399636999998,
"mean": 1.5127998789999992,
"stdev": 0.013678165892381806,
"p50": 1.5045362069999442,
"p90": 1.5265699630000085,
"p95": 1.5293241825000166,
"p99": 1.531527558100023,
"values": [
1.5017850280000289,
1.5320784020000247,
1.5045362069999442
]
},
"throughput": {
"unit": "samples/s",
"value": 3.966155790524097
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_image-classification_google/vit-base-patch16-224",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"model": "google/vit-base-patch16-224",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2442.985472,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 7.2970974209999895,
"mean": 1.459419484199998,
"stdev": 0.05210139006345095,
"p50": 1.4401334369999859,
"p90": 1.521250764199999,
"p95": 1.5379663595999886,
"p99": 1.5513388359199802,
"values": [
1.5546819549999782,
1.4241662819999874,
1.4401334369999859,
1.4711039780000306,
1.4070117690000075
]
},
"throughput": {
"unit": "samples/s",
"value": 6.852039532336137
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2442.985472,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 2.9788482369999656,
"mean": 1.4894241184999828,
"stdev": 0.06525783649999539,
"p50": 1.4894241184999828,
"p90": 1.541630387699979,
"p95": 1.5481561713499787,
"p99": 1.5533767982699782,
"values": [
1.5546819549999782,
1.4241662819999874
]
},
"throughput": {
"unit": "samples/s",
"value": 2.6856017371522416
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2442.985472,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 4.318249184000024,
"mean": 1.4394163946666747,
"stdev": 0.02617044676610726,
"p50": 1.4401334369999859,
"p90": 1.4649098698000216,
"p95": 1.4680069239000262,
"p99": 1.4704845671800297,
"values": [
1.4401334369999859,
1.4711039780000306,
1.4070117690000075
]
},
"throughput": {
"unit": "samples/s",
"value": 4.1683560241714614
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_multiple-choice_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.4.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16766.7712,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.8.0-1014-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": "81ee8771280a17916e9c5b93baa1d4a38cffa99d",
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2867.228672,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.6817775919999463,
"mean": 0.7363555183999893,
"stdev": 0.04850237003827847,
"p50": 0.7179886629999714,
"p90": 0.7889729999999986,
"p95": 0.8104847719999952,
"p99": 0.8276941895999925,
"values": [
0.8319965439999919,
0.7244376840000086,
0.7038921179999988,
0.7034625829999754,
0.7179886629999714
]
},
"throughput": {
"unit": "samples/s",
"value": 13.580396629237981
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2867.228672,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.5564342280000005,
"mean": 0.7782171140000003,
"stdev": 0.053779429999991635,
"p50": 0.7782171140000003,
"p90": 0.8212406579999936,
"p95": 0.8266186009999927,
"p99": 0.8309209553999921,
"values": [
0.8319965439999919,
0.7244376840000086
]
},
"throughput": {
"unit": "samples/s",
"value": 5.139953784156948
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2867.228672,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 2.1253433639999457,
"mean": 0.7084477879999819,
"stdev": 0.006748696017411076,
"p50": 0.7038921179999988,
"p90": 0.7151693539999769,
"p95": 0.7165790084999741,
"p99": 0.7177067320999719,
"values": [
0.7038921179999988,
0.7034625829999754,
0.7179886629999714
]
},
"throughput": {
"unit": "samples/s",
"value": 8.469219752860818
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_multiple-choice_FacebookAI/roberta-base | {
"name": "pytorch",
"version": "2.4.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16766.7712,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.8.0-1014-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": "81ee8771280a17916e9c5b93baa1d4a38cffa99d",
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2867.228672,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.6817775919999463,
"mean": 0.7363555183999893,
"stdev": 0.04850237003827847,
"p50": 0.7179886629999714,
"p90": 0.7889729999999986,
"p95": 0.8104847719999952,
"p99": 0.8276941895999925,
"values": [
0.8319965439999919,
0.7244376840000086,
0.7038921179999988,
0.7034625829999754,
0.7179886629999714
]
},
"throughput": {
"unit": "samples/s",
"value": 13.580396629237981
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2867.228672,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.5564342280000005,
"mean": 0.7782171140000003,
"stdev": 0.053779429999991635,
"p50": 0.7782171140000003,
"p90": 0.8212406579999936,
"p95": 0.8266186009999927,
"p99": 0.8309209553999921,
"values": [
0.8319965439999919,
0.7244376840000086
]
},
"throughput": {
"unit": "samples/s",
"value": 5.139953784156948
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2867.228672,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 2.1253433639999457,
"mean": 0.7084477879999819,
"stdev": 0.006748696017411076,
"p50": 0.7038921179999988,
"p90": 0.7151693539999769,
"p95": 0.7165790084999741,
"p99": 0.7177067320999719,
"values": [
0.7038921179999988,
0.7034625829999754,
0.7179886629999714
]
},
"throughput": {
"unit": "samples/s",
"value": 8.469219752860818
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_multiple-choice_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"model": "FacebookAI/roberta-base",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2845.749248,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.581090587999995,
"mean": 0.716218117599999,
"stdev": 0.043372798377969854,
"p50": 0.697155070000008,
"p90": 0.7645524997999928,
"p95": 0.7826806403999967,
"p99": 0.7971831528799999,
"values": [
0.8008087810000006,
0.710168077999981,
0.6928677590000234,
0.697155070000008,
0.680090899999982
]
},
"throughput": {
"unit": "samples/s",
"value": 13.962227084549829
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2845.749248,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.5109768589999817,
"mean": 0.7554884294999908,
"stdev": 0.04532035150000979,
"p50": 0.7554884294999908,
"p90": 0.7917447106999986,
"p95": 0.7962767458499996,
"p99": 0.7999023739700004,
"values": [
0.8008087810000006,
0.710168077999981
]
},
"throughput": {
"unit": "samples/s",
"value": 5.294588035778844
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2845.749248,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 2.0701137290000133,
"mean": 0.6900379096666711,
"stdev": 0.007248103654729414,
"p50": 0.6928677590000234,
"p90": 0.696297607800011,
"p95": 0.6967263389000096,
"p99": 0.6970693237800083,
"values": [
0.6928677590000234,
0.697155070000008,
0.680090899999982
]
},
"throughput": {
"unit": "samples/s",
"value": 8.695174447587021
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_text-classification_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.4.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16766.7712,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.8.0-1014-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": "81ee8771280a17916e9c5b93baa1d4a38cffa99d",
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2853.994496,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.0089543170000184,
"mean": 0.6017908634000037,
"stdev": 0.04611702098835511,
"p50": 0.5786917049999829,
"p90": 0.6553951017999907,
"p95": 0.6729830384000024,
"p99": 0.6870533876800118,
"values": [
0.6905709750000142,
0.5786917049999829,
0.5669231020000325,
0.5701372430000333,
0.6026312919999555
]
},
"throughput": {
"unit": "samples/s",
"value": 16.617068500345628
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2853.994496,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.269262679999997,
"mean": 0.6346313399999985,
"stdev": 0.05593963500001564,
"p50": 0.6346313399999985,
"p90": 0.679383048000011,
"p95": 0.6849770115000127,
"p99": 0.6894521823000139,
"values": [
0.6905709750000142,
0.5786917049999829
]
},
"throughput": {
"unit": "samples/s",
"value": 6.3028718373725585
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2853.994496,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.7396916370000213,
"mean": 0.5798972123333405,
"stdev": 0.01612888628740313,
"p50": 0.5701372430000333,
"p90": 0.5961324821999711,
"p95": 0.5993818870999632,
"p99": 0.6019814110199571,
"values": [
0.5669231020000325,
0.5701372430000333,
0.6026312919999555
]
},
"throughput": {
"unit": "samples/s",
"value": 10.346661222697927
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_text-classification_FacebookAI/roberta-base | {
"name": "pytorch",
"version": "2.4.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16766.7712,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.8.0-1014-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": "81ee8771280a17916e9c5b93baa1d4a38cffa99d",
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2853.994496,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.0089543170000184,
"mean": 0.6017908634000037,
"stdev": 0.04611702098835511,
"p50": 0.5786917049999829,
"p90": 0.6553951017999907,
"p95": 0.6729830384000024,
"p99": 0.6870533876800118,
"values": [
0.6905709750000142,
0.5786917049999829,
0.5669231020000325,
0.5701372430000333,
0.6026312919999555
]
},
"throughput": {
"unit": "samples/s",
"value": 16.617068500345628
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2853.994496,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.269262679999997,
"mean": 0.6346313399999985,
"stdev": 0.05593963500001564,
"p50": 0.6346313399999985,
"p90": 0.679383048000011,
"p95": 0.6849770115000127,
"p99": 0.6894521823000139,
"values": [
0.6905709750000142,
0.5786917049999829
]
},
"throughput": {
"unit": "samples/s",
"value": 6.3028718373725585
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2853.994496,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.7396916370000213,
"mean": 0.5798972123333405,
"stdev": 0.01612888628740313,
"p50": 0.5701372430000333,
"p90": 0.5961324821999711,
"p95": 0.5993818870999632,
"p99": 0.6019814110199571,
"values": [
0.5669231020000325,
0.5701372430000333,
0.6026312919999555
]
},
"throughput": {
"unit": "samples/s",
"value": 10.346661222697927
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_text-classification_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"model": "FacebookAI/roberta-base",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2826.752,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 2.882509665999976,
"mean": 0.5765019331999952,
"stdev": 0.04978939696949581,
"p50": 0.5569985249999831,
"p90": 0.6300386333999881,
"p95": 0.6525100941999881,
"p99": 0.670487262839988,
"values": [
0.674981554999988,
0.5569985249999831,
0.5424031590000027,
0.5626242509999884,
0.5455021760000136
]
},
"throughput": {
"unit": "samples/s",
"value": 17.34599560576128
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2826.752,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.2319800799999712,
"mean": 0.6159900399999856,
"stdev": 0.05899151500000244,
"p50": 0.6159900399999856,
"p90": 0.6631832519999875,
"p95": 0.6690824034999878,
"p99": 0.673801724699988,
"values": [
0.674981554999988,
0.5569985249999831
]
},
"throughput": {
"unit": "samples/s",
"value": 6.493611487614465
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2826.752,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.6505295860000047,
"mean": 0.5501765286666682,
"stdev": 0.00889233078021606,
"p50": 0.5455021760000136,
"p90": 0.5591998359999935,
"p95": 0.5609120434999909,
"p99": 0.5622818094999888,
"values": [
0.5424031590000027,
0.5626242509999884,
0.5455021760000136
]
},
"throughput": {
"unit": "samples/s",
"value": 10.905590637500968
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_text-generation_openai-community/gpt2",
"backend": {
"name": "pytorch",
"version": "2.4.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "gpt2",
"model": "openai-community/gpt2",
"processor": "openai-community/gpt2",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16766.7712,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.8.0-1014-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": "81ee8771280a17916e9c5b93baa1d4a38cffa99d",
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2844.20096,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.989536197999996,
"mean": 0.7979072395999992,
"stdev": 0.12206507824295237,
"p50": 0.7583344489999888,
"p90": 0.9438610822000101,
"p95": 0.9488353136000114,
"p99": 0.9528146987200126,
"values": [
0.7583344489999888,
0.9289383880000059,
0.9538095450000128,
0.6950243999999941,
0.6534294159999945
]
},
"throughput": {
"unit": "samples/s",
"value": 12.532785145567953
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2844.20096,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.6872728369999948,
"mean": 0.8436364184999974,
"stdev": 0.08530196950000857,
"p50": 0.8436364184999974,
"p90": 0.9118779941000043,
"p95": 0.920408191050005,
"p99": 0.9272323486100058,
"values": [
0.7583344489999888,
0.9289383880000059
]
},
"throughput": {
"unit": "samples/s",
"value": 4.7413790020019295
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2844.20096,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 2.3022633610000014,
"mean": 0.7674211203333338,
"stdev": 0.13288596442259495,
"p50": 0.6950243999999941,
"p90": 0.902052516000009,
"p95": 0.9279310305000109,
"p99": 0.9486338421000124,
"values": [
0.9538095450000128,
0.6950243999999941,
0.6534294159999945
]
},
"throughput": {
"unit": "samples/s",
"value": 7.81839311041357
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_text-generation_openai-community/gpt2 | {
"name": "pytorch",
"version": "2.4.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "gpt2",
"model": "openai-community/gpt2",
"processor": "openai-community/gpt2",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16766.7712,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.8.0-1014-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": "81ee8771280a17916e9c5b93baa1d4a38cffa99d",
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2844.20096,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.989536197999996,
"mean": 0.7979072395999992,
"stdev": 0.12206507824295237,
"p50": 0.7583344489999888,
"p90": 0.9438610822000101,
"p95": 0.9488353136000114,
"p99": 0.9528146987200126,
"values": [
0.7583344489999888,
0.9289383880000059,
0.9538095450000128,
0.6950243999999941,
0.6534294159999945
]
},
"throughput": {
"unit": "samples/s",
"value": 12.532785145567953
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2844.20096,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.6872728369999948,
"mean": 0.8436364184999974,
"stdev": 0.08530196950000857,
"p50": 0.8436364184999974,
"p90": 0.9118779941000043,
"p95": 0.920408191050005,
"p99": 0.9272323486100058,
"values": [
0.7583344489999888,
0.9289383880000059
]
},
"throughput": {
"unit": "samples/s",
"value": 4.7413790020019295
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2844.20096,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 2.3022633610000014,
"mean": 0.7674211203333338,
"stdev": 0.13288596442259495,
"p50": 0.6950243999999941,
"p90": 0.902052516000009,
"p95": 0.9279310305000109,
"p99": 0.9486338421000124,
"values": [
0.9538095450000128,
0.6950243999999941,
0.6534294159999945
]
},
"throughput": {
"unit": "samples/s",
"value": 7.81839311041357
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_text-generation_openai-community/gpt2",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"model": "openai-community/gpt2",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2827.354112,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.1791685380000274,
"mean": 0.6358337076000055,
"stdev": 0.07846941233493662,
"p50": 0.596941285000014,
"p90": 0.7161873142000047,
"p95": 0.7544328206000045,
"p99": 0.7850292257200044,
"values": [
0.7926783270000044,
0.6014507950000052,
0.596941285000014,
0.593667699000008,
0.5944304319999958
]
},
"throughput": {
"unit": "samples/s",
"value": 15.727382616668173
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2827.354112,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.3941291220000096,
"mean": 0.6970645610000048,
"stdev": 0.0956137659999996,
"p50": 0.6970645610000048,
"p90": 0.7735555738000045,
"p95": 0.7831169504000044,
"p99": 0.7907660516800044,
"values": [
0.7926783270000044,
0.6014507950000052
]
},
"throughput": {
"unit": "samples/s",
"value": 5.738349392288174
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2827.354112,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.7850394160000178,
"mean": 0.5950131386666726,
"stdev": 0.0013985114990352936,
"p50": 0.5944304319999958,
"p90": 0.5964391144000103,
"p95": 0.5966901997000121,
"p99": 0.5968910679400136,
"values": [
0.596941285000014,
0.593667699000008,
0.5944304319999958
]
},
"throughput": {
"unit": "samples/s",
"value": 10.083810944822195
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_token-classification_microsoft/deberta-v3-base",
"backend": {
"name": "pytorch",
"version": "2.4.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"library": "transformers",
"model_type": "deberta-v2",
"model": "microsoft/deberta-v3-base",
"processor": "microsoft/deberta-v3-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16766.7712,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.8.0-1014-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": "81ee8771280a17916e9c5b93baa1d4a38cffa99d",
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 4413.366272,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 5.9785980609999,
"mean": 1.19571961219998,
"stdev": 0.07268821849403599,
"p50": 1.164434306999965,
"p90": 1.2728477809999845,
"p95": 1.306518023999979,
"p99": 1.3334542183999747,
"values": [
1.3401882669999736,
1.164434306999965,
1.1488074899999674,
1.1718370520000008,
1.1533309449999933
]
},
"throughput": {
"unit": "samples/s",
"value": 8.363164656638194
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 4413.366272,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 2.5046225739999386,
"mean": 1.2523112869999693,
"stdev": 0.08787698000000432,
"p50": 1.2523112869999693,
"p90": 1.3226128709999727,
"p95": 1.3314005689999733,
"p99": 1.3384307273999736,
"values": [
1.3401882669999736,
1.164434306999965
]
},
"throughput": {
"unit": "samples/s",
"value": 3.194094025601558
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 4413.366272,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 3.4739754869999615,
"mean": 1.157991828999987,
"stdev": 0.009962699139771112,
"p50": 1.1533309449999933,
"p90": 1.1681358305999994,
"p95": 1.1699864413,
"p99": 1.1714669298600007,
"values": [
1.1488074899999674,
1.1718370520000008,
1.1533309449999933
]
},
"throughput": {
"unit": "samples/s",
"value": 5.1813837107827005
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_token-classification_microsoft/deberta-v3-base | {
"name": "pytorch",
"version": "2.4.1+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"library": "transformers",
"model_type": "deberta-v2",
"model": "microsoft/deberta-v3-base",
"processor": "microsoft/deberta-v3-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": null,
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16766.7712,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.8.0-1014-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.15",
"optimum_benchmark_version": "0.4.0",
"optimum_benchmark_commit": "81ee8771280a17916e9c5b93baa1d4a38cffa99d",
"transformers_version": "4.44.2",
"transformers_commit": null,
"accelerate_version": "0.34.2",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 4413.366272,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 5.9785980609999,
"mean": 1.19571961219998,
"stdev": 0.07268821849403599,
"p50": 1.164434306999965,
"p90": 1.2728477809999845,
"p95": 1.306518023999979,
"p99": 1.3334542183999747,
"values": [
1.3401882669999736,
1.164434306999965,
1.1488074899999674,
1.1718370520000008,
1.1533309449999933
]
},
"throughput": {
"unit": "samples/s",
"value": 8.363164656638194
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 4413.366272,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 2.5046225739999386,
"mean": 1.2523112869999693,
"stdev": 0.08787698000000432,
"p50": 1.2523112869999693,
"p90": 1.3226128709999727,
"p95": 1.3314005689999733,
"p99": 1.3384307273999736,
"values": [
1.3401882669999736,
1.164434306999965
]
},
"throughput": {
"unit": "samples/s",
"value": 3.194094025601558
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 4413.366272,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 3.4739754869999615,
"mean": 1.157991828999987,
"stdev": 0.009962699139771112,
"p50": 1.1533309449999933,
"p90": 1.1681358305999994,
"p95": 1.1699864413,
"p99": 1.1714669298600007,
"values": [
1.1488074899999674,
1.1718370520000008,
1.1533309449999933
]
},
"throughput": {
"unit": "samples/s",
"value": 5.1813837107827005
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_token-classification_microsoft/deberta-v3-base",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"model": "microsoft/deberta-v3-base",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 4374.970368,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 5.583659903000068,
"mean": 1.1167319806000138,
"stdev": 0.07853258776030796,
"p50": 1.0731834320000075,
"p90": 1.2025418994000006,
"p95": 1.2374836681999908,
"p99": 1.265437083239983,
"values": [
1.2724254369999812,
1.0731834320000075,
1.0977165930000297,
1.0711078369999996,
1.0692266040000504
]
},
"throughput": {
"unit": "samples/s",
"value": 8.954700119385008
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 4374.970368,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 2.3456088689999888,
"mean": 1.1728044344999944,
"stdev": 0.09962100249998684,
"p50": 1.1728044344999944,
"p90": 1.252501236499984,
"p95": 1.2624633367499825,
"p99": 1.2704330169499816,
"values": [
1.2724254369999812,
1.0731834320000075
]
},
"throughput": {
"unit": "samples/s",
"value": 3.4106283045436583
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 4374.970368,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 3.2380510340000797,
"mean": 1.0793503446666932,
"stdev": 0.01300958794585394,
"p50": 1.0711078369999996,
"p90": 1.0923948418000236,
"p95": 1.0950557174000266,
"p99": 1.097184417880029,
"values": [
1.0977165930000297,
1.0711078369999996,
1.0692266040000504
]
},
"throughput": {
"unit": "samples/s",
"value": 5.558899415419021
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
No dataset card yet
New: Create and edit this dataset card directly on the website!
Contribute a Dataset Card- Downloads last month
- 2