backendbench_tests / README.md
sahancpal's picture
Update README.md
4a3e01e verified
---
configs:
- config_name: torchbench
data_files:
- split: benchmark
path: "backend_bench_problems.parquet"
- config_name: ops_traces_models
data_files:
- split: operator_input_models
path: "operator_input_models_mapping.parquet"
---
# TorchBench
The TorchBench suite of [BackendBench](https://github.com/meta-pytorch/BackendBench) is designed to mimic real-world use cases. It provides operators and inputs derived from 155 model traces found in [TIMM](https://huggingface.co/timm) (67), [Hugging Face Transformers](https://huggingface.co/docs/transformers/en/index) (45), and [TorchBench](https://github.com/pytorch/benchmark) (43). (These are also the models PyTorch developers use to [validate performance](https://hud.pytorch.org/benchmark/compilers).) You can view the origin of these traces by switching the subset in the dataset viewer to `ops_traces_models` and `torchbench` for the full dataset.
When running BackendBench, much of the extra information about what you are testing is abstracted away, so you can simply run `uv run python --suite torchbench ...`. Here, however, we provide the test suite as a dataset that can be explored directly. It includes details about why certain operations and arguments were included or excluded, reflecting the careful consideration behind curating the set.
You can download the dataset in either format:
- `backend_bench_problems.parquet` (default format on Hugging Face)
- `backend_bench_problems.json` (more human-readable)
### Fields
- **uuid** – Unique identifier for the `(op_name, args)` pair.
- **op_name** – Full name of the operator being tested.
- **args** – Serialized form of the inputs from the trace. [See details below](#serialized-arguments-in-backendbench).
- **runnable** – Whether the operator is runnable in BackendBench (some are not yet supported).
- **included_in_benchmark** – Whether this `(op_name, args)` pair is tested in the TorchBench suite.
- **why_excluded** – If not included, a list of reasons for exclusion (e.g., "BackendBench does not support correctness testing for random ops yet," "BackendBench does not support correctness testing for tensor creation and manipulation ops yet").
- **is_synthetic** – Marks synthetically generated inputs (e.g., very large tensors). These are currently excluded from the benchmark.
- **runtime_ms** – Execution time (ms) on our hardware (single GPU from a machine with 8× H100s and an AMD EPYC 9654 96-core processor).
- **relative_runtime_to_kernel_launch** – `runtime_ms` divided by the runtime of a dummy CUDA op (`torch.empty(0, device=cuda)`), representing launch overhead.
- **is_overhead_dominated_op** – Flags operator/argument pairs running close to CUDA overhead as “performance canaries.” [Histogram analysis](https://github.com/meta-pytorch/BackendBench/issues/108) showed that a 1.3× threshold above CUDA overhead is a useful cutoff. These tests can be run for sanity-checking kernels with `uv run python --suite torchbench --check-overhead-dominated-ops ...`.
- **count** – Number of times this operator/input pair appeared in model traces.
- **in_models** – List of models (from real-world traces) where this operator/input pair appears.
- **in_models_count** – Number of distinct models in which this operator/input pair occurs.
# Serialized Arguments in BackendBench
Generally, arguments are serialized by storing tensor shapes and preserving everything else as it's fairly intuitive. For example:
`((T([8, 8, 8, 8, 8], f16), T([8, 8, 8, 8, 8], f16)), {})`
Below we'll go into detail about the format for rigor.
## Format
BackendBench stores function arguments as strings with all parameters needed to reproduce PyTorch operations:
```python
((arg1, arg2, ...), {'key1': val1, 'key2': val2})
```
```python
(([T([5, 5], f32), T([3, 3], i64), 42],), {'weight': T([3, 3], f32)})
```
## Tensor Representation
Tensors use the format `T([shape], dtype)` or `T([shape], dtype, [stride])`:
```python
T([10, 20], f32) # 10×20 float32 tensor
T([1, 512, 768], f16) # 1×512×768 float16 tensor
T([64], i32) # 64-element int32 vector
```
**Data types**: `f16/f32/f64` (float), `bf16` (bfloat16), `i32/i64` (int), `b8` (bool)
## Examples
**Single tensor argument:**
```python
((T([48, 24, 28, 28], f16),), {})
```
48×24×28×28 float16 tensor, no keyword arguments
**Multiple tensors:**
```python
((T([8, 8, 8, 8, 8], f16), T([8, 8, 8, 8, 8], f16)), {})
```
Two 5D tensors of identical shapes
**Mixed arguments:**
```python
((T([128, 256], f16), [1024, 249, 249]), {'dtype': torch.float16, 'device': 'cuda'})
```
Args are a tensor, list, and keyword arguments
**Complex nested:**
```python
(([T([5, 5], f32), T([3, 3], i64), 42],), {'weight': T([3, 3], f32)})
```
List containing tensors and numbers, plus tensor keyword argument
## Argument Types
- **Tensors**: `T([shape], dtype)`
- **Lists**: `[item1, item2, ...]` (can contain tensors)
- **Primitives**: `42`, `'hello'`, `True`, `None`
- **PyTorch objects**: `torch.float16`, `torch.strided`
# Trace Files in BackendBench
This repository includes `.txt` trace files, which were the original output format of model traces and are used to compose the dataset. Here’s their structure:
## Format
Trace files capture PyTorch operations and arguments from real model executions:
```
Operator: operation_name
cnt: count, serialized_arguments
cnt: count, serialized_arguments
...
```
## Structure
**Operator line**: Specifies the PyTorch operation
```
Operator: aten.add.Tensor
Operator: aten.relu.default
Operator: aten.linear.default
```
**Count lines**: Show how often each argument combination was used
```
cnt: 42, ((T([10, 20], f16), T([10, 20], f16)), {})
cnt: 0, ((T([5, 5], f32), T([5, 5], f32)), {})
```
## Reading Count Lines
- **Count `42`**: Argument combination appeared 42 times in traced models
- **`cnt: 0`** = Synthetic/generated arguments (not from real models)
- **`cnt: >0`** = Real usage frequency from model traces
**Arguments**: Same format as serialized arguments – `((args), {kwargs})`
## Example
```
Operator: aten.add.Tensor
cnt: 156, ((T([1, 512, 768], f16), T([1, 512, 768], f16)), {})
cnt: 89, ((T([32, 128], f32), T([32, 128], f32)), {})
cnt: 0, ((T([10, 10], f16), T([10, 10], f16)), {})
Operator: aten.relu.default
cnt: 234, ((T([64, 256], f16),), {})
```
This shows:
- `aten.add.Tensor` called 156 times with 1×512×768 tensors
- Same operation called 89 times with 32×128 tensors
- One synthetic test case (`cnt: 0`)
- `aten.relu.default` called 234 times with a 64×256 tensor
**Note: Traces may be deprecated in the future, but are described here as they are currently included in the dataset/codebase.**
# Acknowledgements
We are extremely grateful to the [TritonBench](https://github.com/pytorch-labs/tritonbench/tree/main) team for these traces and their intuitive format.