mirror of
https://github.com/saymrwulf/pytorch.git
synced 2026-05-15 21:00:47 +00:00
This is one of a series of PRs to update us to PEP585 (changing Dict -> dict, List -> list, etc). Most of the PRs were completely automated with RUFF as follows:
Since RUFF UP006 is considered an "unsafe" fix first we need to enable unsafe fixes:
```
--- a/tools/linter/adapters/ruff_linter.py
+++ b/tools/linter/adapters/ruff_linter.py
@@ -313,6 +313,7 @@
"ruff",
"check",
"--fix-only",
+ "--unsafe-fixes",
"--exit-zero",
*([f"--config={config}"] if config else []),
"--stdin-filename",
```
Then we need to tell RUFF to allow UP006 (as a final PR once all of these have landed this will be made permanent):
```
--- a/pyproject.toml
+++ b/pyproject.toml
@@ -40,7 +40,7 @@
[tool.ruff]
-target-version = "py38"
+target-version = "py39"
line-length = 88
src = ["caffe2", "torch", "torchgen", "functorch", "test"]
@@ -87,7 +87,6 @@
"SIM116", # Disable Use a dictionary instead of consecutive `if` statements
"SIM117",
"SIM118",
- "UP006", # keep-runtime-typing
"UP007", # keep-runtime-typing
]
select = [
```
Finally running `lintrunner -a --take RUFF` will fix up the deprecated uses.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/145101
Approved by: https://github.com/bobrenjc93
94 lines
2.6 KiB
Python
94 lines
2.6 KiB
Python
import random
|
|
|
|
import operator_benchmark as op_bench
|
|
|
|
import torch
|
|
|
|
|
|
"""Microbenchmarks for Stack operator"""
|
|
|
|
# Configs for PT stack operator
|
|
stack_configs_static_runtime = op_bench.config_list(
|
|
attr_names=["sizes", "N"],
|
|
attrs=[
|
|
[(20, 40), 5],
|
|
[(1, 40), 5],
|
|
],
|
|
cross_product_configs={"device": ["cpu", "cuda"], "dim": list(range(3))},
|
|
tags=["static_runtime"],
|
|
)
|
|
|
|
stack_configs_short = op_bench.config_list(
|
|
attr_names=["sizes", "N"],
|
|
attrs=[
|
|
[(1, 1, 1), 2], # noqa: E241
|
|
[(512, 512, 2), 2], # noqa: E241
|
|
[(128, 1024, 2), 2], # noqa: E241
|
|
],
|
|
cross_product_configs={"device": ["cpu", "cuda"], "dim": list(range(4))},
|
|
tags=["short"],
|
|
)
|
|
|
|
stack_configs_long = op_bench.config_list(
|
|
attr_names=["sizes", "N"],
|
|
attrs=[
|
|
[(2**10, 2**10, 2), 2], # noqa: E241
|
|
[(2**10 + 1, 2**10 - 1, 2), 2], # noqa: E226,E241
|
|
[(2**10, 2**10, 2), 2], # noqa: E241
|
|
],
|
|
cross_product_configs={"device": ["cpu", "cuda"], "dim": list(range(4))},
|
|
tags=["long"],
|
|
)
|
|
|
|
# There is a different codepath on CUDA for >4 dimensions
|
|
stack_configs_multidim = op_bench.config_list(
|
|
attr_names=["sizes", "N"],
|
|
attrs=[
|
|
[(2**6, 2**5, 2**2, 2**4, 2**5), 2], # noqa: E241
|
|
[(2**4, 2**5, 2**2, 2**4, 2**5), 8], # noqa: E241
|
|
[
|
|
(2**3 + 1, 2**5 - 1, 2**2 + 1, 2**4 - 1, 2**5 + 1),
|
|
17,
|
|
], # noqa: E226,E241
|
|
],
|
|
cross_product_configs={"device": ["cpu", "cuda"], "dim": list(range(6))},
|
|
tags=["multidim"],
|
|
)
|
|
|
|
|
|
class StackBenchmark(op_bench.TorchBenchmarkBase):
|
|
def init(self, sizes, N, dim, device):
|
|
random.seed(42)
|
|
inputs = []
|
|
gen_sizes = []
|
|
if type(sizes) == list and N == -1:
|
|
gen_sizes = sizes
|
|
else:
|
|
for i in range(N):
|
|
gen_sizes.append(
|
|
[
|
|
old_size() if callable(old_size) else old_size
|
|
for old_size in sizes
|
|
]
|
|
)
|
|
|
|
for s in gen_sizes:
|
|
inputs.append(torch.rand(s, device=device))
|
|
result = torch.rand(gen_sizes[0], device=device)
|
|
self.inputs = {"result": result, "inputs": inputs, "dim": dim}
|
|
self.set_module_name("stack")
|
|
|
|
def forward(self, result: torch.Tensor, inputs: list[torch.Tensor], dim: int):
|
|
return torch.stack(inputs, dim=dim, out=result)
|
|
|
|
|
|
op_bench.generate_pt_test(
|
|
stack_configs_static_runtime
|
|
+ stack_configs_short
|
|
+ stack_configs_long
|
|
+ stack_configs_multidim,
|
|
StackBenchmark,
|
|
)
|
|
|
|
if __name__ == "__main__":
|
|
op_bench.benchmark_runner.main()
|