mirror of
https://github.com/saymrwulf/pytorch.git
synced 2026-05-14 20:57:59 +00:00
This is one of a series of PRs to update us to PEP585 (changing Dict -> dict, List -> list, etc). Most of the PRs were completely automated with RUFF as follows:
Since RUFF UP006 is considered an "unsafe" fix first we need to enable unsafe fixes:
```
--- a/tools/linter/adapters/ruff_linter.py
+++ b/tools/linter/adapters/ruff_linter.py
@@ -313,6 +313,7 @@
"ruff",
"check",
"--fix-only",
+ "--unsafe-fixes",
"--exit-zero",
*([f"--config={config}"] if config else []),
"--stdin-filename",
```
Then we need to tell RUFF to allow UP006 (as a final PR once all of these have landed this will be made permanent):
```
--- a/pyproject.toml
+++ b/pyproject.toml
@@ -40,7 +40,7 @@
[tool.ruff]
-target-version = "py38"
+target-version = "py39"
line-length = 88
src = ["caffe2", "torch", "torchgen", "functorch", "test"]
@@ -87,7 +87,6 @@
"SIM116", # Disable Use a dictionary instead of consecutive `if` statements
"SIM117",
"SIM118",
- "UP006", # keep-runtime-typing
"UP007", # keep-runtime-typing
]
select = [
```
Finally running `lintrunner -a --take RUFF` will fix up the deprecated uses.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/145101
Approved by: https://github.com/bobrenjc93
|
||
|---|---|---|
| .. | ||
| audio_text_models.py | ||
| compare.py | ||
| functional_autograd_benchmark.py | ||
| ppl_models.py | ||
| README.md | ||
| torchaudio_models.py | ||
| torchvision_models.py | ||
| utils.py | ||
| vision_models.py | ||
Benchmarking tool for the autograd API
This folder contain a set of self-contained scripts that allows you to benchmark autograd with different common models. It is designed to run the benchmark before and after your change and will generate a table to share on the PR.
To do so, you can use functional_autograd_benchmark.py to run the benchmarks before your change (using as output before.txt) and after your change (using as output after.txt).
You can then use compare.py to get a markdown table comparing the two runs.
The default arguments of functional_autograd_benchmark.py should be used in general. You can change them though to force a given device or force running even the (very) slow settings.
Sample usage
# Make sure you compile pytorch in release mode and with the same flags before/after
export DEBUG=0
# When running on CPU, it might be required to limit the number of cores to avoid oversubscription
export OMP_NUM_THREADS=10
# Compile pytorch with the base revision
git checkout master
python setup.py develop
# Install dependencies:
# Scipy is required by detr
pip install scipy
# Run the benchmark for the base
# This will use the GPU if available.
pushd benchmarks/functional_autograd_benchmark
python functional_autograd_benchmark.py --output before.txt
# Compile pytorch with your change
popd
git checkout your_feature_branch
python setup.py develop
# Run the benchmark for the new version
pushd benchmarks/functional_autograd_benchmark
python functional_autograd_benchmark.py --output after.txt
# Get the markdown table that you can paste in your github PR
python compare.py
popd
Files in this folder:
functional_autograd_benchmark.pyis the main entry point to run the benchmark.compare.pyis the entry point to run the comparison script that generates a markdown table.torchaudio_models.pyandtorchvision_models.pycontains code extracted from torchaudio and torchvision to be able to run the models without having a specific version of these libraries installed.ppl_models.py,vision_models.pyandaudio_text_models.pycontain all the getter functions used for the benchmark.
Benchmarking against functorch
# Install stable functorch:
pip install functorch
# or install from source:
pip install git+https://github.com/pytorch/functorch
# Run the benchmark for the base
# This will use the GPU if available.
pushd benchmarks/functional_autograd_benchmark
python functional_autograd_benchmark.py --output bench-with-functorch.txt