mirror of
https://github.com/saymrwulf/pytorch.git
synced 2026-05-14 20:57:59 +00:00
Adding support to pass rank and world_size to torchbench model, via its extra_args parameter: https://github.com/pytorch/benchmark/blob/main/torchbenchmark/util/model.py#L83C80-L83C90 This is used for models which distribute over multiple GPUs e.g. simple_gpt https://github.com/pytorch/benchmark/pull/1867 Also add an option to skip multiprocess only gpu models Testing via `python benchmarks/dynamo/torchbench.py -d cuda --output=benchmark_logs/performance.csv --inference --performance --timing --print-memory --multiprocess --only simple_gpt` Pull Request resolved: https://github.com/pytorch/pytorch/pull/108438 Approved by: https://github.com/Chillee |
||
|---|---|---|
| .. | ||
| audio.txt | ||
| data.txt | ||
| fbgemm.txt | ||
| multipy.txt | ||
| numpy_pytorch_interop.txt | ||
| text.txt | ||
| torchbench.txt | ||
| torchrec.txt | ||
| triton.txt | ||
| vision.txt | ||
| xla.txt | ||