mirror of
https://github.com/saymrwulf/pytorch.git
synced 2026-05-14 20:57:59 +00:00
Adding support to pass rank and world_size to torchbench model, via its extra_args parameter: https://github.com/pytorch/benchmark/blob/main/torchbenchmark/util/model.py#L83C80-L83C90 This is used for models which distribute over multiple GPUs e.g. simple_gpt https://github.com/pytorch/benchmark/pull/1867 Also add an option to skip multiprocess only gpu models Testing via `python benchmarks/dynamo/torchbench.py -d cuda --output=benchmark_logs/performance.csv --inference --performance --timing --print-memory --multiprocess --only simple_gpt` Pull Request resolved: https://github.com/pytorch/pytorch/pull/108438 Approved by: https://github.com/Chillee |
||
|---|---|---|
| .. | ||
| actions | ||
| ci_commit_pins | ||
| ISSUE_TEMPLATE | ||
| requirements | ||
| scripts | ||
| templates | ||
| workflows | ||
| actionlint.yaml | ||
| auto_request_review.yml | ||
| ISSUE_TEMPLATE.md | ||
| labeler.yml | ||
| merge_rules.yaml | ||
| PULL_REQUEST_TEMPLATE.md | ||
| pytorch-circleci-labels.yml | ||
| pytorch-probot.yml | ||
| regenerate.sh | ||
| requirements-gha-cache.txt | ||