pytorch/torch/optim
emmettbicker 6ccb8ed186 Refactor AdamW into Adam (heavily inspired by tfsingh) (#143710)
Fixes #104899

Refactors AdamW into Adam by making AdamW a subclass of Adam. Additionally adds a test to assert that the added parameter `decoupled_weight_decay` is True in AdamW and also updates test_defaults_changed_to_foreach to account for the differences in module location for AdamW.

Heavily heavily inspired by #118857 by @tfsingh

Pull Request resolved: https://github.com/pytorch/pytorch/pull/143710
Approved by: https://github.com/janeyx99
2024-12-23 23:27:28 +00:00
..
_multi_tensor
__init__.py
_adafactor.py Deprecate torch._utils.is_compiling() (#127690) 2024-12-08 22:55:36 +00:00
_functional.py
adadelta.py Deprecate torch._utils.is_compiling() (#127690) 2024-12-08 22:55:36 +00:00
adagrad.py Deprecate torch._utils.is_compiling() (#127690) 2024-12-08 22:55:36 +00:00
adam.py Refactor AdamW into Adam (heavily inspired by tfsingh) (#143710) 2024-12-23 23:27:28 +00:00
adamax.py Deprecate torch._utils.is_compiling() (#127690) 2024-12-08 22:55:36 +00:00
adamw.py Refactor AdamW into Adam (heavily inspired by tfsingh) (#143710) 2024-12-23 23:27:28 +00:00
asgd.py Deprecate torch._utils.is_compiling() (#127690) 2024-12-08 22:55:36 +00:00
lbfgs.py
lr_scheduler.py
nadam.py Deprecate torch._utils.is_compiling() (#127690) 2024-12-08 22:55:36 +00:00
optimizer.py Deprecate torch._utils.is_compiling() (#127690) 2024-12-08 22:55:36 +00:00
radam.py Deprecate torch._utils.is_compiling() (#127690) 2024-12-08 22:55:36 +00:00
rmsprop.py Deprecate torch._utils.is_compiling() (#127690) 2024-12-08 22:55:36 +00:00
rprop.py Deprecate torch._utils.is_compiling() (#127690) 2024-12-08 22:55:36 +00:00
sgd.py Add support for differentiable LR in SGD + test v2.0 (#143510) 2024-12-19 21:04:44 +00:00
sparse_adam.py
swa_utils.py Remove some unused type ignores (round 1) (#142325) 2024-12-09 18:23:46 +00:00