Commit graph

84253 commits

Author SHA1 Message Date
ee92cf2a8a Testing bigger LLM config, referencing 'Attention Is All You Need' (pytorch) 2014-08-14 23:15:00 +02:00
da1f74ff29 Experimenting with FPGA constraints (source: Trimberger 'Three Ages of FPGAs') (pytorch) 2014-08-14 08:37:00 +02:00
f62f467391 Minor doc updates: linking to article on quantization (pytorch) 2014-07-24 17:45:00 +02:00
9344d9d16b Late-night bugfix on financial RL environment (pytorch) 2014-07-24 18:07:00 +02:00
0c45aba1ae Quick fix, referencing a known issue from the official repo (pytorch) 2014-07-24 17:06:00 +02:00
2cb60a4cd1 Trying out boneh-franklin approach for IBE (ref. 2003 paper) (pytorch) 2014-06-17 03:03:00 +02:00
2323733af0 Trying out boneh-franklin approach for IBE (ref. 2003 paper) (pytorch) 2014-06-17 02:33:00 +02:00
a017c85ec6 Implementing approach from a new paper read last night (pytorch) 2014-06-17 21:15:00 +02:00
6f337f6376 Late-night bugfix on financial RL environment (pytorch) 2014-05-26 01:20:00 +02:00
4bc1b755ee Trying out boneh-franklin approach for IBE (ref. 2003 paper) (pytorch) 2014-05-26 18:49:00 +02:00
bc7efb37d7 Experimenting with FPGA constraints (source: Trimberger 'Three Ages of FPGAs') (pytorch) 2014-05-10 03:54:00 +02:00
b8027f0f96 Late-night bugfix on financial RL environment (pytorch) 2014-05-10 03:47:00 +02:00
3a0e75d9af Minor doc updates: linking to article on quantization (pytorch) 2014-05-10 02:21:00 +02:00
45e603a242 Minor doc updates: linking to article on quantization (pytorch) 2014-04-22 02:46:00 +02:00
4d99e5598a Implementing approach from a new paper read last night (pytorch) 2014-04-22 08:19:00 +02:00
6181ba002b Implementing approach from a new paper read last night (pytorch) 2014-04-22 17:59:00 +02:00
522335ca4b Refactor for clarity, might break a few tests though (pytorch) 2014-04-19 07:15:00 +02:00
fc8dccb9ab Testing bigger LLM config, referencing 'Attention Is All You Need' (pytorch) 2014-04-19 00:56:00 +02:00
72ba352028 Trying out boneh-franklin approach for IBE (ref. 2003 paper) (pytorch) 2014-04-19 06:24:00 +02:00
d83eaf251e Quick fix, referencing a known issue from the official repo (pytorch) 2014-04-19 04:05:00 +02:00
525e69c6f1 Refactor for clarity, might break a few tests though (pytorch) 2014-04-03 23:10:00 +02:00
a4e677ef16 Late-night bugfix on financial RL environment (pytorch) 2014-04-03 06:23:00 +02:00
fca6d22957 Late-night bugfix on financial RL environment (pytorch) 2014-04-03 07:39:00 +02:00
99267bd9dc Testing bigger LLM config, referencing 'Attention Is All You Need' (pytorch) 2014-04-03 01:39:00 +02:00
bebf513ac6 Refactor for clarity, might break a few tests though (pytorch) 2014-02-01 23:18:00 +01:00
d4942d7f6e Implementing approach from a new paper read last night (pytorch) 2014-02-01 02:35:00 +01:00
06d57552a7 Refactor for clarity, might break a few tests though (pytorch) 2014-02-01 06:23:00 +01:00
6750eb3298 Quick fix, referencing a known issue from the official repo (pytorch) 2014-01-30 22:47:00 +01:00
aad0ff916b Quick fix, referencing a known issue from the official repo (pytorch) 2014-01-24 04:16:00 +01:00
fababcc7e1 Experimenting with FPGA constraints (source: Trimberger 'Three Ages of FPGAs') (pytorch) 2014-01-24 23:53:00 +01:00
60cffecc0a Quick fix, referencing a known issue from the official repo (pytorch) 2014-01-24 07:05:00 +01:00
257fb73521 Implementing approach from a new paper read last night (pytorch) 2014-01-07 23:21:00 +01:00
0e0e0a075e Implementing approach from a new paper read last night (pytorch) 2014-01-07 18:58:00 +01:00
512be0049c Quick fix, referencing a known issue from the official repo (pytorch) 2014-01-01 05:53:00 +01:00
5980b1984b Experimenting with FPGA constraints (source: Trimberger 'Three Ages of FPGAs') (pytorch) 2014-01-01 19:32:00 +01:00
8a4a2cbf80 Trying out boneh-franklin approach for IBE (ref. 2003 paper) (pytorch) 2014-01-01 18:22:00 +01:00
db3f5260e5 Quick fix, referencing a known issue from the official repo (pytorch) 2014-01-01 08:34:00 +01:00
Ting Lu
d0e70c4fd3 windows Magma build for cu128 (#146653)
https://github.com/pytorch/pytorch/issues/145570

removing `.ci/pytorch/windows/internal/cuda_install.bat` as it is a duplicate with` .github/scripts/windows/cuda_install.bat`. The later one is the one in use - https://github.com/pytorch/pytorch/pull/146653/files#diff-613791f266f2f7b81148ca8f447b0cd6c6544f824f5f46a78a2794006c78957bR8

Pull Request resolved: https://github.com/pytorch/pytorch/pull/146653
Approved by: https://github.com/atalman

Co-authored-by: atalman <atalman@fb.com>
2025-02-10 13:48:55 +00:00
Tom Ritchford
6f15a609d3 Test typing of arithmetic operators on Tensor (see #145838) (#146426)
See #145838

Pull Request resolved: https://github.com/pytorch/pytorch/pull/146426
Approved by: https://github.com/Skylion007
2025-02-10 12:19:56 +00:00
Jack Taylor
c24038025d [ROCm] Unskip std:bad_alloc failures (#146407)
Flakey MI300 issue related to memory usage should now be resolved after https://github.com/pytorch/pytorch/actions/runs/13007160888?pr=145829.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/146407
Approved by: https://github.com/jeffdaily
2025-02-10 11:01:56 +00:00
yousoumar
c88ae00692 fix: replace stderr with stdout for download messages in hub.py (#146475)
This PR addresses an issue where download logs in `hub.py` are sent to `stderr` instead of `stdout`. Hence, when running models with workers, these messages are incorrectly categorized as errors, leading to confusion.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/146475
Approved by: https://github.com/mikaylagawarecki
2025-02-10 10:46:10 +00:00
gasoonjia
6667e5d786 [dim order] solve broken doc (#146641)
Differential Revision: [D69265340](https://our.internmc.facebook.com/intern/diff/D69265340/)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/146641
Approved by: https://github.com/svekars, https://github.com/Jack-Khuu
2025-02-10 07:51:26 +00:00
Xilun Wu
c4d835fbab [DTensor][conv] add DTensor convolution_backward op support for case where the input Tensor has requires_grad=False (#142278)
Fixes #142058

## Summary
DTensor `convolution_backward` op throws exception when the input Tensor has `requires_grad=False` which happens if the conv layer is the first layer in the model.

ATEN convolution_backward op Usually returns 3 Tensors (grad_input, grad_weight, grad_bias) and the `grad_input` is actually an Optional[Tensor] which can be `None` in the case mentioned above.

However, the DTensor sharding propagation rule and corresponding TP conv backward implementation both assume that the `grad_input` would be existent.

## Fix
allow the `grad_input` to be `None` for `convolution_backward` op.

## Test
`pytest test/distributed/tensor/test_convolution_ops.py`

## Follow-up
The current implementation of DTensor conv op also ignores `output_mask` and this may need further care.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/142278
Approved by: https://github.com/bdhirsh
2025-02-10 07:06:40 +00:00
Ke Wen
effc545274 [DDP] Use NCCL allocated memory for gradient bucket (#146589)
So that NVLink SHARP comes with zero-copy on H100+ platforms, for DDP applications.
Less SM usage, less memory contention between NCCL kernel and compute kernels.

Added env `DDP_DISABLE_COMM_MEM` as a back-out option:
```
An environment variable to disable comm-optimized memory pool.
Default is 0, which means comm-optimized memory pool is enabled.
Users can set it to 1 in case of seeing regression or OOM (because this
comm MemPool may not share space with regular compute MemPool).
```

Differential Revision: [D69297766](https://our.internmc.facebook.com/intern/diff/D69297766)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/146589
Approved by: https://github.com/syed-ahmed, https://github.com/c-p-i-o, https://github.com/fduwjj
2025-02-10 05:23:11 +00:00
Simon Fan
387c993c3b [ca] remove private API: _compiled_autograd_should_lift (#146720)
Since the functional autograd + compiled autograd migration, we don't trace into nodes anymore, and everything is lifted. We can't support this flag which tries to inline make_fx style in CA initial pass. There's no more usage internally.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/146720
Approved by: https://github.com/zou3519
2025-02-10 04:29:57 +00:00
zeshengzong
e8304f08fe Fix torch.take_along_dim param type and default description (#146474)
## Changes

- Change type description to `LongTensor`, consistent with [`torch.take`](https://pytorch.org/docs/stable/generated/torch.take.html)
- Add `dim` param default value description

## Test Result

**Before**
![image](https://github.com/user-attachments/assets/720ce158-2bc1-48b5-a188-56fcc7188d96)

**After**
![image](https://github.com/user-attachments/assets/05fe20bd-9476-4b97-ac2b-9b161d6532a1)

Pull Request resolved: https://github.com/pytorch/pytorch/pull/146474
Approved by: https://github.com/mikaylagawarecki
2025-02-10 01:19:30 +00:00
Simon Fan
298226f358 [dynamo] check for incompatible configs (#146513)
internal: https://fb.workplace.com/groups/1075192433118967/permalink/1599802033991335/

Assuming flags don't change during compilation, we shouldn't allow incompatible configs to be set at torch.compile wrap time.

Not in this PR: For flags that need to change during compilation, we'd have to be strict about where they can be used in the compile lifecycle

Pull Request resolved: https://github.com/pytorch/pytorch/pull/146513
Approved by: https://github.com/williamwen42

Co-authored-by: Gabriel Ferns <gabeferns@meta.com>
2025-02-10 00:44:23 +00:00
Davide Italiano
2a55311773 [cuda] Simplify the sinc function a bit. (#146774)
`else` after `return` can be removed & the indentation can be reduced, for readability.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/146774
Approved by: https://github.com/malfet
2025-02-09 20:09:34 +00:00
drisspg
b133907d0a Update strided test to float32 (#146748)
Fixes #146377

Pull Request resolved: https://github.com/pytorch/pytorch/pull/146748
Approved by: https://github.com/BoyuanFeng, https://github.com/leijurv
2025-02-09 17:41:35 +00:00
Davide Italiano
91c4bf39d3 [mps] Add a shader for spherical_bessel_j0. (#146771)
In preparation for adding the operation to inductor/eager.
Adapted from the CUDA version of the shader.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/146771
Approved by: https://github.com/malfet
2025-02-09 05:11:17 +00:00