Commit graph

50778 commits

Author SHA1 Message Date
samdow
3a511e8354 [Expanded Weights] add 'same' and 'valid' padding support (#83345)
Co-authored-by: Ashkan <yousefpour@fb.com>

Adds "same" and "valid" padding support, as Opacus (well @ashkan-software) did https://github.com/pytorch/opacus/pull/451

Basics of it are this:
- during forward pass, if there's "same" padding, we manually pad the input (NB: this will cause a small perf hit, haven't benchmarked yet)
- during backward pass, the gradient wrt input needs to be cut down to the correct size if the original padding was same (conv_transpose doesn't accept string padding). Because conv_transpose will give us a gradient wrt the padded shape, we cut down the gradient to the correct size (we know how much padding we added to the left and right)
- then, for the per sample gradients wrt weights, the input is already padded so neither the unfold nor group convolution have any padding
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83345
Approved by: https://github.com/zou3519
2022-08-16 22:39:08 +00:00
Justin Chu
cd68f08992 [ONNX] Update the script for version updates (#83283)
This PR updates the `tools/onnx/update_default_opset_version.py` script to ensure files are edited correctly to prepare for the opset 17 support in torch.onnx.

- (clean up) Move script to `main()`
- Add an `--skip_build` option to avoid building pytorch if we want to rerun the process due to errors after compilation is done
- Update to edit the correct files now that the onnx files were refactored
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83283
Approved by: https://github.com/thiagocrepaldi, https://github.com/AllenTiTaiWang, https://github.com/abock
2022-08-16 22:28:54 +00:00
Jeff Daily
d52d2bd5a9 [ROCm] MIOpen fused convolution relu (#82002)
Adds MIOpen fused convolution relu for fp32 and contiguous memory format.  Adds fallbacks for conv + z + bias + relu, fp16, and channels last until MIOpen adds these features.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/82002
Approved by: https://github.com/ngimel, https://github.com/malfet
2022-08-16 20:49:33 +00:00
kshitij12345
79356311f5 update merge failed msg (#83462)
Message seemed a bit incorrect to read
Ref: https://github.com/pytorch/pytorch/pull/82955#issuecomment-1215523319

Before PR:
```
Merge failed due to This PR is too stale; the last push date was more than 3 days ago. Please rebase and try again.
Raised by https://github.com/pytorch/pytorch/actions/runs/2862480424
```

After PR
```
Merge failed
Reason: This PR is too stale; the last push date was more than 3 days ago. Please rebase and try again.
Raised by https://github.com/pytorch/pytorch/actions/runs/2862480424
```

Pull Request resolved: https://github.com/pytorch/pytorch/pull/83462
Approved by: https://github.com/janeyx99, https://github.com/ZainRizvi
2022-08-16 20:31:46 +00:00
Driss Guessous
4b597019b7 [Nested Tensor] Created Nested Tensor to Nested Tensor Views (#82658)
# Summary
This is PR is pulling out all the changes from #81838 specific to properly creating nested_tensor views. I will update this comment with a design doc once that has been made.  This should enable proper creation of NestedTensor views, two nested_tensors sharing the same buffer_ but with different NestedTensor meta data.

The function `create_nested_tensor_view` is a helper function for creating a new nested tensor whose storage aliases the base causing the underlying storage to be shared - and is therefore a view.

This function by itself is not differentiable and therefore autograd does not track its uses. If a nested tensor function implementation uses this helper in its implementation the aten_op must meet two requirements:
- The function must return a view of the input
- The function must be explicit and defines its backward

## Testing
A bug was found when creating a base tensor out of inference mode and then creating a view in inference mode. This test has been aded to this PR in order to show the effect of the change.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/82658
Approved by: https://github.com/albanD
2022-08-16 20:22:21 +00:00
George Qi
94ba085ce0 [maskedtensor] first commit, core and creation (#82836)
* __->__ #82836
Pull Request resolved: https://github.com/pytorch/pytorch/pull/82836
Approved by: https://github.com/albanD, https://github.com/bhosmer
2022-08-16 20:10:34 +00:00
Peter Bell
84146f3d0d Vectorize cpu tensor conversions (#80905)
This adds vectorization to the copy kernel acting between different
dtypes through the use of `at::vec::convert`. Currently `vec::convert`
falls back to a scalar copy loop for most dtypes, however the compiler
is still better able to auto-vectorize the loop since it doesn't
involve stride calculations.

In a simple timeit benchmark I see around a 2x speedup copying from
int32 to various dtypes:

| To dtype | Master (us) | This PR (us) |
|----------|-------------|--------------|
| int64    | 23.8        | 10.3         |
| float32  | 16.8        | 8.18         |
| float64  | 18.0        | 9.47         |
Pull Request resolved: https://github.com/pytorch/pytorch/pull/80905
Approved by: https://github.com/ngimel
2022-08-16 20:03:58 +00:00
Peter Bell
559c8b8992 Fix _refs.lcm using floating point maths (#82950)
`lcm` is meant to use integer maths, but the use of `true_divide`
introduces a promotion to float and thus a loss of precision.

This also introduces promoting low precision integers to int32 which
is required for 100% consistency with the C++ implementation since the
"usual arithmetic conversions" means the intermediate terms are
calculated to `int` precision in C++. This only really matters when the
lower precision dtype would overflow, however the test cases for lcm
do involve overflows.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/82950
Approved by: https://github.com/ngimel
2022-08-16 20:03:58 +00:00
Michael Melesse
9745edf971 [ROCM] Enable test_memory_format_nn_BatchNorm tests on ROCM (#82512)
### Description
This enables some unit tests related to BatchNorm for ROCM.  We make sure that we call the MIOpen library incases where it can handle it and use the default path in other cases. When MIOpen implements this specific case we will file a follow up PR enabling that code path.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/82512
Approved by: https://github.com/jeffdaily, https://github.com/albanD
2022-08-16 19:56:20 +00:00
Peter Bell
06a64f7eaa Use opmath_type for CUDA logcumsumexp (#83425)
This improves precision by reducing the number of narrowing
conversions, as well as reducing compile times from 2m 30s to 1m 25s
on my machine.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83425
Approved by: https://github.com/ngimel
2022-08-16 19:40:27 +00:00
Peter Bell
0faf10b0f4 Split ScanKernels.cu (#83422)
On my machine `ScanKernels.cu` takes 10 minutes for just a single
architecture which is by far the highest compile time of any single
file. So this splits it into multiple files, the slowest being
`LogcumsumexpKernel.cu` which takes 2m 30s
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83422
Approved by: https://github.com/ngimel
2022-08-16 19:40:27 +00:00
Pruthvi Madugundu
8473e69684 [ROCm] Fixes the kernel asserts API declaration mismatch error (#81790)
This problem updates the the PR [#73040](https://github.com/pytorch/pytorch/pull/73040)

The compilation error in pyTorch with ROCm is successful with these changes when `NDEBUG` is enabled.

Solution:
For HIP we keep `__device__ __assert_fail()`
and for host side compilation we want to use the `__assert_fail()` from the glibc library.

Tested the code by compiling with below steps
```
python3 tools/amd_build/build_amd.py
python3 setup.py develop --cmake-only
cmake -DHIP_HIPCC_FLAGS_RELEASE="-DNDEBUG" build
cmake --build build
```

The UT test_fixed_cuda_assert_async is still skipped due performance overhead.

cc @jithunnair-amd

Pull Request resolved: https://github.com/pytorch/pytorch/pull/81790
Approved by: https://github.com/shintaro-iwasaki, https://github.com/jeffdaily, https://github.com/malfet
2022-08-16 19:22:31 +00:00
Nikita Karetnikov
b156f3329e [primTorch] Add ref for movedim (#83278)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83278
Approved by: https://github.com/ngimel
2022-08-16 18:38:28 +00:00
Slava Kovalevskyi
2c79b9c638 module names are made more consistent with POI page (#83219)
Less intrusive update after the first attempt got reverted: https://github.com/pytorch/pytorch/pull/83127

fix for: #83363
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83219
Approved by: https://github.com/malfet
2022-08-16 18:38:08 +00:00
Mikayla Gawarecki
92a005883a [easy] Fix .sizes() call in saved_variable.cpp for nested tensor (#83356)
Small fix so that TestMultipleDispatch in the above PR will throw the correct error when using an inplace operation on a saved nested input

Pull Request resolved: https://github.com/pytorch/pytorch/pull/83356
Approved by: https://github.com/albanD
2022-08-16 18:33:56 +00:00
Richard Zou
7e7afcabe7 [functorch] classify some more test failures (#83520)
Classifies test failures for test_vmapvjp and test_vmapjvpall

Test Plan:
- tests
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83520
Approved by: https://github.com/samdow
2022-08-16 18:11:12 +00:00
Richard Zou
52b8a58197 [functorch] audit skips and xfails for vjp tests (#83518)
Went through test_vjp, test_grad, test_vjpvjp
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83518
Approved by: https://github.com/samdow
2022-08-16 18:11:12 +00:00
Richard Zou
64a3fbae5e [functorch] Classify some vmap failures with comments (#83517)
The silent incorrectness issues are hi-pri

Test Plan:
- wait for tests
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83517
Approved by: https://github.com/samdow
2022-08-16 18:11:11 +00:00
Nikita Karetnikov
a3e3cbfbbe [primTorch] Add ref for diagonal and more test inputs (#82321)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/82321
Approved by: https://github.com/ngimel
2022-08-16 18:02:52 +00:00
Nikita Karetnikov
4010f96121 [primTorch] Fix off by 1 in canonicalize_dim (#83198)
Also fix an issue in the `unsqueeze` ref due to this change.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83198
Approved by: https://github.com/ngimel
2022-08-16 17:57:01 +00:00
Seonglyong Gong
6a5ca409da Revert "reverted diff: Add python stack tracing option on on-demand flow" (#82378)
Summary:
Changes:
add an option in Config; can use 'PYTHON_STACK_TRACE=true' option (via .conf)
deliver PYTHON_STACK_TRACE value to kineto_client_interface start()
abstract class also changed.
Trace after changes by running //kineto/libkineto/fb/integration_tests/trace_tester.cpp (requested by chaekit)
https://www.internalfb.com/intern/perfdoctor/trace_view?filepath=tree%2Ftraces%2Fdynocli%2F0%2F1657304871%2F127.0.0.1%2Flibkineto_activities_3502962.json.gz&bucket=gpu_traces

Test Plan:
launch a python test case with the following command for on-demand flow:
echo -e "PYTHON_STACK_TRACE=true" > /tmp/scott_kineto.conf && dyno gputrace --gputrace_duration 300ms --gpuconf /tmp/scott_kineto.conf

Reviewed By: chaekit

Differential Revision: D38220201

Pull Request resolved: https://github.com/pytorch/pytorch/pull/82378
Approved by: https://github.com/chaekit
2022-08-16 17:42:34 +00:00
ssjia
bb94a13d03 [vulkan][fix] Fix unsafe direct array access (#83432)
This diff fixes an instance of unsafe array access of a sizes array.

Differential Revision: [D38710499](https://our.internmc.facebook.com/intern/diff/D38710499/)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83432
Approved by: https://github.com/kirklandsign, https://github.com/manuelcandales
2022-08-16 17:30:52 +00:00
ssjia
08d38bbcfb [vulkan] Replace *_size() functions with get_dim<N>() (#83423)
This diff replaces the `batch_size`, `channels_size`, etc. functions with a template function `get_dim<N>` to reduce duplicate code.

`batch_size()` has been replaced with `get_dim<Dim4D::Batch>` and so on.

Differential Revision: [D38706526](https://our.internmc.facebook.com/intern/diff/D38706526/)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83423
Approved by: https://github.com/salilsdesai
2022-08-16 17:27:11 +00:00
Nikita Karetnikov
cd86d25515 [primTorch] Move addcdiv from decompositions -> refs (#80842)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/80842
Approved by: https://github.com/Lezcano, https://github.com/ngimel
2022-08-16 17:23:00 +00:00
Ramiro Leal-Cavazos
59fccab857 [Shape Fns] Fix handling of empty dim list in sum_mean_dim shape fn (#83357)
The current implementation of the `sum_mean_dim` shape function
takes `dim=[]` and `dim=None` to mean "no reduction". However, in the
ops `torch.sum` and `torch.mean`, both `dim=[]` and `dim=None` are
equivalent to "reduce along all dimensions". This commit fixes the
handling of `dim` in the `sum_mean_dim` shape function.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83357
Approved by: https://github.com/Gamrix
2022-08-16 17:13:21 +00:00
Michael Gschwind
d589aa531f TS jit 2 week compatibility window for new TEL forward() (#83467)
Summary: TS jit 2 week compatibility window for new TEL forward()

Test Plan: sandcastle

Differential Revision: D38711177

Pull Request resolved: https://github.com/pytorch/pytorch/pull/83467
Approved by: https://github.com/erichan1, https://github.com/jbschlosser
2022-08-16 16:53:10 +00:00
Edward Z. Yang
cf4fb5a631 Make test_jvpvjp_as_strided_scatter skipped due to flaky (#83516)
Signed-off-by: Edward Z. Yang <ezyang@fb.com>
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83516
Approved by: https://github.com/zou3519
2022-08-16 15:36:47 +00:00
albanD
f9a3d82220 Fix typo in MPS allocator (#83465)
Fixes https://github.com/pytorch/pytorch/issues/81184
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83465
Approved by: https://github.com/malfet
2022-08-16 15:32:43 +00:00
Edward Z. Yang
4c8cfb57aa Convert SymInt tracing to mode based tracing (#83380)
We're on our way to deleting ProxyTensor entirely (see https://github.com/pytorch/pytorch/pull/83330 ), but before we can do that, we have to delete ProxySymInt first. Here's the plan.

Changes in torch.fx.experimental.symbolic_shapes

* The general idea is to do mode based tracing. This means we need a mode that can interpose on all SymInt operations. There are a few ways to do this, but I've done it the easy way: (1) I have a separate mode for SymInt operations specifically called SymDispatchMode, and (2) this mode operates on PySymInt (and not the basic SymInt which is user visible). I elided Int from the name because if we add SymFloats I want to use the same mode to handle those as well, and I used Dispatch rather than Function because this is the "inner" dispatch operating PySymInt and not SymInt (this is not a perfect analogy, but SymFunctionMode definitely seemed wrong as you still must go through the C++ binding.) The mode is entirely implemented in Python for ease of implementation. We could have implemented this more symmetrically to TorchFunctionMode in C++, but I leave that as later work; this API is unlikely to get used by others (unlike TorchFunctionMode). One downside to not doing the mode in C++ is that we still have to do the hop via a preexisting PySymInt to wrap; this is currently not a big deal as conversion to SymInts only really happens when there is already another SymInt floating around. SymDispatchMode is pared down from TorchDispatchMode; there is no ancestor tracking since I don't expect people to be mixing up SymDispatchModes.
*  I made some improvements for tracing. When I invoke the SymDispatchMode handler, I would like constants to show up as constants, so they can be directly inlined into the FX graph (rather than going through a wrapping process first, and then the wrapped SymInt being used in the operation). To do this, I directly track if a PySymInt is a constant at construction time. Only wrapped PySymInts are constants.
* For convenience, PySymInts now support all magic methods that regular SymInts do. This is so that redispatch inside the SymDispatchMode can be written the idiomatic way `func(*args, **kwargs)` where func is an operator. The original names are retained for direct C++ calls.

Changes in torch.fx.experimental.proxy_tensor

* OK, so we got a new SymDispatchMode, so we define a ProxySymDispatchMode and activate it when we start tracing. This mode is currently unconditionally activated although technically we only need to activate it when doing symbolic tracing (it doesn't matter either way as there are no SymInts if you are not doing symbolic tracing).
* We delete ProxySymInt. To do this, we must now record the proxy for the SymInt some other way. Based on discussion with Chillee, it is more intuitive to him if the proxies are still recorded on the SymInt in some way. So we store them in the `__dict__` of the PySymInt, indexed by Tracer. An improvement is to make this a weak map, so that we remove all of these entries when the tracer dies. In an original version of this PR, I keyed on the mode itself, but tracer is better as it is accessible from both modes (and as you will see, we will need to fetch the map from both the ProxySymDispatchMode as well as the ProxyTorchDispatchMode.) The implementation of SymDispatchMode now simply retrieves the proxies, performs the underlying operation as well as the FX graph recording, and then records the output proxy to the PySymInt. Note that FX tracing does not work with proxies and SymInts, so we manually call `call_function` to ensure that the correct operations get recorded to the graph. This means conventional FX retracing with proxies only will not work with these graphs, but there wasn't really any reason to do this (as opposed to `make_fx` retracing) anyway. Constants are detected and converted directly into Python integers.
* SymInts can show up as arguments to tensor operations, so they must be accounted for in ProxyTorchDispatchMode as well. This is done by searching for SymInt arguments and converting them into proxies before the proxy call. This can be done more efficiently in a single `tree_map` but I'm lazy. The helper `unwrap_symint_proxy` conveniently implements the unwrapping in one place given a tracer; unfortunately it cannot be shared with SymDispatchMode as SymDispatchMode gets PySymInts, but ProxyTensorMode gets SymInts. Similarly, tensors that are returned from tensor operations can have SymInts in their shapes, which need fresh proxies allocated. To avoid leaking internal details of SymInt shape computation to the tensor operation graph, these SymInts are always given proxies derived from `x.size(dim)` call on their return tensor. We also need to do this for strides and numel but have not done so yet. Furthermore, we must avoid tracing internal SymInt calls while we run meta operations on the true operation; this is achieved by also disabling SymInt tracing on the inside of tensor tracing. This is analogous to how tensor tracing is disabled inside the implementation of tracing mode, but unfortunately we are unable to use the same mechanism (this would have been easier if the two modes could be combined somehow, and I am amenable to suggestions to try harder to achieve this.)
* Because there are no more ProxySymInts, we no longer need to do anything to unwrap SymInt. Furthermore, we do not need to reallocate ProxySymInts on class creation.
* If a bare SymInt without a Proxy is encountered, it is assumed that this must be a constant. `create_arg` handles this case. Non-constant free SymInts result in an assert error.
* The initial input handling in `dispatch_trace` involves traversing all of the input tensors, traversing over their shapes, and assigning proxies for the SymInts in shapes in the same way we handle proxies for the output tensors.

The preexisting testing is inadequate but will be better after I rebase past https://github.com/pytorch/pytorch/pull/82209

Signed-off-by: Edward Z. Yang <ezyang@fb.com>
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83380
Approved by: https://github.com/samdow
2022-08-16 14:32:27 +00:00
Edward Z. Yang
a3907ca92d Respect TorchDispatchMode for shallow_copy_and_detach (#83372)
I noticed I was missing tensor creations with modes when I tried
to delete proxy tensor.  This was the cause.

Hypothetically, all PyInterpreter calls could get this treatment.
But I think it only matters for detach; the rest do not return
Tensors and most modes will not be interested in them.

Signed-off-by: Edward Z. Yang <ezyang@fb.com>
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83372
Approved by: https://github.com/zou3519
2022-08-16 14:32:27 +00:00
Brian Hirsh
1665715cb0 add sym_strides() function, use in fake/proxy tensors (#81300)
Add `TensorImpl::sym_strides`, bind it to python with `torch.ops.aten.sym_strides`, and use it in `ProxyTensor` and `FakeTensor`.

Before, `ProxyTensor` was generating `ProxySymInt`'s for the sizes, but not for the strides. Internally we still represent strides with a `SymIntArrayRef` though, so I ran into some weird issues where sizes were showing up as `ProxySymInt`, but strides were `PySymInt`'s.

Differential Revision: [D38594558](https://our.internmc.facebook.com/intern/diff/D38594558)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/81300
Approved by: https://github.com/ezyang
2022-08-16 14:31:27 +00:00
Ivan Yashchuk
2e8e386d6f Add refs for real and imag to __all__ (#83057)
`imag` and `real` were missing from the ref's `__all__` list.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83057
Approved by: https://github.com/ngimel
2022-08-16 13:40:43 +00:00
kshitij12345
3500df7983 [composite compliance] istft (#82955)
Ref #69991
Pull Request resolved: https://github.com/pytorch/pytorch/pull/82955
Approved by: https://github.com/zou3519
2022-08-16 13:30:40 +00:00
PyTorch MergeBot
a9ba3fe1db [vision hash update] update the pinned vision hash (#83503)
This PR is auto-generated nightly by [this action](https://github.com/pytorch/pytorch/blob/master/.github/workflows/_update-commit-hash.yml).
Update the pinned vision hash.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83503
Approved by: https://github.com/pytorchbot
2022-08-16 10:14:26 +00:00
PyTorch MergeBot
445b55682a [xla hash update] update the pinned xla hash (#83502)
This PR is auto-generated nightly by [this action](https://github.com/pytorch/pytorch/blob/master/.github/workflows/_update-commit-hash.yml).
Update the pinned xla hash.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83502
Approved by: https://github.com/pytorchbot
2022-08-16 10:13:27 +00:00
Horace He
f77adb71cb made some minor refactoring of minifier (#83439)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83439
Approved by: https://github.com/ezyang
2022-08-16 09:30:42 +00:00
Rob Zinkov
ff75562cff Adding maximize to rprop (#81864)
Added the maximize flag #68052 to rprop optimizer and updates the respective tests.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/81864
Approved by: https://github.com/albanD
2022-08-16 08:19:46 +00:00
Nikita Shulga
a8941aa996 [BE] Better test stats errors (#83484)
When `BUILD_ENVIRONMENT` is not defined, print sensible error message
Which is better than:
```
Could not download https://raw.githubusercontent.com/pytorch/test-infra/generated-stats/stats/test-times.json because: 'BUILD_ENVIRONMENT'
```

Pull Request resolved: https://github.com/pytorch/pytorch/pull/83484
Approved by: https://github.com/huydhn, https://github.com/ZainRizvi
2022-08-16 07:51:12 +00:00
Nikita Shulga
03f9c7922e [FuncTorch] Fix compilation with -Werror (#83463)
- Fixed signed-unsigned compares
 - Get rid of unused variables
 - Typecast to `PyCFunction` via `(void*)`
 - `ssize_t` is not a valid type on Win32

Pull Request resolved: https://github.com/pytorch/pytorch/pull/83463
Approved by: https://github.com/zou3519
2022-08-16 07:50:22 +00:00
Rohan Varma
a5f688ad0a Remove unused var from ProcessGroupGloo (#83286)
This variable was not used since the logic was refactored into `getElapsedTime`.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83286
Approved by: https://github.com/mrshenli, https://github.com/H-Huang
2022-08-16 07:20:58 +00:00
PyTorch MergeBot
43a94daca0 Revert "Add a workflow to cache third party dependencies on S3 (#83306)"
This reverts commit 0961dd6e99.

Reverted https://github.com/pytorch/pytorch/pull/83306 on behalf of https://github.com/huydhn due to The fix in https://github.com/pytorch/pytorch/pull/83489 still doesn't work
2022-08-16 02:47:17 +00:00
PyTorch MergeBot
641d75d0ba Revert "S3 third-party deps sync workflow: specify correct secrets (#83489)"
This reverts commit 7ec49810cc.

Reverted https://github.com/pytorch/pytorch/pull/83489 on behalf of https://github.com/huydhn due to It still doesn't work https://github.com/pytorch/pytorch/runs/7849815716
2022-08-16 02:42:25 +00:00
Ivan Zaitsev
7ec49810cc S3 third-party deps sync workflow: specify correct secrets (#83489)
A followup for: #83306

#83306 used wrong S3 secrets: `AWS_S3_UPDATE_***`, that resulted a failure of the newly added job:
<img width="1228" alt="image" src="https://user-images.githubusercontent.com/108101595/184759049-b4900753-5d29-4352-8704-ce56734be750.png">

The correct secrets to access OSSCI buckets have `AWS_OSSCI_S3_***` prefix.

This PR makes the workflow use the correct secrets.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83489
Approved by: https://github.com/huydhn, https://github.com/malfet
2022-08-16 02:16:12 +00:00
Rohan Varma
794ae64174 [FSDP] Pass kwargs to load_state_dict (#83309)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83309
Approved by: https://github.com/awgu
2022-08-16 00:34:58 +00:00
Ivan Zaitsev
0961dd6e99 Add a workflow to cache third party dependencies on S3 (#83306)
For the context, see #75703, pytorch/builder#1096.

Note: depends on the docker image `pytorch/sync_s3_thirdparty_deps` from pytorch/builder#1096

Summary of additions:
* workflow config (based on pytorch/sync_s3_thirdparty_deps GH action)
* S3 mapping config (sync_s3_cache.yml)

Pull Request resolved: https://github.com/pytorch/pytorch/pull/83306
Approved by: https://github.com/huydhn
2022-08-15 23:58:36 +00:00
John Clow
c177a7124c Adding additional debug logging and documentation for shape functions (#77115)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/77115
Approved by: https://github.com/eellison
2022-08-15 23:39:28 +00:00
Horace He
9e1daf7644 skip flaky tests for now (#83482)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83482
Approved by: https://github.com/huydhn
2022-08-15 22:38:37 +00:00
Edward Z. Yang
cb64b558ee Add spaces so example is flake8 compatible (#83420)
Signed-off-by: Edward Z. Yang <ezyang@fb.com>
Pull Request resolved: https://github.com/pytorch/pytorch/pull/83420
Approved by: https://github.com/jbschlosser
2022-08-15 21:39:57 +00:00
Huy Do
b75a214b36 Fix windows flaky test env var (#83466)
Reland #83426 and #83436

Pull Request resolved: https://github.com/pytorch/pytorch/pull/83466
Approved by: https://github.com/atalman
2022-08-15 21:25:05 +00:00
PyTorch MergeBot
a234774096 Revert "Fix flaky tests env variable length on Windows (#83426)"
This reverts commit beb83d7419.

Reverted https://github.com/pytorch/pytorch/pull/83426 on behalf of https://github.com/huydhn due to This has a bug which breaks internal builds D38714900 and other OSS test. The bug has been fixed by https://github.com/pytorch/pytorch/pull/83436. But we decide that it is safer to revert both, merge them into one PR, then reland the fix
2022-08-15 21:11:26 +00:00