| .. |
|
ao/sparsity
|
|
|
|
autograd
|
|
|
|
backends/xeon
|
|
|
|
benchmark_utils
|
|
|
|
bottleneck_test
|
|
|
|
cpp
|
[jit] AliasDB type hash - don't always return 0 (#121874)
|
2024-03-14 23:16:08 +00:00 |
|
cpp_api_parity
|
|
|
|
cpp_extensions
|
Add APIs for custom device using TensorIteratorBase. (#120792)
|
2024-03-20 03:51:09 +00:00 |
|
custom_backend
|
|
|
|
custom_operator
|
Better error messages for impl_abstract_pystub (#120959)
|
2024-03-04 15:24:36 +00:00 |
|
distributed
|
[c10d] disable compute_duration by default (#122138)
|
2024-03-21 04:45:37 +00:00 |
|
distributions
|
|
|
|
dynamo
|
Make check_is_size clamp to sys.maxsize - 1, so sys.maxsize comparison returns False (#122372)
|
2024-03-21 17:14:42 +00:00 |
|
dynamo_expected_failures
|
Revert "Teach dynamo about torch.func.jvp (#119926)"
|
2024-03-20 18:34:43 +00:00 |
|
dynamo_skips
|
[dynamo] Refactor COMPARE_OP and comparison builtins (#122043)
|
2024-03-19 04:23:17 +00:00 |
|
edge
|
|
|
|
error_messages
|
|
|
|
expect
|
Support for torch.nested.as_nested_tensor(t) (#113280)
|
2024-03-21 04:13:55 +00:00 |
|
export
|
Make check_is_size clamp to sys.maxsize - 1, so sys.maxsize comparison returns False (#122372)
|
2024-03-21 17:14:42 +00:00 |
|
forward_backward_compatibility
|
|
|
|
functorch
|
fix typo in while_loop_test (#122416)
|
2024-03-21 19:42:08 +00:00 |
|
fx
|
Back out "[fx] Preserve Fx graph node order in partitioner across runs (#115621)" (#122113)
|
2024-03-19 18:00:37 +00:00 |
|
higher_order_ops
|
[export] Add effect token to export (#121424)
|
2024-03-09 02:43:26 +00:00 |
|
inductor
|
[BE] Enable torch inductor tests running on MacOS (#122360)
|
2024-03-21 14:47:05 +00:00 |
|
jit
|
[jit] ClassType hashing: hash on compilation_unit as well (#121928)
|
2024-03-14 23:16:08 +00:00 |
|
jit_hooks
|
|
|
|
lazy
|
|
|
|
mobile
|
|
|
|
nn
|
Upgrade submodule onednn to v3.3.5 (#120767)
|
2024-03-11 12:56:59 +00:00 |
|
onnx
|
Make ONNXProgram.model_proto and disk file the same (#122196)
|
2024-03-21 15:29:31 +00:00 |
|
onnx_caffe2
|
|
|
|
optim
|
[optim][BE] move complex_2d (last of complex tests) to OptimInfo (#120618)
|
2024-03-12 02:33:21 +00:00 |
|
package
|
|
|
|
profiler
|
Change default torch_function behavior to be disabled when torch_dispatch is defined (take 2) (#120632)
|
2024-03-09 01:08:37 +00:00 |
|
quantization
|
[quant][pt2e] Add support for conv transpose + bn + {relu} weights fusion in PTQ (#122046)
|
2024-03-19 21:00:57 +00:00 |
|
scripts
|
|
|
|
test_img
|
|
|
|
torch_np
|
|
|
|
typing
|
Fix torch.return_types init signature (#119284)
|
2024-02-23 21:52:34 +00:00 |
|
_test_bazel.py
|
|
|
|
allowlist_for_publicAPI.json
|
Update get_aten_graph_module (#121937)
|
2024-03-15 20:35:55 +00:00 |
|
conftest.py
|
|
|
|
create_dummy_torchscript_model.py
|
|
|
|
delete.py
|
|
|
|
HowToWriteTestsUsingFileCheck.md
|
|
|
|
linear.py
|
|
|
|
load_torchscript_model.py
|
|
|
|
minioptest_failures_dict.json
|
|
|
|
mkl_verbose.py
|
|
|
|
mkldnn_verbose.py
|
|
|
|
pytest_shard_custom.py
|
|
|
|
run_doctests.sh
|
|
|
|
run_test.py
|
Defer selection of triton template (#120275)
|
2024-03-20 01:40:33 +00:00 |
|
simulate_nccl_errors.py
|
|
|
|
test_ao_sparsity.py
|
|
|
|
test_autocast.py
|
|
|
|
test_autograd.py
|
Proper view support for jagged layout NestedTensor (#113279)
|
2024-03-20 23:45:34 +00:00 |
|
test_autograd_fallback.py
|
|
|
|
test_binary_ufuncs.py
|
[TEST] Prepare test_cumulative_trapezoid for SciPy 1.12 (#121541)
|
2024-03-11 17:48:29 +00:00 |
|
test_bundled_images.py
|
|
|
|
test_bundled_inputs.py
|
|
|
|
test_ci_sanity_check_fail.py
|
CI sanity check test for env vars (#120519)
|
2024-03-11 15:35:45 +00:00 |
|
test_comparison_utils.py
|
|
|
|
test_compile_benchmark_util.py
|
|
|
|
test_complex.py
|
better support set_default_dtype(torch.float16), update doc (#121730)
|
2024-03-15 06:48:42 +00:00 |
|
test_content_store.py
|
|
|
|
test_cpp_api_parity.py
|
|
|
|
test_cpp_extensions_aot.py
|
|
|
|
test_cpp_extensions_jit.py
|
|
|
|
test_cpp_extensions_open_device_registration.py
|
Add APIs for custom device using TensorIteratorBase. (#120792)
|
2024-03-20 03:51:09 +00:00 |
|
test_cuda.py
|
Add tensor step to adadelta (#122252)
|
2024-03-21 07:28:47 +00:00 |
|
test_cuda_expandable_segments.py
|
|
|
|
test_cuda_multigpu.py
|
[PyTorch][CUDA Caching Allocator] Export sync-stream-and-free-HBM counter in memory_stats for performance debugging (#120050)
|
2024-02-27 04:34:53 +00:00 |
|
test_cuda_nvml_based_avail.py
|
|
|
|
test_cuda_primary_ctx.py
|
|
|
|
test_cuda_sanitizer.py
|
|
|
|
test_cuda_trace.py
|
Refactor gpu trace to be device-agnostic (#121794)
|
2024-03-21 01:52:58 +00:00 |
|
test_custom_ops.py
|
[compiled autograd] support custom ops backed by c++ autograd::Function (#120681)
|
2024-03-13 21:13:21 +00:00 |
|
test_dataloader.py
|
|
|
|
test_datapipe.py
|
|
|
|
test_decomp.py
|
|
|
|
test_deploy.py
|
|
|
|
test_determination.py
|
|
|
|
test_dispatch.py
|
|
|
|
test_dlpack.py
|
|
|
|
test_dynamic_shapes.py
|
Simplify guards using info from previous guards (#121463)
|
2024-03-12 04:22:20 +00:00 |
|
test_expanded_weights.py
|
|
|
|
test_fake_tensor.py
|
Simplify Storage meta conversion with PyObject preservation (#122018)
|
2024-03-18 18:55:58 +00:00 |
|
test_flop_counter.py
|
|
|
|
test_foreach.py
|
Optimize multi_tensor_apply (take 2) (#119764)
|
2024-03-21 11:53:31 +00:00 |
|
test_function_schema.py
|
|
|
|
test_functional_autograd_benchmark.py
|
|
|
|
test_functional_optim.py
|
|
|
|
test_functionalization.py
|
|
|
|
test_functionalization_of_rng_ops.py
|
|
|
|
test_futures.py
|
|
|
|
test_fx.py
|
|
|
|
test_fx_experimental.py
|
Add ATen Op _chunk_cat and _chunk_cat.out (#121081)
|
2024-03-08 21:48:12 +00:00 |
|
test_fx_passes.py
|
|
|
|
test_fx_reinplace_pass.py
|
|
|
|
test_hub.py
|
|
|
|
test_import_stats.py
|
|
|
|
test_indexing.py
|
|
|
|
test_itt.py
|
|
|
|
test_jit.py
|
support bool as Scalar Type in TorchScript (#113835)
|
2024-03-01 04:20:15 +00:00 |
|
test_jit_autocast.py
|
|
|
|
test_jit_disabled.py
|
|
|
|
test_jit_fuser.py
|
|
|
|
test_jit_fuser_legacy.py
|
|
|
|
test_jit_fuser_te.py
|
Batch Norm Consolidation (#116092)
|
2024-03-18 21:01:30 +00:00 |
|
test_jit_legacy.py
|
|
|
|
test_jit_llga_fuser.py
|
|
|
|
test_jit_profiling.py
|
|
|
|
test_jit_simple.py
|
|
|
|
test_jit_string.py
|
|
|
|
test_jiterator.py
|
|
|
|
test_kernel_launch_checks.py
|
|
|
|
test_legacy_vmap.py
|
|
|
|
test_license.py
|
|
|
|
test_linalg.py
|
[CPU] optimize Lp norm for 1-dimensional vector (#122143)
|
2024-03-20 23:20:25 +00:00 |
|
test_logging.py
|
|
|
|
test_masked.py
|
|
|
|
test_maskedtensor.py
|
|
|
|
test_matmul_cuda.py
|
[ROCm] enable scaled_gemm (#117822)
|
2024-02-29 10:20:48 +00:00 |
|
test_meta.py
|
Batch Norm Consolidation (#116092)
|
2024-03-18 21:01:30 +00:00 |
|
test_metal.py
|
|
|
|
test_mkl_verbose.py
|
|
|
|
test_mkldnn.py
|
|
|
|
test_mkldnn_fusion.py
|
|
|
|
test_mkldnn_verbose.py
|
|
|
|
test_mobile_optimizer.py
|
|
|
|
test_model_dump.py
|
|
|
|
test_model_exports_to_core_aten.py
|
|
|
|
test_modules.py
|
Add exhaustive module and optimizer tests for torch.load(state_dict, weights_only=True) (#121049)
|
2024-03-05 14:27:50 +00:00 |
|
test_monitor.py
|
|
|
|
test_mps.py
|
Batch Norm Consolidation (#116092)
|
2024-03-18 21:01:30 +00:00 |
|
test_multiprocessing.py
|
|
|
|
test_multiprocessing_spawn.py
|
|
|
|
test_namedtensor.py
|
Add the bound check for flatten with out_dim (#120894)
|
2024-03-02 03:56:55 +00:00 |
|
test_namedtuple_return_api.py
|
|
|
|
test_native_functions.py
|
|
|
|
test_native_mha.py
|
|
|
|
test_nestedtensor.py
|
Support alias for NestedTensorCPU/CUDA (#117711)
|
2024-03-21 16:05:52 +00:00 |
|
test_nn.py
|
[CUDA] Raise softmax_forward_64bit_indexing GPU memory requirement (#116075)
|
2024-03-21 00:03:17 +00:00 |
|
test_nnapi.py
|
|
|
|
test_numba_integration.py
|
special-case cuda array interface of zero size (#121458)
|
2024-03-18 15:21:38 +00:00 |
|
test_numpy_interop.py
|
Support numpy array in Tensor.__eq__ (#122249)
|
2024-03-21 04:55:01 +00:00 |
|
test_openmp.py
|
|
|
|
test_ops.py
|
Revert "Avoid COW materialize in conv, log sigmoid, repeat, group_norm, batch_norm (#121537)"
|
2024-03-21 00:03:30 +00:00 |
|
test_ops_fwd_gradients.py
|
|
|
|
test_ops_gradients.py
|
|
|
|
test_ops_jit.py
|
|
|
|
test_optim.py
|
[optim][BE] move complex_2d (last of complex tests) to OptimInfo (#120618)
|
2024-03-12 02:33:21 +00:00 |
|
test_out_dtype_op.py
|
[export][reland] Disable exported_program.__call__ (#120019)
|
2024-03-05 05:29:46 +00:00 |
|
test_overrides.py
|
|
|
|
test_package.py
|
|
|
|
test_per_overload_api.py
|
|
|
|
test_prims.py
|
|
|
|
test_proxy_tensor.py
|
Batch Norm Consolidation (#116092)
|
2024-03-18 21:01:30 +00:00 |
|
test_pruning_op.py
|
|
|
|
test_public_bindings.py
|
|
|
|
test_python_dispatch.py
|
Change default torch_function behavior to be disabled when torch_dispatch is defined (take 2) (#120632)
|
2024-03-09 01:08:37 +00:00 |
|
test_pytree.py
|
[pytree][reland] Require pytree serialized_type_name (#120636)
|
2024-02-27 06:53:33 +00:00 |
|
test_quantization.py
|
|
|
|
test_reductions.py
|
relax tol for flaky nansum_out_dtype_cuda_float32 test (#121550)
|
2024-03-14 15:28:45 +00:00 |
|
test_scatter_gather_ops.py
|
|
|
|
test_schema_check.py
|
Change default torch_function behavior to be disabled when torch_dispatch is defined (take 2) (#120632)
|
2024-03-09 01:08:37 +00:00 |
|
test_segment_reductions.py
|
|
|
|
test_serialization.py
|
add complex32 to v3_dtypes (#120388)
|
2024-02-28 02:32:29 +00:00 |
|
test_set_default_mobile_cpu_allocator.py
|
|
|
|
test_shape_ops.py
|
|
|
|
test_show_pickle.py
|
|
|
|
test_sort_and_select.py
|
|
|
|
test_sparse.py
|
[ROCm] enable complex128 in test_addmm_sizes_all_sparse_csr for rocm for trivial (k,n,m) cases (#120504)
|
2024-03-12 07:29:57 +00:00 |
|
test_sparse_csr.py
|
[ROCm] enable complex128 in test_addmm_sizes_all_sparse_csr for rocm for trivial (k,n,m) cases (#120504)
|
2024-03-12 07:29:57 +00:00 |
|
test_sparse_semi_structured.py
|
Revert "[Sparsity] add support for H100 compute capability 9.x (#121768)"
|
2024-03-21 10:42:08 +00:00 |
|
test_spectral_ops.py
|
|
|
|
test_stateless.py
|
|
|
|
test_static_runtime.py
|
[PyTorch] Split StaticModule out of test_static_runtime (#121028)
|
2024-03-05 23:14:07 +00:00 |
|
test_subclass.py
|
Change default torch_function behavior to be disabled when torch_dispatch is defined (take 2) (#120632)
|
2024-03-09 01:08:37 +00:00 |
|
test_sympy_utils.py
|
|
|
|
test_tensor_creation_ops.py
|
Move doc links to point to main (#121823)
|
2024-03-15 19:49:37 +00:00 |
|
test_tensorboard.py
|
[tensorboard] Handle bfloat16 type in add_histogram (#120087)
|
2024-03-05 00:27:21 +00:00 |
|
test_tensorexpr.py
|
|
|
|
test_tensorexpr_pybind.py
|
|
|
|
test_testing.py
|
|
|
|
test_throughput_benchmark.py
|
|
|
|
test_torch.py
|
Enable fx graph cache in torch_test.py when using PYTORCH_TEST_WITH_INDUCTOR=1 (#122010)
|
2024-03-19 02:17:10 +00:00 |
|
test_transformers.py
|
Revert "Add Flash Attention support on ROCM (#121561)"
|
2024-03-19 17:14:28 +00:00 |
|
test_type_hints.py
|
|
|
|
test_type_info.py
|
|
|
|
test_type_promotion.py
|
|
|
|
test_typing.py
|
|
|
|
test_unary_ufuncs.py
|
add decomposition for frexp (#119217)
|
2024-02-23 21:52:42 +00:00 |
|
test_utils.py
|
Use tree_map to get device ids and device types for activation checkpointing (#121462)
|
2024-03-20 21:09:21 +00:00 |
|
test_view_ops.py
|
|
|
|
test_vulkan.py
|
|
|
|
test_weak.py
|
|
|
|
test_xnnpack_integration.py
|
|
|
|
test_xpu.py
|
Revert "Introduce XPU implementation for PyTorch ATen operators (#120891)"
|
2024-03-21 20:30:20 +00:00 |