pytorch/torch
Xiaodong Wang e7cf7d00be Support torch.bool in torch.sort + CUDA (#139409)
Summary: This might be out-dated, so I'm adding it back and see if we pass all the tests. I'm pretty sure cuda12 is ok.

Test Plan: CI

Differential Revision: D65282650

Pull Request resolved: https://github.com/pytorch/pytorch/pull/139409
Approved by: https://github.com/zou3519, https://github.com/ngimel, https://github.com/eqy
2024-11-06 00:02:54 +00:00
..
_awaits
_C [Inductor][CPP] Add oneDNN BRGEMM config for Half cpp gemm template (#136255) 2024-11-05 05:33:29 +00:00
_C_flatbuffer
_custom_op
_decomp [Inductor][CPU] Enable the oneDNN Linear fusion for special case (#139172) 2024-11-05 23:49:53 +00:00
_dispatch
_dynamo Improvements for associative_scan - slicing of xs (#138858) 2024-11-05 23:38:21 +00:00
_export [export] Add support for symbool to make it usable for torch.cond (#138765) 2024-11-04 23:31:49 +00:00
_functorch Revert "Deprecate torch._utils.is_compiling() and torch._dynamo.external_utils.is_compiling() (#127690)" 2024-11-05 23:10:38 +00:00
_higher_order_ops Improvements for associative_scan - slicing of xs (#138858) 2024-11-05 23:38:21 +00:00
_inductor typing ir.py - part 2 (#131846) 2024-11-06 00:01:15 +00:00
_lazy
_library Optimize mutable torch.library.custom_op overhead (#139513) 2024-11-05 18:30:53 +00:00
_logging Profile guided optimization for automatic_dynamic (#139001) 2024-11-03 06:29:57 +00:00
_numpy
_prims use torch.sym_sum instead of incremental sum in _cat_meta (#139653) 2024-11-05 07:24:24 +00:00
_prims_common check fake/real mismatches during real tensor prop (#137747) 2024-11-04 23:39:48 +00:00
_refs Revert "Fix unbind_copy and add its decomposition (#134319)" 2024-10-29 04:54:37 +00:00
_strobelight Increase default COMPILE_STROBELIGHT_MAX_STACK_LENGTH to 500 (#138006) 2024-10-17 07:31:32 +00:00
_subclasses check fake/real mismatches during real tensor prop (#137747) 2024-11-04 23:39:48 +00:00
_vendor
accelerator Introduce a device-agnostic runtime API design (#132204) 2024-10-27 10:37:09 +00:00
amp [MPS] Update error message for supported autocast type (#139192) 2024-10-30 16:48:29 +00:00
ao Revert "Tighten type hints for tensor arithmetic (#135392)" 2024-11-04 23:30:15 +00:00
autograd [Profiler] Create Auto-Trace Frontend for Trace ID (#139310) 2024-10-31 19:02:57 +00:00
backends Revert "[sparse] add search for optimal alg_id to torch.compile (#137427)" 2024-10-24 17:27:06 +00:00
compiler Profile guided optimization for automatic_dynamic (#139001) 2024-11-03 06:29:57 +00:00
contrib Remove unused Python variables in torch/[b-z]* (#136963) 2024-10-19 16:45:22 +00:00
cpu [Inductor][CPP] Add oneDNN BRGEMM config for Half cpp gemm template (#136255) 2024-11-05 05:33:29 +00:00
csrc [pgnccl][simple] log started work numel (#139773) 2024-11-05 23:11:19 +00:00
cuda Adds snapshot API for MemPools to get pool memory segments (#133601) 2024-10-29 01:01:47 +00:00
distributed Revert "Deprecate torch._utils.is_compiling() and torch._dynamo.external_utils.is_compiling() (#127690)" 2024-11-05 23:10:38 +00:00
distributions Remove unused Python variables in torch/[b-z]* (#136963) 2024-10-19 16:45:22 +00:00
export [export] Dedup data-dependent errors based on stacktrace (#139540) 2024-11-05 18:16:05 +00:00
fft
func
futures
fx check fake/real mismatches during real tensor prop (#137747) 2024-11-04 23:39:48 +00:00
jit Remove unused Python variables in torch/[b-z]* (#136963) 2024-10-19 16:45:22 +00:00
legacy
lib
linalg
masked [BE]: Update Typeguard to TypeIs for better type inference (#133814) 2024-10-26 15:07:13 +00:00
monitor
mps
mtia
multiprocessing Remove unused Python variables in torch/[b-z]* (#136963) 2024-10-19 16:45:22 +00:00
nested Fix and test several NJT reductions (#139317) 2024-10-31 20:55:38 +00:00
nn Revert "Deprecate torch._utils.is_compiling() and torch._dynamo.external_utils.is_compiling() (#127690)" 2024-11-05 23:10:38 +00:00
onnx [ONNX] Remove type promotion rule for pow (#139527) 2024-11-02 02:19:50 +00:00
optim Revert "Deprecate torch._utils.is_compiling() and torch._dynamo.external_utils.is_compiling() (#127690)" 2024-11-05 23:10:38 +00:00
package Remove unused Python variables in torch/[b-z]* (#136963) 2024-10-19 16:45:22 +00:00
profiler [fx graph cache] Support freezing with FX graph caching (#136505) 2024-11-01 18:29:29 +00:00
quantization Remove unused Python variables in torch/[b-z]* (#136963) 2024-10-19 16:45:22 +00:00
signal
sparse Revert "[sparse] add search for optimal alg_id to torch.compile (#137427)" 2024-10-24 17:27:06 +00:00
special
testing Support torch.bool in torch.sort + CUDA (#139409) 2024-11-06 00:02:54 +00:00
utils Fix existing lint issues in ir.py (#139237) 2024-11-05 06:06:12 +00:00
xpu Add torch.xpu.get_arch_list and torch.xpu.get_gencode_flags for XPU (#137773) 2024-10-18 02:28:08 +00:00
__config__.py
__future__.py
__init__.py add hex method on SymFloat (#139451) 2024-11-02 05:33:19 +00:00
_appdirs.py
_classes.py
_compile.py
_custom_ops.py
_deploy.py
_environment.py
_guards.py [hierarchical-compilation][invoke_subgraph] Use tracing context to cache artifacts of dispatch keys (#137965) 2024-10-22 15:33:42 +00:00
_jit_internal.py
_linalg_utils.py
_lobpcg.py
_lowrank.py
_meta_registrations.py check fake/real mismatches during real tensor prop (#137747) 2024-11-04 23:39:48 +00:00
_namedtensor_internals.py
_ops.py remove redundant a (#139046) 2024-10-28 17:47:24 +00:00
_python_dispatcher.py
_size_docs.py
_sources.py
_storage_docs.py
_streambase.py
_tensor.py Remove numpy dependency for maia serialization (#137600) 2024-10-28 20:57:35 +00:00
_tensor_docs.py
_tensor_str.py
_thread_safe_fork.py
_torch_docs.py make equation behind torch.isclose element-wise (#138459) 2024-11-01 18:18:33 +00:00
_utils.py Revert "Deprecate torch._utils.is_compiling() and torch._dynamo.external_utils.is_compiling() (#127690)" 2024-11-05 23:10:38 +00:00
_utils_internal.py Profile guided optimization for automatic_dynamic (#139001) 2024-11-03 06:29:57 +00:00
_VF.py
_vmap_internals.py
_weights_only_unpickler.py [BE] Change _marked_safe_globals_list to set (#139303) 2024-11-04 23:50:55 +00:00
abi-check.cpp
CMakeLists.txt Add torch.xpu.get_arch_list and torch.xpu.get_gencode_flags for XPU (#137773) 2024-10-18 02:28:08 +00:00
custom_class.h Remove some pre-cpp17 stuff (#138410) 2024-10-23 00:38:03 +00:00
custom_class_detail.h Remove some pre-cpp17 stuff (#138410) 2024-10-23 00:38:03 +00:00
extension.h
functional.py
hub.py Remove unused Python variables in torch/[b-z]* (#136963) 2024-10-19 16:45:22 +00:00
library.h [1/N] Enable cppcoreguidelines-special-member-functions (#137405) 2024-10-23 00:16:53 +00:00
library.py no-op torch.library.custom_op APIs on torch.deploy (#139509) 2024-11-04 18:01:08 +00:00
overrides.py Add Weighted Loss Functions to PyTorch : WMSE, WMAE, and Weighted Huber Loss (#132049) 2024-10-31 21:59:43 +00:00
py.typed
quasirandom.py
random.py [Torch] Support meta device in random.fork_rng (#137715) 2024-10-16 18:00:39 +00:00
README.txt
return_types.py
script.h
serialization.py Forward fix D65441551 for T206731737 (#139767) 2024-11-05 23:19:08 +00:00
storage.py Fix .to(cpu) for Storage (#138011) 2024-10-23 01:31:48 +00:00
torch_version.py
types.py
version.py.tpl

Note [TH abstraction violation]
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~

TH/THC provide some hpp headers, which are proper C++ headers rather than
C headers.  These headers serve double duty as *internal implementation
detail* headers, whose contents should largely not be used by external
clients.

Ideally, we would not install these headers at all; instead, you should
use public functions (in headers like `THTensor.h`, NOT `THTensor.hpp`)
to manipulate these structs.  However, there are a few places
in torch/csrc where we violate this abstraction.  They are marked with
a pointer to this note.  Each of those sites will have to be refactored
when we refactor the guts of THTensor and related structures.