mirror of
https://github.com/saymrwulf/pytorch.git
synced 2026-05-14 20:57:59 +00:00
Fix most documentation warnings (#27782)
Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/27782 Warnings show up when running `make html` to build documentation. All of the warnings are very reasonable and point to bugs in our docs. This PR attempts to fix most of those warnings. In the future we will add something to the CI that asserts that there are no warnings in our docs. Test Plan: - build and view changes locally Differential Revision: D17887067 Pulled By: zou3519 fbshipit-source-id: 6bf4d08764759133b20983d6cd7f5d27e5ee3166
This commit is contained in:
parent
446a79b959
commit
23bffc4f14
17 changed files with 51 additions and 39 deletions
|
|
@ -67,6 +67,8 @@ Tensor autograd functions
|
|||
.. autoclass:: torch.Tensor
|
||||
:members: grad, requires_grad, is_leaf, backward, detach, detach_, register_hook, retain_grad
|
||||
|
||||
:noindex:
|
||||
|
||||
:hidden:`Function`
|
||||
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||
|
||||
|
|
|
|||
|
|
@ -49,7 +49,7 @@ here is the basic process.
|
|||
operator/optimizer?” Giving evidence for its utility, e.g., usage
|
||||
in peer reviewed papers, or existence in other frameworks, helps a
|
||||
bit when making this case.
|
||||
- **Adding operators / algorithms from recently-released research**
|
||||
- **Adding operators / algorithms from recently-released research**
|
||||
is generally not accepted, unless there is overwhelming evidence that
|
||||
this newly published work has ground-breaking results and will eventually
|
||||
become a standard in the field. If you are not sure where your method falls,
|
||||
|
|
|
|||
|
|
@ -63,8 +63,7 @@ Distributed
|
|||
|
||||
- Pieter Noordhuis (`pietern <https://github.com/pietern>`__)
|
||||
- Shen Li (`mrshenli <https://github.com/mrshenli>`__)
|
||||
..
|
||||
- (proposed) Pritam Damania
|
||||
- (proposed) Pritam Damania
|
||||
(`pritamdamania87 <https://github.com/pritamdamania87>`__)
|
||||
|
||||
Multiprocessing and DataLoaders
|
||||
|
|
|
|||
|
|
@ -52,7 +52,9 @@ Memory management
|
|||
.. autofunction:: reset_max_memory_allocated
|
||||
.. autofunction:: memory_reserved
|
||||
.. autofunction:: max_memory_reserved
|
||||
.. autofunction:: reset_max_memory_reserved
|
||||
.. FIXME The following doesn't seem to exist. Is it supposed to?
|
||||
https://github.com/pytorch/pytorch/issues/27785
|
||||
.. autofunction:: reset_max_memory_reserved
|
||||
.. autofunction:: memory_cached
|
||||
.. autofunction:: max_memory_cached
|
||||
.. autofunction:: reset_max_memory_cached
|
||||
|
|
|
|||
|
|
@ -410,7 +410,7 @@ both python2 and python3.
|
|||
Spawn utility
|
||||
-------------
|
||||
|
||||
The :doc:`torch.multiprocessing` package also provides a ``spawn``
|
||||
The :ref:`multiprocessing-doc` package also provides a ``spawn``
|
||||
function in :func:`torch.multiprocessing.spawn`. This helper function
|
||||
can be used to spawn multiple processes. It works by passing in the
|
||||
function that you want to run and spawns N processes to run it. This
|
||||
|
|
|
|||
|
|
@ -16,6 +16,7 @@ PyTorch is an optimized tensor library for deep learning using GPUs and CPUs.
|
|||
:caption: Notes
|
||||
|
||||
notes/*
|
||||
|
||||
* `PyTorch on XLA Devices <http://pytorch.org/xla/>`_
|
||||
|
||||
.. toctree::
|
||||
|
|
|
|||
|
|
@ -1,3 +1,6 @@
|
|||
:orphan:
|
||||
.. _multiprocessing-doc:
|
||||
|
||||
Multiprocessing package - torch.multiprocessing
|
||||
===============================================
|
||||
|
||||
|
|
|
|||
|
|
@ -373,7 +373,7 @@ Top-level quantization APIs
|
|||
.. autofunction:: convert
|
||||
.. autoclass:: QConfig
|
||||
.. autoclass:: QConfigDynamic
|
||||
.. autoattr:: default_qconfig
|
||||
.. autoattribute:: default_qconfig
|
||||
|
||||
Preparing model for quantization
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
|
|
|||
|
|
@ -8,14 +8,13 @@ torch.random
|
|||
|
||||
Random Number Generator
|
||||
-------------------------
|
||||
.. FIXME: We're missing torch.random.cuda docs.
|
||||
https://github.com/pytorch/pytorch/issues/27778
|
||||
|
||||
.. autofunction:: get_rng_state
|
||||
.. autofunction:: get_rng_state_all
|
||||
.. autofunction:: set_rng_state
|
||||
.. autofunction:: set_rng_state_all
|
||||
.. autofunction:: manual_seed
|
||||
.. autofunction:: manual_seed_all
|
||||
.. autofunction:: seed
|
||||
.. autofunction:: seed_all
|
||||
.. autofunction:: initial_seed
|
||||
.. autofunction:: fork_rng
|
||||
|
||||
|
|
|
|||
|
|
@ -148,6 +148,7 @@ view of a storage and defines numeric operations on it.
|
|||
.. autoattribute:: is_cuda
|
||||
.. autoattribute:: device
|
||||
.. autoattribute:: grad
|
||||
:noindex:
|
||||
.. autoattribute:: ndim
|
||||
.. autoattribute:: T
|
||||
|
||||
|
|
@ -183,6 +184,7 @@ view of a storage and defines numeric operations on it.
|
|||
.. automethod:: atan2_
|
||||
.. automethod:: atan_
|
||||
.. automethod:: backward
|
||||
:noindex:
|
||||
.. automethod:: baddbmm
|
||||
.. automethod:: baddbmm_
|
||||
.. automethod:: bernoulli
|
||||
|
|
@ -222,7 +224,9 @@ view of a storage and defines numeric operations on it.
|
|||
.. automethod:: det
|
||||
.. automethod:: dense_dim
|
||||
.. automethod:: detach
|
||||
:noindex:
|
||||
.. automethod:: detach_
|
||||
:noindex:
|
||||
.. automethod:: diag
|
||||
.. automethod:: diag_embed
|
||||
.. automethod:: diagflat
|
||||
|
|
@ -295,12 +299,13 @@ view of a storage and defines numeric operations on it.
|
|||
.. automethod:: irfft
|
||||
.. automethod:: is_contiguous
|
||||
.. automethod:: is_floating_point
|
||||
.. automethod:: is_leaf
|
||||
.. autoattribute:: is_leaf
|
||||
:noindex:
|
||||
.. automethod:: is_pinned
|
||||
.. automethod:: is_set_to
|
||||
.. automethod:: is_shared
|
||||
.. automethod:: is_signed
|
||||
.. automethod:: is_sparse
|
||||
.. autoattribute:: is_sparse
|
||||
.. automethod:: item
|
||||
.. automethod:: kthvalue
|
||||
.. automethod:: le
|
||||
|
|
@ -382,6 +387,7 @@ view of a storage and defines numeric operations on it.
|
|||
.. automethod:: reciprocal_
|
||||
.. automethod:: record_stream
|
||||
.. automethod:: register_hook
|
||||
:noindex:
|
||||
.. automethod:: remainder
|
||||
.. automethod:: remainder_
|
||||
.. automethod:: real
|
||||
|
|
@ -389,13 +395,14 @@ view of a storage and defines numeric operations on it.
|
|||
.. automethod:: renorm_
|
||||
.. automethod:: repeat
|
||||
.. automethod:: repeat_interleave
|
||||
.. automethod:: requires_grad
|
||||
.. autoattribute:: requires_grad
|
||||
.. automethod:: requires_grad_
|
||||
.. automethod:: reshape
|
||||
.. automethod:: reshape_as
|
||||
.. automethod:: resize_
|
||||
.. automethod:: resize_as_
|
||||
.. automethod:: retain_grad
|
||||
:noindex:
|
||||
.. automethod:: rfft
|
||||
.. automethod:: roll
|
||||
.. automethod:: rot90
|
||||
|
|
|
|||
|
|
@ -93,10 +93,13 @@ Random sampling
|
|||
.. autofunction:: set_rng_state
|
||||
.. autoattribute:: torch.default_generator
|
||||
:annotation: Returns the default CPU torch.Generator
|
||||
.. autoattribute:: torch.cuda.default_generators
|
||||
:annotation: If cuda is available, returns a tuple of default CUDA torch.Generator-s.
|
||||
The number of CUDA torch.Generator-s returned is equal to the number of
|
||||
GPUs available in the system.
|
||||
|
||||
.. The following doesn't actually seem to exist.
|
||||
https://github.com/pytorch/pytorch/issues/27780
|
||||
.. autoattribute:: torch.cuda.default_generators
|
||||
:annotation: If cuda is available, returns a tuple of default CUDA torch.Generator-s.
|
||||
The number of CUDA torch.Generator-s returned is equal to the number of
|
||||
GPUs available in the system.
|
||||
.. autofunction:: bernoulli
|
||||
.. autofunction:: multinomial
|
||||
.. autofunction:: normal
|
||||
|
|
@ -151,8 +154,7 @@ The context managers :func:`torch.no_grad`, :func:`torch.enable_grad`, and
|
|||
:func:`torch.set_grad_enabled` are helpful for locally disabling and enabling
|
||||
gradient computation. See :ref:`locally-disable-grad` for more details on
|
||||
their usage. These context managers are thread local, so they won't
|
||||
work if you send work to another thread using the :module:`threading`
|
||||
module, etc.
|
||||
work if you send work to another thread using the ``threading`` module, etc.
|
||||
|
||||
Examples::
|
||||
|
||||
|
|
|
|||
|
|
@ -399,6 +399,7 @@ Example::
|
|||
tensor([[False, True]], dtype=torch.bool)
|
||||
>>> a.any()
|
||||
tensor(True, dtype=torch.bool)
|
||||
|
||||
.. function:: any(dim, keepdim=False, out=None) -> Tensor
|
||||
|
||||
Returns True if any elements in each row of the tensor in the given
|
||||
|
|
|
|||
|
|
@ -6167,7 +6167,7 @@ Calculates determinant of a square matrix or batches of square matrices.
|
|||
:meth:`~torch.svd` for details.
|
||||
|
||||
Arguments:
|
||||
input (Tensor): the input tensor of size (*, n, n) where `*` is zero or more
|
||||
input (Tensor): the input tensor of size ``(*, n, n)`` where ``*`` is zero or more
|
||||
batch dimensions.
|
||||
|
||||
Example::
|
||||
|
|
@ -6254,7 +6254,7 @@ Calculates log determinant of a square matrix or batches of square matrices.
|
|||
:meth:`~torch.svd` for details.
|
||||
|
||||
Arguments:
|
||||
input (Tensor): the input tensor of size (*, n, n) where `*` is zero or more
|
||||
input (Tensor): the input tensor of size ``(*, n, n)`` where ``*`` is zero or more
|
||||
batch dimensions.
|
||||
|
||||
Example::
|
||||
|
|
@ -6295,7 +6295,7 @@ Calculates the sign and log absolute value of the determinant(s) of a square mat
|
|||
See :meth:`~torch.svd` for details.
|
||||
|
||||
Arguments:
|
||||
input (Tensor): the input tensor of size (*, n, n) where `*` is zero or more
|
||||
input (Tensor): the input tensor of size ``(*, n, n)`` where ``*`` is zero or more
|
||||
batch dimensions.
|
||||
|
||||
Returns:
|
||||
|
|
|
|||
|
|
@ -27,13 +27,13 @@ class Stream(torch._C._CudaStreamBase):
|
|||
Arguments:
|
||||
event (Event): an event to wait for.
|
||||
|
||||
.. note:: This is a wrapper around ``cudaStreamWaitEvent()``: see `CUDA
|
||||
documentation`_ for more info.
|
||||
.. note:: This is a wrapper around ``cudaStreamWaitEvent()``: see
|
||||
`CUDA Stream documentation`_ for more info.
|
||||
|
||||
This function returns without waiting for :attr:`event`: only future
|
||||
operations are affected.
|
||||
|
||||
.. _CUDA documentation:
|
||||
.. _CUDA Stream documentation:
|
||||
http://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__STREAM.html
|
||||
"""
|
||||
event.wait(self)
|
||||
|
|
@ -78,10 +78,7 @@ class Stream(torch._C._CudaStreamBase):
|
|||
r"""Wait for all the kernels in this stream to complete.
|
||||
|
||||
.. note:: This is a wrapper around ``cudaStreamSynchronize()``: see
|
||||
`CUDA documentation`_ for more info.
|
||||
|
||||
.. _CUDA documentation:
|
||||
http://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__STREAM.html
|
||||
`CUDA Stream documentation`_ for more info.
|
||||
"""
|
||||
super(Stream, self).synchronize()
|
||||
|
||||
|
|
@ -121,8 +118,8 @@ class Event(torch._C._CudaEventBase):
|
|||
interprocess (bool): if ``True``, the event can be shared between processes
|
||||
(default: ``False``)
|
||||
|
||||
.. _CUDA documentation:
|
||||
https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__EVENT.html
|
||||
.. _CUDA Event Documentation:
|
||||
https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__EVENT.html
|
||||
"""
|
||||
|
||||
def __new__(cls, enable_timing=False, blocking=False, interprocess=False):
|
||||
|
|
@ -174,11 +171,8 @@ class Event(torch._C._CudaEventBase):
|
|||
Waits until the completion of all work currently captured in this event.
|
||||
This prevents the CPU thread from proceeding until the event completes.
|
||||
|
||||
.. note:: This is a wrapper around ``cudaEventSynchronize()``: see `CUDA
|
||||
documentation`_ for more info.
|
||||
|
||||
.. _CUDA documentation:
|
||||
https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__EVENT.html
|
||||
.. note:: This is a wrapper around ``cudaEventSynchronize()``: see
|
||||
`CUDA Event documentation`_ for more info.
|
||||
"""
|
||||
super(Event, self).synchronize()
|
||||
|
||||
|
|
|
|||
|
|
@ -47,6 +47,7 @@ class LowRankMultivariateNormal(Distribution):
|
|||
r"""
|
||||
Creates a multivariate normal distribution with covariance matrix having a low-rank form
|
||||
parameterized by :attr:`cov_factor` and :attr:`cov_diag`::
|
||||
|
||||
covariance_matrix = cov_factor @ cov_factor.T + cov_diag
|
||||
|
||||
Example:
|
||||
|
|
@ -69,6 +70,7 @@ class LowRankMultivariateNormal(Distribution):
|
|||
`matrix determinant lemma <https://en.wikipedia.org/wiki/Matrix_determinant_lemma>`_.
|
||||
Thanks to these formulas, we just need to compute the determinant and inverse of
|
||||
the small size "capacitance" matrix::
|
||||
|
||||
capacitance = I + cov_factor.T @ inv(cov_diag) @ cov_factor
|
||||
"""
|
||||
arg_constraints = {"loc": constraints.real,
|
||||
|
|
|
|||
|
|
@ -13,7 +13,7 @@ class FloatFunctional(torch.nn.Module):
|
|||
This class does not provide a ``forward`` hook. Instead, you must use
|
||||
one of the underlying functions (e.g. ``add``).
|
||||
|
||||
.. Examples::
|
||||
Examples::
|
||||
|
||||
>>> f_add = FloatFunctional()
|
||||
>>> a = torch.tensor(3.0)
|
||||
|
|
@ -91,7 +91,7 @@ class QFunctional(torch.nn.Module):
|
|||
This class does not provide a ``forward`` hook. Instead, you must use
|
||||
one of the underlying functions (e.g. ``add``).
|
||||
|
||||
.. Examples::
|
||||
Examples::
|
||||
|
||||
>>> q_add = QFunctional('add')
|
||||
>>> a = torch.quantize_per_tensor(torch.tensor(3.0), 1.0, 0, torch.qint32)
|
||||
|
|
|
|||
|
|
@ -639,7 +639,7 @@ class RecordingObserver(_ObserverBase):
|
|||
class NoopObserver(Observer):
|
||||
r"""
|
||||
Observer that doesn't do anything and just passes its configuration to the
|
||||
quantized module's ``.from_float()`.
|
||||
quantized module's ``.from_float()``.
|
||||
|
||||
Primarily used for quantization to float16 which doesn't require determining
|
||||
ranges.
|
||||
|
|
|
|||
Loading…
Reference in a new issue