mirror of
https://github.com/zebrajr/pytorch.git
synced 2025-12-07 00:21:07 +01:00
Summary:
xref gh-38010 and gh-38011.
After this PR, there should be only two warnings:
```
pytorch/docs/source/index.rst:65: WARNING: toctree contains reference to nonexisting \
document 'torchvision/index'
WARNING: autodoc: failed to import class 'tensorboard.writer.SummaryWriter' from module \
'torch.utils'; the following exception was raised:
No module named 'tensorboard'
```
If tensorboard and torchvision are prerequisites to building docs, they should be added to the `requirements.txt`.
As for breaking up quantization into smaller pieces: I split out the list of supported operations and the list of modules to separate documents. I think this makes the page flow better, makes it much "lighter" in terms of page cost, and also removes some warnings since the same class names appear in multiple sub-modules.
Pull Request resolved: https://github.com/pytorch/pytorch/pull/41321
Reviewed By: ngimel
Differential Revision: D22753099
Pulled By: mruberry
fbshipit-source-id: d504787fcf1104a0b6e3d1c12747ec53450841da
69 lines
1.8 KiB
ReStructuredText
69 lines
1.8 KiB
ReStructuredText
.. _torch_quantization:
|
|
|
|
torch.quantization
|
|
------------------
|
|
.. automodule:: torch.quantization
|
|
|
|
This module implements the functions you call
|
|
directly to convert your model from FP32 to quantized form. For
|
|
example the :func:`~torch.quantization.prepare` is used in post training
|
|
quantization to prepares your model for the calibration step and
|
|
:func:`~torch.quantization.convert` actually converts the weights to int8 and
|
|
replaces the operations with their quantized counterparts. There are
|
|
other helper functions for things like quantizing the input to your
|
|
model and performing critical fusions like conv+relu.
|
|
|
|
Top-level quantization APIs
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
.. autofunction:: quantize
|
|
.. autofunction:: quantize_dynamic
|
|
.. autofunction:: quantize_qat
|
|
.. autofunction:: prepare
|
|
.. autofunction:: prepare_qat
|
|
.. autofunction:: convert
|
|
.. autoclass:: QConfig
|
|
.. autoclass:: QConfigDynamic
|
|
|
|
.. FIXME: The following doesn't display correctly.
|
|
.. autoattribute:: default_qconfig
|
|
|
|
Preparing model for quantization
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
.. autofunction:: fuse_modules
|
|
.. autoclass:: QuantStub
|
|
.. autoclass:: DeQuantStub
|
|
.. autoclass:: QuantWrapper
|
|
.. autofunction:: add_quant_dequant
|
|
|
|
Utility functions
|
|
~~~~~~~~~~~~~~~~~
|
|
.. autofunction:: add_observer_
|
|
.. autofunction:: swap_module
|
|
.. autofunction:: propagate_qconfig_
|
|
.. autofunction:: default_eval_fn
|
|
|
|
Observers
|
|
~~~~~~~~~~~~~~~
|
|
.. autoclass:: ObserverBase
|
|
:members:
|
|
.. autoclass:: MinMaxObserver
|
|
.. autoclass:: MovingAverageMinMaxObserver
|
|
.. autoclass:: PerChannelMinMaxObserver
|
|
.. autoclass:: MovingAveragePerChannelMinMaxObserver
|
|
.. autoclass:: HistogramObserver
|
|
.. autoclass:: FakeQuantize
|
|
.. autoclass:: NoopObserver
|
|
|
|
Debugging utilities
|
|
~~~~~~~~~~~~~~~~~~~
|
|
.. autofunction:: get_observer_dict
|
|
.. autoclass:: RecordingObserver
|
|
|
|
.. currentmodule:: torch
|
|
|
|
.. autosummary::
|
|
:nosignatures:
|
|
|
|
nn.intrinsic
|
|
|