Skip to content

Commit

Permalink
Fix (docs): bullet formatting in getting started guide (Xilinx#952)
Browse files Browse the repository at this point in the history
  • Loading branch information
timkpaine authored May 6, 2024
1 parent 564f64a commit f20a16f
Showing 1 changed file with 1 addition and 0 deletions.
1 change: 1 addition & 0 deletions docsrc/source/getting_started.rst
Original file line number Diff line number Diff line change
Expand Up @@ -186,6 +186,7 @@ Weights, activations, biases quantization
..
Compared to the previous scenario:

- We now set ``return_quant_tensor=True`` in every quantized activations to propagate a ``QuantTensor`` to the next layer. This informs each ``QuantLinear`` or ``QuantConv2d`` of how the input passed in has been quantized.
- A ``QuantTensor`` is just a tensor-like data structure providing metadata about how a tensor has been quantized, similar to a `torch.qint` dtype, but training friendly. Setting ``return_quant_tensor=True`` does not affect the way quantization is performed, it only changes the way the output is represented.
- We enable bias quantization by setting the `Int32Bias` quantizer. What it does is to perform bias quantization with ```bias_scale = input_scale * weight_scale``, as it commonly done across inference toolchains. This is why we have to set ``return_quant_tensor=True``: each layer with ``Int32Bias`` can read the input scale from the ``QuantTensor`` passed in and use for bias quantization.
Expand Down

0 comments on commit f20a16f

Please sign in to comment.