Skip to content

Commit

Permalink
Update docs/articles_en/openvino-workflow/model-optimization-guide/we…
Browse files Browse the repository at this point in the history
…ight-compression.rst

Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
  • Loading branch information
AlexKoff88 and tsavina authored Sep 10, 2024
1 parent 1ff1040 commit e01cc5a
Showing 1 changed file with 3 additions and 3 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -226,9 +226,9 @@ For data-aware weight compression refer to the following

.. note::

Some of the methods can be stacked one on top of another to achieve a better
accuracy-performance trade-off after weight quantization. For example, Scale Estimation
method can be applied along with AWQ and mixed-precision quantization (``ratio`` parameter).
Some methods can be stacked on top of one another to achieve a better
accuracy-performance trade-off after weight quantization. For example, the Scale Estimation
method can be applied along with AWQ and mixed-precision quantization (the ``ratio`` parameter).

The example below shows data-free 4-bit weight quantization
applied on top of OpenVINO IR. Before trying the example, make sure Optimum Intel
Expand Down

0 comments on commit e01cc5a

Please sign in to comment.