Skip to content

Commit

Permalink
[Doc] Installed version of llmcompressor for int8/fp8 quantization (#…
Browse files Browse the repository at this point in the history
…11103)

Signed-off-by: Guangda Liu <[email protected]>
Co-authored-by: Guangda Liu <[email protected]>
  • Loading branch information
bingps and bingps authored Dec 11, 2024
1 parent b2f7754 commit fd22220
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 3 deletions.
2 changes: 1 addition & 1 deletion docs/source/quantization/fp8.rst
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,7 @@ To produce performant FP8 quantized models with vLLM, you'll need to install the

.. code-block:: console
$ pip install llmcompressor==0.1.0
$ pip install llmcompressor
Quantization Process
--------------------
Expand Down
4 changes: 2 additions & 2 deletions docs/source/quantization/int8.rst
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ To use INT8 quantization with vLLM, you'll need to install the `llm-compressor <

.. code-block:: console
$ pip install llmcompressor==0.1.0
$ pip install llmcompressor
Quantization Process
--------------------
Expand Down Expand Up @@ -142,4 +142,4 @@ Best Practices
Troubleshooting and Support
---------------------------

If you encounter any issues or have feature requests, please open an issue on the ``vllm-project/llm-compressor`` GitHub repository.
If you encounter any issues or have feature requests, please open an issue on the ``vllm-project/llm-compressor`` GitHub repository.

0 comments on commit fd22220

Please sign in to comment.