-
Notifications
You must be signed in to change notification settings - Fork 193
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add to README a command for export of diffusion models with hybrid quantization #1228
Add to README a command for export of diffusion models with hybrid quantization #1228
Conversation
README.md
Outdated
@@ -163,6 +163,9 @@ For more examples check out our [LLM Inference Guide](https://docs.openvino.ai/2 | |||
```sh | |||
#Download and convert to OpenVINO dreamlike-anime-1.0 model | |||
optimum-cli export openvino --model dreamlike-art/dreamlike-anime-1.0 --task stable-diffusion --weight-format fp16 dreamlike_anime_1_0_ov/FP16 | |||
|
|||
#Download, convert to OpenVINO and apply int8 hybrid quantization to dreamlike-anime-1.0 model |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Maybe explicitly state in the comment that this is an alternative, e.g.:
Or use INT8 hybrid quantization to optimize and speed up the model
or
Also you can use INT8 hybrid quantization to further optimize the model and reduce the inference latency
, etc,
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Updated
…antization (openvinotoolkit#1228) Added a command for exporting quantized diffusion models similar to how it is done for LMs.
Added a command for exporting quantized diffusion models similar to how it is done for LMs.