-
Notifications
You must be signed in to change notification settings - Fork 259
Issues: intel/neural-compressor
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
How to set the pruned weight blocks as a same learnable value?
#1361
opened Oct 30, 2023 by
hobbitlzy
Quantized Neural compress model not generating expected results in AMD processor
#1531
opened Jan 10, 2024 by
Bhuvaneswaran-R
AWQ fails on ONNX model when a MatMul node's input is a model input/initializer
#1571
opened Jan 25, 2024 by
jstoecker
PostTrainingQuantConfig(quant_level='auto', device='npu', backend="onnxrt_dml_ep") produces fp32 ops.
#1580
opened Jan 26, 2024 by
kleiti
How to quantify google/vit-base-patch16-224 pytorch_model.bin to int8 type with neural-compressor
#1612
opened Feb 19, 2024 by
yingmuying
neural_compressor/adaptor/ox_utils/quantizer.py dfs crash during "basic" tuning
#1621
opened Feb 22, 2024 by
kmn1024
'q_config' is needed when export an INT8 model
aitce
AI TCE to handle it firstly
#1736
opened Apr 18, 2024 by
ZhangShuoAlreadyExists
how to extract int8 weights from quantized model
aitce
AI TCE to handle it firstly
#1817
opened May 25, 2024 by
chensterliu
Previous Next
ProTip!
What’s not been updated in a month: updated:<2024-12-19.