diff --git a/docs/source/mixed_precision.md b/docs/source/mixed_precision.md index fa134e6c0b9..eab37d62504 100644 --- a/docs/source/mixed_precision.md +++ b/docs/source/mixed_precision.md @@ -20,14 +20,99 @@ The recently launched 3rd Gen Intel® Xeon® Scalable processor (codenamed Coope ## Mixed Precision Support Matrix -|Framework |BF16 |FP16 | -|--------------|:-----------:|:-----------:| -|TensorFlow |✔ |:x: | -|PyTorch |✔ |:x: | -|ONNX Runtime |✔ |✔ | -|MXNet |✔ |:x: | - -> **During quantization, BF16 conversion is default enabled, FP16 can be executed if 'device' of config is 'gpu'. Please refer to this [document](./quantization_mixed_precision.md) for its workflow.** +
Framework | +Backend | +Backend Library | +Backend Value | +Support Device(cpu as default) | +Support BF16 | +Support FP16 | +
---|---|---|---|---|---|---|
PyTorch | +FX | +FBGEMM | +"default" | +cpu | +✔ | +:x: | +
IPEX | +OneDNN | +"ipex" | +cpu | +✔ | +:x: | +|
ONNX Runtime | +CPUExecutionProvider | +MLAS | +"default" | +cpu | +:x: | +:x: | +
TensorrtExecutionProvider | +TensorRT | +"onnxrt_trt_ep" | +gpu | +:x: | +:x: | +|
CUDAExecutionProvider | +CUDA | +"onnxrt_cuda_ep" | +gpu | +✔ | +✔ | +|
DnnlExecutionProvider | +OneDNN | +"onnxrt_dnnl_ep" | +cpu | +✔ | +:x: | +|
Tensorflow | +Tensorflow | +OneDNN | +"default" | +cpu | +✔ | +:x: | +
ITEX | +OneDNN | +"itex" | +cpu | gpu | +✔ | +:x: | +|
MXNet | +OneDNN | +OneDNN | +"default" | +cpu | +✔ | +:x: | +