diff --git a/neural_compressor/transformers/quantization/utils.py b/neural_compressor/transformers/quantization/utils.py index 8dba085a553..6f209344348 100644 --- a/neural_compressor/transformers/quantization/utils.py +++ b/neural_compressor/transformers/quantization/utils.py @@ -353,9 +353,9 @@ def convert_to_quantized_model(model, config, device="cpu"): import intel_extension_for_pytorch assert hasattr(torch, "xpu") and torch.xpu.is_available(), "There is no xpu device in this system!" - os.environ["FORCE_DEVICE"] = "cpu" + os.environ["INC_TARGET_DEVICE"] = "cpu" logger.info( - "Set the environment variable FORCE_DEVICE='cpu' to ensure the quantization process occurs on the CPU." + "Set the environment variable INC_TARGET_DEVICE='cpu' to ensure the quantization process occurs on the CPU." ) orig_dtype = torch.float32