- Blog by Intel: Intel® AMX Enhances AI Inference Performance (Jan 2023)
- Blog by TensorFlow: Optimizing TensorFlow for 4th Gen Intel Xeon Processors (Jan 2023)
- Blog on Medium: MLefficiency — Optimizing transformer models for efficiency (Dec 2022)
- Blog on Medium: One-Click Acceleration of Hugging Face Transformers with Intel’s Neural Coder (Dec 2022)
- Blog on Medium: One-Click Quantization of Deep Learning Models with the Neural Coder Extension (Dec 2022)
- Blog on Medium: Accelerate Stable Diffusion with Intel Neural Compressor (Dec 2022)
- Blog on WeChat: Intel together with Tencent deepens the cooperation to build a cloud foundation for digital and intelligent industry (Dec 2022)
- Blog on VMware: Intel Neural Compressor for TF Virtual Appliance packaged by Bitnami (Nov 2022)
- Blog on Tencent Cloud: Neural Compressor: an open-source Python library for network compression (Nov 2022)
- Blog on Medium: Running Fast Transformers on CPUs: Intel Approach Achieves Significant Speed Ups and SOTA Performance (Nov 2022)
- Blog on Medium: Personalized Stable Diffusion with Few-Shot Fine-Tuning (Nov 2022)
- NeurIPS'2022: Fast Distilbert on CPUs (Oct 2022)
- NeurIPS'2022: QuaLA-MiniLM: a Quantized Length Adaptive MiniLM (Oct 2022)
- Blog by Intel: Meet the Innovation of Intel AI Software: Intel® Extension for TensorFlow* (Oct 2022)
- Blog by Intel: PyTorch* Inference Acceleration with Intel® Neural Compressor (Oct 2022)
- Post on Social Media: Neural Coder, a new plug-in for Intel Neural Compressor was covered by Twitter, LinkedIn, and Intel Developer Zone from Intel, and Twitter and LinkedIn from Hugging Face. (Oct 2022)
- Marketplace Distribute: Intel Neural Compressor successfully landed on GCP, AWS, and Azure marketplace. (Oct 2022)
- Post on Social Media: Neural Coder (Intel Neural Compressor Plug-in): One-Click, No-Code Solution (Pat's Keynote IntelON 2022) (Sep 2022)
- Blog on Medium: Alibaba Cloud and Intel Neural Compressor Deliver Better Productivity for PyTorch Users [Chinese version] (Sep 2022)
- Blog on Medium: Efficient Text Classification with Intel Neural Compressor (Sep 2022)
- Blog on Medium: Dynamic Neural Architecture Search with Intel Neural Compressor (Sep 2022)
- Blog on Medium: Easy Quantization in PyTorch Using Fine-Grained FX (Sep 2022)
- Blog on Medium: One-Click Enabling of Intel Neural Compressor Features in PyTorch Scripts (Aug 2022)
- Blog by Alibaba: Deep learning inference optimization for Address Purification (Aug 2022)
- Blog by Intel: Accelerate AI Inference without Sacrificing Accuracy (Jun 2022)
- Blog by Meta: PyTorch Inference Acceleration with Intel® Neural Compressor (Jun 2022)
- Blog by Hugging Face: Intel and Hugging Face Partner to Democratize Machine Learning Hardware Acceleration (Jun 2022)
- Blog by Intel: Intel® Neural Compressor oneAPI (Jun 2022)
- Blog by Intel: Intel® Deep Learning Boost - Boost Network Security AI Inference Performance in Google Cloud Platform (GCP) (Apr 2022)
- PyTorch Ecosystem: INC as PT ecosystem project (Apr 2022)
- Blog by Intel: New instructions in the Intel® Xeon® Scalable processors combined with optimized software frameworks enable real-time AI within network workloads (Feb 2022)
- Joint blog with MSFT: Quantizing ONNX Models using Intel® Neural Compressor (Feb 2022)
- Blog by Intel: Quantize AI Model by Intel® oneAPI AI Analytics Toolkit on Alibaba Cloud (Feb 2022)
- Blog by SigOpt: Intel Neural Compressor Quantization with SigOpt (Jan 2022)
- Post on Social Media: AI Performance and Productivity with Intel® Neural Compressor (Jan 2022)
- PyTorch Ecosystem: Ease-of-use quantization for PyTorch with Intel® Neural Compressor (Jan 2022)
- Tutorial on BiliBili: Intel Neural Compressor Tutorial on BiliBili (Dec 2021)
- Blog on GESTALT IT: Faster AI/ML Results With Intel Neural Compressor (Dec 2021)
- AI Submit’21: Dynamic Quantization with Intel Neural Compressor and Transformers (Nov 2021)
- NeurIPS’21: Prune Once for All: Sparse Pre-Trained Language Models (Nov 2021)
- Blog by Intel: Faster, Easier Optimization with Intel® Neural Compressor (Nov 2021)
- Blog by Intel: Accelerate Deep Learning with Intel® Extension for TensorFlow* (Oct 2021)
- ONNX Ecosystem: Intel® Neural Compressor: A Scalable Quantization Tool for ONNX Models (Oct 2021)
- Blog by Intel: A "Double Play" for MLPerf™ Inference Performance Gains with 3rd Generation Intel® Xeon® Scalable Processors (Sep 2021)
- Blog by Intel: Optimize TensorFlow Pre-trained Model for Inference (Jun 2021)
- Blog by Intel: 3D Digital Face Reconstruction Solution enabled by 3rd Gen Intel® Xeon® Scalable Processors (Apr 2021)
- Blog by Intel: Accelerating Alibaba Transformer model performance with 3rd Gen Intel® Xeon® Scalable Processors (Ice Lake) and Intel® Deep Learning Boost (Apr 2021)
- Blog by Intel: MLPerf™ Performance Gains Abound with latest 3rd Generation Intel® Xeon® Scalable Processors (Apr 2021)
- Blog by Intel: Using Low-Precision Optimizations for High-Performance DL Inference Applications (Apr 2021)
- ONNX Ecosystem: Quantization support for ONNX using LPOT (Low precision optimization tool) (Mar 2021)
- Blog on NextPlatform:DL Boost Quantization with CERN's 3D-GANs model (Feb 2021)
- Joint presentation with CERN: Reduced Precision Strategies for Deep Learning: 3DGAN Use Case - presentation on 4th IML Machine Learning Workshop (Oct 2020)
- Blog by Intel: Intel Neural Compressor (Sep 2020)
- Blog by Intel: Lower Numerical Precision Deep Learning Inference and Training (May 2018)
- ASPLOS’18: Highly Efficient 8-bit Low Precision Inference of Convolutional Neural Networks with IntelCaffe (May 2018)