- Blog on Medium: Quantization Accuracy Loss Diagnosis with Neural Insights (Aug 2023)
- Blog on Medium: Faster Stable Diffusion Inference with Intel Extension for Transformers (July 2023)
- Post on Social Media: ONNXCommunityMeetup2023: INT8 Quantization for Large Language Models with Intel Neural Compressor (July 2023)
- Blog by Intel: Accelerate Llama 2 with Intel AI Hardware and Software Optimizations (July 2023)
- Blog on Medium: Model quantization diagnosis with Neural Insights (July 2023)
- Blog on Medium: Simplify Your Custom Chatbot Deployment (June 2023)
- Blog by MSFT: Olive: A user-friendly toolchain for hardware-aware model optimization (June 2023)
- Blog by MSFT: Automate optimization techniques for transformer models (June 2023)
- Post on Social Media: Get Started Post-Training Dynamic Quantization | AI Model Optimization with Intel® Neural Compressor (June 2023)
- Post on Social Media: How to Choose AI Model Quantization Techniques | AI Model Optimization with Intel® Neural Compressor (June 2023)
- Post on Social Media: What is AI Model Optimization | AI Model Optimization with Intel® Neural Compressor | Intel Software (June 2023)
- Blog on Medium: Streamlining Model Optimization as a Service with Intel Neural Compressor (June 2023)
- Blog on Medium: Intel Optimization at Netflix (May 2023)
- Blog on Medium: Effective Post-training Quantization for Large Language Models with Enhanced SmoothQuant Approach (Apr 2023)
- Blog by Intel: Intel® Xeon® Processors Are Still the Only CPU With MLPerf Results, Raising the Bar By 5x (Apr 2023)
- Post on Social Media: Adopt with Tencent TACO: Heterogeneous optimization is also key to improving AI computing power (Mar 2023)
- Blog on Medium: Structured Pruning for Transformer-Based Models (Jan 2023)
- Post on Social Media: Training and Inference for Stable Diffusion | Intel Business (Jan 2023)
- Blog by Intel: Intel® AMX Enhances AI Inference Performance (Jan 2023)
- Blog by TensorFlow: Optimizing TensorFlow for 4th Gen Intel Xeon Processors (Jan 2023)
- Blog on Medium: From Innovation to Ecosystem: A Journey of Intel Neural Compressor (Dec 2022)
- Blog on Medium: MLefficiency — Optimizing transformer models for efficiency (Dec 2022)
- Blog on Medium: One-Click Acceleration of Hugging Face Transformers with Intel’s Neural Coder (Dec 2022)
- Blog on Medium: One-Click Quantization of Deep Learning Models with the Neural Coder Extension (Dec 2022)
- Blog on Medium: Accelerate Stable Diffusion with Intel Neural Compressor (Dec 2022)
- Blog on WeChat: Intel together with Tencent deepens the cooperation to build a cloud foundation for digital and intelligent industry (Dec 2022)
- Blog on VMware: Intel Neural Compressor for TF Virtual Appliance packaged by Bitnami (Nov 2022)
- Blog on Tencent Cloud: Neural Compressor: an open-source Python library for network compression (Nov 2022)
- Blog on Medium: Running Fast Transformers on CPUs: Intel Approach Achieves Significant Speed Ups and SOTA Performance (Nov 2022)
- Blog on Medium: Personalized Stable Diffusion with Few-Shot Fine-Tuning (Nov 2022)
- NeurIPS'2022: Fast Distilbert on CPUs (Oct 2022)
- NeurIPS'2022: QuaLA-MiniLM: a Quantized Length Adaptive MiniLM (Oct 2022)
- Blog by Intel: Meet the Innovation of Intel AI Software: Intel® Extension for TensorFlow* (Oct 2022)
- Blog by Intel: PyTorch* Inference Acceleration with Intel® Neural Compressor (Oct 2022)
- Post on Social Media: Neural Coder, a new plug-in for Intel Neural Compressor was covered by Twitter, LinkedIn, and Intel Developer Zone from Intel, and Twitter and LinkedIn from Hugging Face. (Oct 2022)
- Marketplace Distribute: Intel Neural Compressor successfully landed on GCP, AWS, and Azure marketplace. (Oct 2022)
- Post on Social Media: Neural Coder (Intel Neural Compressor Plug-in): One-Click, No-Code Solution (Pat's Keynote IntelON 2022) (Sep 2022)
- Blog on Medium: Alibaba Cloud and Intel Neural Compressor Deliver Better Productivity for PyTorch Users [Chinese version] (Sep 2022)
- Blog on Medium: Efficient Text Classification with Intel Neural Compressor (Sep 2022)
- Blog on Medium: Dynamic Neural Architecture Search with Intel Neural Compressor (Sep 2022)
- Blog on Medium: Easy Quantization in PyTorch Using Fine-Grained FX (Sep 2022)
- Blog on Medium: One-Click Enabling of Intel Neural Compressor Features in PyTorch Scripts (Aug 2022)
- Blog by Alibaba: Deep learning inference optimization for Address Purification (Aug 2022)
- Blog by Intel: Accelerate AI Inference without Sacrificing Accuracy (Jun 2022)
- Blog by Meta: PyTorch Inference Acceleration with Intel® Neural Compressor (Jun 2022)
- Blog by Hugging Face: Intel and Hugging Face Partner to Democratize Machine Learning Hardware Acceleration (Jun 2022)
- Blog by Intel: Intel® Neural Compressor oneAPI (Jun 2022)
- Blog by Intel: Intel® Deep Learning Boost - Boost Network Security AI Inference Performance in Google Cloud Platform (GCP) (Apr 2022)
- PyTorch Ecosystem: INC as PT ecosystem project (Apr 2022)
- Blog by Intel: New instructions in the Intel® Xeon® Scalable processors combined with optimized software frameworks enable real-time AI within network workloads (Feb 2022)
- Joint blog with MSFT: Quantizing ONNX Models using Intel® Neural Compressor (Feb 2022)
- Blog by Intel: Quantize AI Model by Intel® oneAPI AI Analytics Toolkit on Alibaba Cloud (Feb 2022)
- Blog by SigOpt: Intel Neural Compressor Quantization with SigOpt (Jan 2022)
- Post on Social Media: AI Performance and Productivity with Intel® Neural Compressor (Jan 2022)
- PyTorch Ecosystem: Ease-of-use quantization for PyTorch with Intel® Neural Compressor (Jan 2022)
- Tutorial on BiliBili: Intel Neural Compressor Tutorial on BiliBili (Dec 2021)
- Blog on GESTALT IT: Faster AI/ML Results With Intel Neural Compressor (Dec 2021)
- AI Submit’21: Dynamic Quantization with Intel Neural Compressor and Transformers (Nov 2021)
- NeurIPS’21: Prune Once for All: Sparse Pre-Trained Language Models (Nov 2021)
- Blog by Intel: Faster, Easier Optimization with Intel® Neural Compressor (Nov 2021)
- Blog by Intel: Accelerate Deep Learning with Intel® Extension for TensorFlow* (Oct 2021)
- ONNX Ecosystem: Intel® Neural Compressor: A Scalable Quantization Tool for ONNX Models (Oct 2021)
- Blog by Intel: A "Double Play" for MLPerf™ Inference Performance Gains with 3rd Generation Intel® Xeon® Scalable Processors (Sep 2021)
- Blog by Intel: Optimize TensorFlow Pre-trained Model for Inference (Jun 2021)
- Blog by Intel: 3D Digital Face Reconstruction Solution enabled by 3rd Gen Intel® Xeon® Scalable Processors (Apr 2021)
- Blog by Intel: Accelerating Alibaba Transformer model performance with 3rd Gen Intel® Xeon® Scalable Processors (Ice Lake) and Intel® Deep Learning Boost (Apr 2021)
- Blog by Intel: MLPerf™ Performance Gains Abound with latest 3rd Generation Intel® Xeon® Scalable Processors (Apr 2021)
- Blog by Intel: Using Low-Precision Optimizations for High-Performance DL Inference Applications (Apr 2021)
- ONNX Ecosystem: Quantization support for ONNX using LPOT (Low precision optimization tool) (Mar 2021)
- Blog on NextPlatform:DL Boost Quantization with CERN's 3D-GANs model (Feb 2021)
- Joint presentation with CERN: Reduced Precision Strategies for Deep Learning: 3DGAN Use Case - presentation on 4th IML Machine Learning Workshop (Oct 2020)
- Blog by Intel: Intel Neural Compressor (Sep 2020)
- Blog by Intel: Lower Numerical Precision Deep Learning Inference and Training (May 2018)
- ASPLOS’18: Highly Efficient 8-bit Low Precision Inference of Convolutional Neural Networks with IntelCaffe (May 2018)