- New instructions in the Intel® Xeon® Scalable processors combined with optimized software frameworks enable real-time AI within network workloads (Feb 2022)
- Quantizing ONNX Models using Intel® Neural Compressor (Feb 2022)
- Quantize AI Model by Intel® oneAPI AI Analytics Toolkit on Alibaba Cloud (Feb 2022)
- AI Performance and Productivity with Intel® Neural Compressor (Jan 2022)
- Ease-of-use quantization for PyTorch with Intel® Neural Compressor (Jan 2022)
- Intel Neural Compressor Tutorial on BiliBili (Dec 2021)
- Faster AI/ML Results With Intel Neural Compressor (Dec 2021)
- Prune Once for All: Sparse Pre-Trained Language Models (Nov 2021)
- Faster, Easier Optimization with Intel® Neural Compressor (Nov 2021)
- Intel® Neural Compressor: A Scalable Quantization Tool for ONNX Models (Oct 2021)
- A "Double Play" for MLPerf™ Inference Performance Gains with 3rd Generation Intel® Xeon® Scalable Processors (Sep 2021)
- Optimize TensorFlow Pre-trained Model for Inference (Jun 2021)
- 3D Digital Face Reconstruction Solution enabled by 3rd Gen Intel® Xeon® Scalable Processors (Apr 2021)
- Accelerating Alibaba Transformer model performance with 3rd Gen Intel® Xeon® Scalable Processors (Ice Lake) and Intel® Deep Learning Boost (Apr 2021)
- MLPerf™ Performance Gains Abound with latest 3rd Generation Intel® Xeon® Scalable Processors (Apr 2021)
- Using Low-Precision Optimizations for High-Performance DL Inference Applications (Apr 2021)
- Quantization support for ONNX using LPOT (Low precision optimization tool) (Mar 2021)
- DL Boost Quantization with CERN's 3D-GANs model (Feb 2021)
- Reduced Precision Strategies for Deep Learning: 3DGAN Use Case - presentation on 4th IML Machine Learning Workshop (Oct 2020)
- Intel Neural Compressor (Sep 2020)