Intel® Neural Compressor v1.10 Release
Features
- Quantization
- Supported the quantization on latest deep learning frameworks
- Supported the quantization for a new model domain (Audio)
- Supported the compatible quantization recipes for framework upgrade
- Pruning & Knowledge distillation
- Supported fine-tuning and quantization using INC & Optimum for “Prune Once for All: Sparse Pre-Trained Language Models” published at ENLSP NeurIPS Workshop 2021
- Structured sparsity
- Proved the sparsity training recipes across multiple model domains (CV, NLP, and Recommendation System)
Productivity
- Improved INC GUI for easy quantization
- Supported Windows OS conda installation
Ecosystem
- Upgraded INC v1.9 into HuggingFace Optimum
- Upsteamed INC quantized mobilenet & faster-rcnn models to ONNX Model Zoo
Examples
- Supported quantization on 300 random models
- Added bare-metal examples for Bert-mini and DLRM
Validated Configurations
- Python 3.7, 3.8, 3.9
- Centos 8.3 & Ubuntu 18.04 & Win10
- TensorFlow 2.6.2, 2.7, 2.8
- Intel TensorFlow 1.15.0 UP3, 2.7, 2.8
- PyTorch 1.8.0+cpu, 1.9.0+cpu, 1.10.0+cpu
- IPEX 1.8.0, 1.9.0, 1.10.0
- MxNet 1.6.0, 1.7.0, 1.8.0
- ONNX Runtime 1.8.0, 1.9.0, 1.10.0
Distribution:
Channel | Links | Install Command | |
---|---|---|---|
Source | Github | https://github.com/intel/neural-compressor.git | $ git clone https://github.com/intel/neural-compressor.git |
Binary | Pip | https://pypi.org/project/neural-compressor | $ pip install neural-compressor |
Binary | Conda | https://anaconda.org/intel/neural-compressor | $ conda install neural-compressor -c conda-forge -c intel |
Contact:
Please feel free to contact [email protected], if you get any questions.