Skip to content

Intel® Neural Compressor v1.10 Release

Compare
Choose a tag to compare
@ftian1 ftian1 released this 28 Feb 05:27
1eb6529

Features

  • Quantization
    • Supported the quantization on latest deep learning frameworks
    • Supported the quantization for a new model domain (Audio)
    • Supported the compatible quantization recipes for framework upgrade
  • Pruning & Knowledge distillation
    • Supported fine-tuning and quantization using INC & Optimum for “Prune Once for All: Sparse Pre-Trained Language Models” published at ENLSP NeurIPS Workshop 2021
  • Structured sparsity
    • Proved the sparsity training recipes across multiple model domains (CV, NLP, and Recommendation System)

Productivity

  • Improved INC GUI for easy quantization
  • Supported Windows OS conda installation

Ecosystem

  • Upgraded INC v1.9 into HuggingFace Optimum
  • Upsteamed INC quantized mobilenet & faster-rcnn models to ONNX Model Zoo

Examples

  • Supported quantization on 300 random models
  • Added bare-metal examples for Bert-mini and DLRM

Validated Configurations

  • Python 3.7, 3.8, 3.9
  • Centos 8.3 & Ubuntu 18.04 & Win10
  • TensorFlow 2.6.2, 2.7, 2.8
  • Intel TensorFlow 1.15.0 UP3, 2.7, 2.8
  • PyTorch 1.8.0+cpu, 1.9.0+cpu, 1.10.0+cpu
  • IPEX 1.8.0, 1.9.0, 1.10.0
  • MxNet 1.6.0, 1.7.0, 1.8.0
  • ONNX Runtime 1.8.0, 1.9.0, 1.10.0

Distribution:

  Channel Links Install Command
Source Github https://github.com/intel/neural-compressor.git $ git clone https://github.com/intel/neural-compressor.git
Binary Pip https://pypi.org/project/neural-compressor $ pip install neural-compressor
Binary Conda https://anaconda.org/intel/neural-compressor $ conda install neural-compressor -c conda-forge -c intel

Contact:

Please feel free to contact [email protected], if you get any questions.