Skip to content

Intel® Neural Compressor v1.7 Release

Compare
Choose a tag to compare
@ftian1 ftian1 released this 01 Oct 06:05
· 2439 commits to master since this release

Intel® Neural Compressor(formerly known as Intel® Low Precision Optimization Tool) v1.7 release is featured by:

Features

  • Quantization
    • Improved quantization accuracy in SSD-Reset34 and MobileNet v3 on TensorFlow
  • Pruning
    • Supported magnitude pruning on TensorFlow
  • Knowledge distillation
    • Supported knowledge distillation on PyTorch
  • Multi-node support
    • Supported multi-node pruning with distributed dataloader on PyTorch
    • Supported multi-node inference for benchmark on PyTorch
  • Acceleration library
    • Added a domain-specific acceleration library for NLP models

Productivity

  • Supported the configuration-free (pure Python) quantization
  • Improved ease-of-use user interface for quantization with few clicks

Ecosystem

  • Integrated into HuggingFace optimization library (Optimum)
  • Upstreamed INC quantized models (RN50, VGG16) to ONNX Model Zoo

Documentation

  • Add tutorial and examples for knowledge distillation
  • Add tutorial and examples for multi-node training
  • Add tutorial and examples for acceleration library

Validated Configurations

  • Python 3.6 & 3.7 & 3.8 & 3.9
  • Centos 8.3 & Ubuntu 18.04
  • TensorFlow 2.6.0
  • Intel TensorFlow 2.4.0, 2.5.0 and 1.15.0 UP3
  • PyTorch 1.8.0+cpu, 1.9.0+cpu, IPEX 1.8.0
  • MxNet 1.6.0, 1.7.0, 1.8.0
  • ONNX Runtime 1.6.0, 1.7.0, 1.8.0

Distribution:

  Channel Links Install Command
Source Github https://github.com/intel/neural-compressor.git $ git clone https://github.com/intel/neural-compressor.git
Binary Pip https://pypi.org/project/neural-compressor $ pip install neural-compressor
Binary Conda https://anaconda.org/intel/neural-compressor $ conda install neural-compressor -c conda-forge -c intel

Contact:

Please feel free to contact [email protected], if you get any questions.