Main Content

Get Started with Network Compression

Learn the basics of the Deep Learning Toolbox™ Model Compression Library

Use Deep Learning Toolbox together with the Deep Learning Toolbox Model Compression Library support package to reduce the memory footprint and computational requirements of a deep neural network:

  • Prune filters from convolution layers by using first-order Taylor approximation.

  • Project layers by performing principal component analysis (PCA) on the layer activations.

  • Quantize the weights, biases, and activations of layers to reduced precision scaled integer data types.

You can then generate code from the compressed network to deploy to your desired hardware.

Diagram of suggested compression workflow: first pruning, then projection, then quantization, then code generation.

Topics

Featured Examples