Size of a quantized DNN
8 views (last 30 days)
Show older comments
Hello,
I am using the deep network quantizer to quantize a deep CNN. I followed instructions given at https://www.mathworks.com/help/deeplearning/ref/deepnetworkquantizer-app.html . I used the Export button and selected Export Quantizer to export my quantized network. But when I compare the size of both networks (the original one before quantization and the quantized one stored in the dlquantizer object), I see they both have the same number of bytes in the workspace.
Is my way for checking the size of network correct and resonable?
The app shows 73% size reduction which I can not find when comparing the networks!
Any help os appreciated.
0 Comments
Answers (1)
Jacob Mathew
on 31 Jul 2025
Quantizing a neural network does not necessarily lead to reduction in the physical storage size needed to represent the model. The goal is to replace floating point computations with fixed point or look up table based activation functions as well as reducing the bits needed for activation function calculation. A detailed definition and workflows that go behind quantizing a network can be found in the documentation below:
As such the goal is to reduce execution memory requirements and specialised hardware requirements allowing ML and DL models to run on lower power hardware
0 Comments
See Also
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!