Why do the values of learnables in a quantized dlnetwork still stored as float32(single precision)?

조회 수: 4 (최근 30일)
Even though the dlquantizer is quantizing the weights of the fully connected layer to int8 and bias of the layer to int32, why do I see in the quantized dlnetwork the values are still stored as float32(single precision)?
Also, I would like to find out if dlquantizer can quantize a particular layer or not?

채택된 답변

MathWorks Fixed Point Team
MathWorks Fixed Point Team 2025년 7월 18일
편집: MathWorks Fixed Point Team 2025년 7월 18일
Yes, the learnables on the dlnetwork/quantized network are still stored as single precision.
Consider estimating parameter memory of the quantized network once deployed using the API: https://www.mathworks.com/help/deeplearning/ref/estimatenetworkmetrics.html.
The layers that it decided to quantize: https://www.mathworks.com/help/deeplearning/ug/supported-layers-for-quantization.html. It changes across releases and varies among intended targets.
The 'Analyze for Compression' feature (available in R2025a) in the Deep Network designer app -- it'll show you which layers in your network are supported for quantization, which can be friendlier than manually comparing to the supported layers doc page. It currently only analyzes for the MATLAB execution environment.

추가 답변 (0개)

카테고리

Help CenterFile Exchange에서 Quantization, Projection, and Pruning에 대해 자세히 알아보기

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by