OpenCV dnn import quantized model
Is it possible to load NN quantized model of tensorflow?
The nodes and weights of a neural network are originally stored as 32-bit floating point numbers. The simplest motivation for quantization is to compress each float value to an eight-bit integer.
If so how?
Thank you.
If you mean https://github.com/tensorflow/tensorf... so answer is yes. It supports import of quantized models from TensorFlow. However all the computations are made in FP32.
So what's benifit of that though?
I guess even TensorFlow will convert the weights in FP32 because it' s just a way to compress model size (read the docs:
This is mostly useful for shrinking file sizes
). You need to usequantize_nodes
instead.Can dnn module leverage on that (some speed-up) if used quantized model?