-
Notifications
You must be signed in to change notification settings - Fork 104
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Get scale value after Full Integer Quantization #132
Comments
The quantizers (e.g. |
a navie question, for INT8 quantization using Mnist example, is the Input divided by 255 (normilized to [0 1] ) or not ? |
Well, I usually don't divide the input to normalize it 0 to 1. That way the inputs are integer. You could divide it, and that way you will get non-integer inputs, that on PC can work with floating-point. Then if you want to do deploy this to some custom hardware with out floating-point units, you can use fixed-point. With regards on how this affects training, in my experience it does not, at least for (relatively) shallow networks I usually train. |
when doing the Inference on edge device with INT8 weights/bias, the accuracy is very low, |
Hm. This is a bit unusual. I am not exactly sure what is causing your issues, but in general you can also skip softmax in inference, if you are only using it for picking the best class. Regarding large integers. One thing you could do (if you are not doing it yet) is to saturate the intermediate activations. E.g. have a ReLU with saturation. |
There is a related issue #111,
But, I'm still confused about how to get the scale value after full integer quantization , any one knows how to get it ?
The text was updated successfully, but these errors were encountered: