-
Notifications
You must be signed in to change notification settings - Fork 6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Implementation detail of Quantizaiton? #2
Comments
Hi @llCurious , thanks for your interest in our work and sorry for late reply. You are right, the input to the quantization function q_k is always normalized to |
Thanks for your replay.
Since the raw weight/input may have much larger range. E.g. the input (maybe the output of some FC layer) can be significantly large if the size of the neuron for this layer is like 128 and the data dimension is 1,000. In this case, after normalization, the magnitude seems to change a lot.
|
Hi @llCurious , I do not see difference between denormalization and dequantization. We also clip the input before scaling and quantization. |
By the way, what is the underlying data type for the whole quantization procedure? It seems to be |
Hey, your work is well-presented and i just wonder one detail:
How do you ensure that the input to your quantization function is in the range [0,1]?
As you mentioned in
models/quant_ops
(this link), do you require that the input is normalized in advance?The text was updated successfully, but these errors were encountered: