-
-
Notifications
You must be signed in to change notification settings - Fork 157
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] bitlinear fix #42
Comments
Can you elaborate please? Can you go deeper? |
I encountered the same problem. When passing a tensor of 4,2 int to a BitLinear(2,8), I get an error at the line |
I think so, but I am confused that since self.gamma is related to activations while self.beta is related to weights, should we explicitly broadcast these two matrices [quantization about activations ('group_size = x.shape[0] // self.num_groups') should be grouped in the dim=1(x.shape[1]) because of the batch_size?], thus 'x * self.gamma * self.beta' in the dequantization process can do hadamard product? If I make wrong, pls point out. Thanks. |
emmm, I see the owner update the new code. (without group quantization) |
Stale issue message |
beta and gamma sizes to be (1, weight.shape[0], not (weight.shape[0], 1) ???
Upvote & Fund
The text was updated successfully, but these errors were encountered: