Skip to content

Commit d474616

Browse files
authored
Fix int4 quantization
Discovered by @HDCharles
1 parent 7d45270 commit d474616

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

quantize.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -486,7 +486,7 @@ def __init__(
486486
bias=True, device=None, dtype=None, groupsize: int = 128, inner_k_tiles: int = 8, use_cuda=True,
487487
) -> None:
488488
super().__init__()
489-
self.padding = _check_linear_int4_k(in_features, groupsize, inner_k_tiles)
489+
self.padding = not _check_linear_int4_k(in_features, groupsize, inner_k_tiles)
490490
if self.padding:
491491
from model import find_multiple
492492
self.origin_in_features = in_features

0 commit comments

Comments
 (0)