fix bug
This commit is contained in:
parent
f26615fc0c
commit
234004ceb5
|
|
@ -108,9 +108,9 @@ def matmul4bit(x, qweight, scales, zeros, groupsize=-1):
|
|||
if np.prod(x.shape[:-1]) > auto_switch_thd:
|
||||
output = _matmul4bit_v1_recons(x, qweight, scales, zeros)
|
||||
else:
|
||||
output = _matmul4bit_v1(x, qweight, scales, zeros)
|
||||
output = _matmul4bit_v1(x, qweight, scales.float(), zeros)
|
||||
else:
|
||||
output = _matmul4bit_v1(x, qweight, scales, zeros)
|
||||
output = _matmul4bit_v1(x, qweight, scales.float(), zeros)
|
||||
else:
|
||||
# use v2
|
||||
if use_new:
|
||||
|
|
@ -118,9 +118,9 @@ def matmul4bit(x, qweight, scales, zeros, groupsize=-1):
|
|||
if np.prod(x.shape[:-1]) > auto_switch_thd:
|
||||
output = _matmul4bit_v2_recons(x, qweight, scales, zeros, groupsize)
|
||||
else:
|
||||
output = _matmul4bit_v2(x, qweight, scales, zeros, groupsize)
|
||||
output = _matmul4bit_v2(x, qweight, scales.float(), zeros, groupsize)
|
||||
else:
|
||||
output = _matmul4bit_v2(x, qweight, scales, zeros, groupsize)
|
||||
output = _matmul4bit_v2(x, qweight, scales.float(), zeros, groupsize)
|
||||
return output
|
||||
|
||||
|
||||
|
|
|
|||
Loading…
Reference in New Issue