We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 85a01b0 commit 2c489f8Copy full SHA for 2c489f8
bitsandbytes/backends/cpu_xpu_common.py
@@ -377,6 +377,7 @@ def quantize_4bit_impl(
377
int(lowp_mode),
378
-1, # act_quant_mode. -1 means don't quant activation
379
)
380
+ state.absmax = torch.Tensor()
381
return torch.Tensor(), state
382
383
return out, state
@@ -444,6 +445,7 @@ def dequantize_4bit_impl(
444
445
assert quant_state.op_context is not None
446
A = quant_state.op_context.to_public(quant_state.op_context.get_weight())
447
A = A.reshape(-1)
448
+ absmax = quant_state.op_context.get_scales().reshape(-1)
449
450
if out is None:
451
out = torch.empty(quant_state.shape, dtype=quant_state.dtype, device=A.device)
0 commit comments