diff --git a/bitsandbytes/autograd/_functions.py b/bitsandbytes/autograd/_functions.py index 79ff48f44..e9610d554 100644 --- a/bitsandbytes/autograd/_functions.py +++ b/bitsandbytes/autograd/_functions.py @@ -517,7 +517,6 @@ def forward(ctx, A, B, out=None, bias=None, quant_state: Optional[F.QuantState] # 1. Dequantize # 2. MatmulnN - print("*******quant_state absmax: ", quant_state.absmax) output = torch.nn.functional.linear(A, F.dequantize_4bit(B, quant_state).to(A.dtype).t(), bias) # 3. Save state @@ -596,5 +595,4 @@ def matmul_4bit( out += bias return out else: - print("^^^^^^^^^grad seperate path^^^^^^^^^") return MatMul4Bit.apply(A, B, out, bias, quant_state)