You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I want to use bitblas to implement gemm of w4a4, README shows that bitblas supports this operation, but in QuickStart it shows that A_type as int4 is not supported?
Also, I have some questions about the bitblas API. The w4a4 matrix multiplication I need to implement will quantize both weights and activations, accumulate them via int32, and finally dequantize them to fp16. The dequantization requires the “scales” of both weight and activation, but it seems that the current API only supports providing the “scales” of the weights?
The text was updated successfully, but these errors were encountered:
Thanks for your great work!
I want to use bitblas to implement gemm of w4a4, README shows that bitblas supports this operation, but in QuickStart it shows that A_type as int4 is not supported?
Also, I have some questions about the bitblas API. The w4a4 matrix multiplication I need to implement will quantize both weights and activations, accumulate them via int32, and finally dequantize them to fp16. The dequantization requires the “scales” of both weight and activation, but it seems that the current API only supports providing the “scales” of the weights?
The text was updated successfully, but these errors were encountered: