You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
hi @yaoyaoding , thanks for your attention! currently we do not have implementations of these bits which is not a pow of 2 (which may lead to some empty slot when compress data, for example, only 2 int3 can be stored at a int8 storage value which lead to only 6/8 utilization).
But the implementation is simple, if you really request these datatypes, I'll be pleasure to add them :)
hi @yaoyaoding , thanks for your attention! currently we do not have implementations of these bits which is not a pow of 2 (which may lead to some empty slot when compress data, for example, only 2 int3 can be stored at a int8 storage value which lead to only 6/8 utilization).
But the implementation is simple, if you really request these datatypes, I'll be pleasure to add them :)
@yaoyaoding It depends, bitblas has about three different templates for dequantization (in global, in shared, in register), we can select one to deploy this computation :)
Thanks for the great work!
I am curious whether bitblas/ladder has support for the quantized data types like int3, int5, int6 (or uint3, uint5, uint6, float6_e3m2, etc)?
The text was updated successfully, but these errors were encountered: