-
Notifications
You must be signed in to change notification settings - Fork 21
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
PrepareB but take integers instead of float #42
Comments
Prepare B if B is quantized and transposed: Prepare B if B is transposed I think we can merge them to the master first and then try to do some optimizations. |
Ooh |
Merged prepare-b-quantized-transposed in 03a4a9d |
We need prepareB if B is only quantized too. |
Also, a slight enhancement, it would be nice (and probably more important from performance point of view) to have transpose and Quantize for prepareA. The affine and dot operators take |
So we need all combinations?:
|
The current PrepareB function combines quantization and rearrangement. The rearragement is dependent on register length. We're going to want to distribute int8 models in an architecture-independent fashion (probably as row major) then have them rearranged at load. The Quantize function already converts to int8 format without rearranging. So what's needed is an int8 rearrangement function.
Possibly with a preprocessing template, though that sounds complicated.
Also worth considering if this should be done in-place or copying.
The text was updated successfully, but these errors were encountered: