-
Notifications
You must be signed in to change notification settings - Fork 205
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Lack of half precision as datatype for PTQ. #1173
Comments
I will look into that. It shouldn't be a problem. Thanks for this, I'll update once the PR is opened/merged |
Many thanks for the quick feedback @Giuseppe5. I can put PR asap. I have tested and seen no problems at all. |
Feel free to open a PR, but I might ask please to also update the README? We have the list with all the options, and we would like to that up-to-date. We generally just use the output of the argparse helper. Thanks! |
Also, please check the CONTRIBUTING.md file about how to sign off your commit |
Hi @Giuseppe5, I've put the PR Fantastic work btw, many thanks for this API. I needed the custom bit-width implementations for FP16, and this API provides that. |
Sometimes we struggle to keep up with our own rules, but we're working hard to improve :) |
I might be missing something very simple, so I wanted to ask for clarification regarding the data type support in
ptq.evaluate.py.
Currently, the script includes the following argument for data type selection:
Have we considered extending this to support FP16 (i.e. torch.half) by modifying the argument as follows?
My question is: Is there a specific reason we don’t include FP16 support by default but include BF16?
The text was updated successfully, but these errors were encountered: