Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

8-bit quantization MVP #347

Closed
21 of 22 tasks
robertknight opened this issue Sep 6, 2024 · 1 comment
Closed
21 of 22 tasks

8-bit quantization MVP #347

robertknight opened this issue Sep 6, 2024 · 1 comment
Labels
quantization Issues related to support for quantized data types or operations

Comments

@robertknight
Copy link
Owner

robertknight commented Sep 6, 2024

The issue tracks the work involved in an MVP of 8-bit quantization support. The goal is to be able to convert and run:

@robertknight robertknight pinned this issue Oct 13, 2024
@robertknight robertknight added the quantization Issues related to support for quantized data types or operations label Feb 3, 2025
@robertknight
Copy link
Owner Author

Initial support for running quantized models has been released as part of v0.16.0. The quantization guide has more details and steps for quantizing ONNX models with recommended settings.

@robertknight robertknight unpinned this issue Feb 8, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
quantization Issues related to support for quantized data types or operations
Projects
None yet
Development

No branches or pull requests

1 participant