Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

oonxruntime-gpu: Reduce overall project size by levearging oonx cpu and gpu runtimes #71

Open
arky opened this issue Feb 20, 2025 · 2 comments

Comments

@arky
Copy link

arky commented Feb 20, 2025

The current overall deployment of AddaxAI requires 4GB install file and 20 GB of free disk space. The most of the space is taken by python environments.

Python projects are now leverage on oonxruntime (https://pypi.org/project/onnxruntime-gpu/) packages to directly models. This issue is filed to explore the possibility of exploring this new feature to reduce the footprint of AddaxAI

@PetervanLunteren
Copy link
Owner

Thanks for sharing @arky! Do I understand correctly that onnxruntime-gpu only works for ONNX format models?

@arky
Copy link
Author

arky commented Mar 3, 2025

@PetervanLunteren Yes, onnxruntime-[cpu,gpu] is specific to Microsoft ONNX format.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants