Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GPU offloading proper number #252

Open
alexjp opened this issue Mar 4, 2025 · 1 comment
Open

GPU offloading proper number #252

alexjp opened this issue Mar 4, 2025 · 1 comment

Comments

@alexjp
Copy link

alexjp commented Mar 4, 2025

Hi,

I have a model that I usually have to use in llama-cpp because of the proper number of GPU offloading.

Example, in the model in the image below, LM Studio allows me in configuration to select GPU offloading from 0 to 40.
If I select 40, I see in the logs, that it offloaded 41/41 layers to GPU.
If I select 39 in the configuration, it shows in the logs that it offloaded 39/41 layers to GPU.

How do I offload 40/41 layers to the GPU? I can do this in llama-cpp directly.

Image

Thanks for LM Studio!

@yagil
Copy link
Member

yagil commented Mar 4, 2025

Good question. cc @mattjcly

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants