-
Notifications
You must be signed in to change notification settings - Fork 31
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Running initLlama with lora / lora_scaled set, fails to load context and crashes app. #86
Comments
The For Line 201 in c1d15a3
|
Thanks for incredibly quick help, really appreciate it! I've tried different ways of running initLlama e.g with/without the 'file://' extension on both model and lora. But I can't find a configuration of it that will work without closing the app, sorry if I'm missing anything obvious here. In this example file.uri and file.lora are:
This is my current setup for initLlama().
Thanks again! |
Tested with bartowski/Meta-Llama-3.1-8B-Instruct-GGUF as the base model and grimjim/Llama-3-Instruct-abliteration-LoRA-8B (converted) as the lora adapter, no issue on my Android device (Pixel 6). Could you share what model & lora you are using? Also, Android hardware info that may be helpful. |
I'm successfully managing to initLlama and inference on it, without lora:
But if I add lora / lora_scaled, it fails to load and crashes without erroring.
Any help would be greatly appreciated - running on Android. I'm loading the files in from assets/models then moving them to DocumentDirectoryPath and calling them from there. 🥸
The text was updated successfully, but these errors were encountered: