-
-
Notifications
You must be signed in to change notification settings - Fork 15.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ollama: 0.5.4 -> 0.5.5 #373234
ollama: 0.5.4 -> 0.5.5 #373234
Conversation
0.5.4 was just merged, please rebase |
Yes, I would prefer if we wait for it to be tagged as a proper release. |
looks like 0.5.5 is official now. |
22dc139
to
ad52a9b
Compare
|
Fails on
|
Tested and can verify that now 0.5.5 runs well on x86_64 linux + Nvidia GPU. More specifically the |
Commit message should read |
This issue is still making the |
Yes, because @liberodark still has not applied the suggestion from my comment #373234 (comment) Can you try if that fixes the build for you? |
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM !
|
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Good job, thanks !
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
looks good
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Tested ollama-rocm
: it appears to function correctly.
postInstall = lib.optionalString (stdenv.hostPlatform.isx86 || enableRocm || enableCuda) '' | ||
# copy libggml_*.so and runners into lib | ||
# https://github.com/ollama/ollama/blob/v0.4.4/llama/make/gpu.make#L90 | ||
mkdir -p $out/lib | ||
cp -r dist/*/lib/* $out/lib/ | ||
''; |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Seems good enough!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Oh, someone should check that x86_64-darwin's happy. I can't.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
My only lingering worry is that devices like Jetson have Cuda and aarch64.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
My only lingering worry is that devices like Jetson have Cuda and aarch64.
The condition (stdenv.hostPlatform.isx86 || enableRocm || enableCuda)
only contains ||
(or), not &&
(and), so it should work fine if CUDA is requested on ARM (as long as ollama and llama-cpp correctly support that combination). However, enableCuda
does require Linux; most programs aren't designed to run on bare metal, though, so that's probably not especially notable.
|
@GaetanLepage but why are the darwin build failures acceptable? |
they are spurious - both chatd and ollama build just fine on my aarch64-darwin and also on x86_64-darwin |
Yes, in particular, the |
not sure if its related but encountered this on updating unstable and rebuilding:
amd gpu + rocm config:
|
Changelog :
https://github.com/ollama/ollama/releases/tag/v0.5.4 > https://github.com/ollama/ollama/releases/tag/v0.5.5
Also is pre-release so probably can wait for release.
Things done
nix.conf
? (See Nix manual)sandbox = relaxed
sandbox = true
nix-shell -p nixpkgs-review --run "nixpkgs-review rev HEAD"
. Note: all changes have to be committed, also see nixpkgs-review usage./result/bin/
)Add a 👍 reaction to pull requests you find important.