-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: alt-text auto generation #22
base: main
Are you sure you want to change the base?
feat: alt-text auto generation #22
Conversation
@Chanbinski, I took a quick check of the code so far, with an interest in identifying how it was doing the alt text generation. I find Have you found a way to put the model in a shared location for all users in our jupyterhub / container context? |
Thanks @gmerritt for the review. This code is intended to run on the JupyterHub and assumes that
I was testing on a hub where the ollama binary was installed in the right place. The issue is that the code doesn't run ollama that is installed when the extension is launched. |
@Chanbinski, I believe your are conflating the separate, distinct notions of the framework vs. the model? Ollama is an application that implements a framework that can load & host models and present a chat completion API. The binary you reference is just the framework. You have to load a model. Even if the binary for the ollama application is in a shared location, when you ask ollama (via the api) to do a chat completion with a certain model, ollama will load that model in user filespace. There is no way to tell ollama to use a shared filespace for its models. Please send a reference if you have found otherwise -- but I spent hours on this, it was very frustrating. Edit: see addendum in comment to follow! |
Oh, and to refine -- I believe that the real deal-breaker that ollama insists to write to the model location directory. I was mis-remembering the exact issue! |
@gmerritt Got it, thanks for letting me know. I will let you know if I come up with any solutions to that. But the problem prior to that is that the extension doesn't call the function to start ollama in the first place. I've added code (that can be seen in the latest commit) so that the extension serves as a jupyter server extension, but it seems like it's not calling _load_jupyter_server_extension (from init.py) when the extension is first launched. I've followed the exact process from the juputer server docs and referred to a similar PR, but it seems like there is a problem. |
The code doesn't work but created a PR for code review.