Skip to content

Latest commit

 

History

History
27 lines (25 loc) · 2.81 KB

setup-local.md

File metadata and controls

27 lines (25 loc) · 2.81 KB

Ollama Setup

  • Go to Ollama's Linux download page and run the simple curl command they provide. The command should be curl https://ollama.ai/install.sh | sh.
  • Since Ollama will run as a systemd service, there is no need to run ollama serve unless you disable it. If you do disable it or have an older ollama version, do the following:
    • In terminal 1 -> ollama serve to setup ollama
    • In terminal 2 -> ollama run [model name], for example ollama run llama2
      • The models can vary as you can create your own model. You can also view ollama's library of models.
  • Otherwise, if you have the latest ollama, you can just run ollama run [model name] rather than running this in 2 terminals.
  • If there are any issues running ollama because of missing LLMs, run ollama pull [model name] as it will pull the model if Ollama has it in their library.
    • This can also be done in wsl for Windows machines.
    • This should also not be a problem is a future feature that allows for pulling of models via discord client. For now, they must be pulled manually.
  • You can now interact with the model you just ran (it might take a second to startup).
    • Response time varies with processing power!

Note

You can now pull models directly from the Discord client using /pull-model <model-name> or /switch-model <model-name>. They must exist from your local model library or from the Ollama Model Library

To Run Locally (without Docker)

  • Run npm install to install the npm packages.
  • Ensure that your .env file's OLLAMA_IP is 127.0.0.1 to work properly.
    • You only need your CLIENT_TOKEN, OLLAMA_IP, OLLAMA_PORT.
    • The ollama ip and port should just use it's defaults by nature. If not, utilize OLLAMA_IP = 127.0.0.1 and OLLAMA_PORT = 11434.
  • Now, you can run the bot by running npm run client which will build and run the decompiled typescript and run the setup for ollama.
    • IMPORTANT: This must be ran in the wsl/Linux instance to work properly! Using Command Prompt/Powershell/Git Bash/etc. will not work on Windows (at least in my experience).
    • Refer to the resources on what node version to use.
  • If you are using wsl, open up a separate terminal/shell to startup the ollama service. Again, if you are running an older ollama, you must run ollama serve in that shell.
    • If you are on an actual Linux machine/VM there is no need for another terminal (unless you have an older ollama version).
    • If you do not have a model, you can optionally run ollama pull [model name] in wsl prior to application start. You are not required as it can be pulled from the Discord client.