Skip to content

Commit

Permalink
Revert changes to ollama compose
Browse files Browse the repository at this point in the history
Signed-off-by: JeffMboya <[email protected]>
  • Loading branch information
JeffMboya committed Nov 6, 2024
1 parent ae3914c commit 2145e09
Show file tree
Hide file tree
Showing 3 changed files with 85 additions and 85 deletions.
8 changes: 4 additions & 4 deletions docker/.env
Original file line number Diff line number Diff line change
Expand Up @@ -157,7 +157,7 @@ UV_CUBE_PROXY_INSTANCE_ID=
UV_CUBE_UI_NEXTAUTH_SECRET="cZAcFIdjxebC1XDULvfoXs_sO7ufCTRo3hW2lXtMoCvcSKkTyP"
UV_CUBE_UI_NAME="Cube AI"
# change IP address to your local IP address
UV_CUBE_UI_BASE_URL=http://localhost:3001
UV_CUBE_NEXTAUTH_URL=http://localhost:3001/api/auth
UV_CUBE_PUBLIC_BASE_URL=http://localhost:3001
UV_CUBE_PUBLIC_UI_TYPE=cube-ai
UV_CUBE_UI_BASE_URL=http://109.92.195.153:6193
UV_CUBE_NEXTAUTH_URL=http://109.92.195.153:6193/api/auth
UV_CUBE_PUBLIC_BASE_URL=http://109.92.195.153:6193
UV_CUBE_PUBLIC_UI_TYPE=cube-ai
160 changes: 80 additions & 80 deletions docker/ollama-compose.yaml
Original file line number Diff line number Diff line change
@@ -1,87 +1,87 @@
# # Copyright (c) Ultraviolet
# # SPDX-License-Identifier: Apache-2.0
# Copyright (c) Ultraviolet
# SPDX-License-Identifier: Apache-2.0

# volumes:
# open-webui:
# driver: local
# ollama:
# driver: local
volumes:
open-webui:
driver: local
ollama:
driver: local

# services:
# ollama:
# container_name: ollama
# image: ollama/ollama:0.3.12 # For AMD GPU, use ollama/ollama:0.3.8-rocm
# restart: unless-stopped
# volumes:
# - ollama:/root/.ollama
# tty: true
# networks:
# - cube-network
# # # Uncomment the following lines to enable AMD GPU support
# # devices:
# # - /dev/dri:/dev/dri
# # - /dev/kfd:/dev/kfd
# # environment:
# # - "HSA_OVERRIDE_GFX_VERSION=${HSA_OVERRIDE_GFX_VERSION-11.0.0}"
services:
ollama:
container_name: ollama
image: ollama/ollama:0.3.12 # For AMD GPU, use ollama/ollama:0.3.8-rocm
restart: unless-stopped
volumes:
- ollama:/root/.ollama
tty: true
networks:
- cube-network
# # Uncomment the following lines to enable AMD GPU support
# devices:
# - /dev/dri:/dev/dri
# - /dev/kfd:/dev/kfd
# environment:
# - "HSA_OVERRIDE_GFX_VERSION=${HSA_OVERRIDE_GFX_VERSION-11.0.0}"

# # # Uncomment the following lines to enable Nvidia GPU support
# # deploy:
# # resources:
# # reservations:
# # devices:
# # - driver: ${OLLAMA_GPU_DRIVER-nvidia}
# # count: ${OLLAMA_GPU_COUNT-1}
# # capabilities:
# # - gpu
# # Uncomment the following lines to enable Nvidia GPU support
# deploy:
# resources:
# reservations:
# devices:
# - driver: ${OLLAMA_GPU_DRIVER-nvidia}
# count: ${OLLAMA_GPU_COUNT-1}
# capabilities:
# - gpu

# open-webui:
# container_name: open-webui
# image: ghcr.io/open-webui/open-webui:0.3.32-ollama
# restart: unless-stopped
# volumes:
# - open-webui:/app/backend/data
# ports:
# - 3000:8080
# environment:
# - OLLAMA_BASE_URL=http://ollama:11434
# networks:
# - cube-network
open-webui:
container_name: open-webui
image: ghcr.io/open-webui/open-webui:0.3.32-ollama
restart: unless-stopped
volumes:
- open-webui:/app/backend/data
ports:
- 3000:8080
environment:
- OLLAMA_BASE_URL=http://ollama:11434
networks:
- cube-network

# pull-tinyllama:
# image: docker:27.3.1
# container_name: pull-tinyllama
# restart: on-failure
# depends_on:
# - ollama
# entrypoint: /bin/sh
# command: -c "docker exec ollama ollama run tinyllama:1.1b"
# volumes:
# - /var/run/docker.sock:/var/run/docker.sock
# networks:
# - cube-network
pull-tinyllama:
image: docker:27.3.1
container_name: pull-tinyllama
restart: on-failure
depends_on:
- ollama
entrypoint: /bin/sh
command: -c "docker exec ollama ollama run tinyllama:1.1b"
volumes:
- /var/run/docker.sock:/var/run/docker.sock
networks:
- cube-network

# pull-starcoder2:
# image: docker:27.3.1
# container_name: pull-starcoder2
# restart: on-failure
# depends_on:
# - ollama
# entrypoint: /bin/sh
# command: -c "docker exec ollama ollama pull starcoder2:3b"
# volumes:
# - /var/run/docker.sock:/var/run/docker.sock
# networks:
# - cube-network
pull-starcoder2:
image: docker:27.3.1
container_name: pull-starcoder2
restart: on-failure
depends_on:
- ollama
entrypoint: /bin/sh
command: -c "docker exec ollama ollama pull starcoder2:3b"
volumes:
- /var/run/docker.sock:/var/run/docker.sock
networks:
- cube-network

# pull-nomic-embed-text:
# image: docker:27.3.1
# container_name: pull-nomic-embed-text
# restart: on-failure
# depends_on:
# - ollama
# entrypoint: /bin/sh
# command: -c "docker exec ollama ollama pull nomic-embed-text:v1.5"
# volumes:
# - /var/run/docker.sock:/var/run/docker.sock
# networks:
# - cube-network
pull-nomic-embed-text:
image: docker:27.3.1
container_name: pull-nomic-embed-text
restart: on-failure
depends_on:
- ollama
entrypoint: /bin/sh
command: -c "docker exec ollama ollama pull nomic-embed-text:v1.5"
volumes:
- /var/run/docker.sock:/var/run/docker.sock
networks:
- cube-network
2 changes: 1 addition & 1 deletion ui/config.json
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@
},
"allowMultipleThemes": true,
"themes": {
"availableTheme": ["default", "midnightsky", "tealtide", "graywave"],
"availableTheme": "default",
"defaultTheme": "default"
},
"favicon": {
Expand Down

0 comments on commit 2145e09

Please sign in to comment.