diff --git a/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-cpu b/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-cpu index 5bba90f..7fcce73 100644 --- a/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-cpu +++ b/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-cpu @@ -32,6 +32,11 @@ RUN mv localai_entrypoint.sh /build/entrypoint_md_ai.sh RUN python3 -m pip install --no-cache-dir cryptography aiohttp tk RUN python3 -m pip install --no-cache-dir -r midori_program_requirments.txt RUN python3 -m pip cache purge + +RUN mkdir /tempmodels + +RUN midori_ai_downloader llava-v1.5-13b-Q6_K.gguf && mv llava-v1.5-13b-Q6_K.gguf /tempmodels/llava-v1.5-13b-Q6_K.gguf +RUN midori_ai_downloader mmproj-model-f16.gguf && mv mmproj-model-f16.gguf /tempmodels/mmproj-model-f16.gguf RUN apt-get autoclean && apt-get clean diff --git a/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-hipblas-gpu b/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-hipblas-gpu index 93acb2b..7235ef2 100644 --- a/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-hipblas-gpu +++ b/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-hipblas-gpu @@ -33,6 +33,11 @@ RUN python3 -m pip install --no-cache-dir cryptography aiohttp tk RUN python3 -m pip install --no-cache-dir -r midori_program_requirments.txt RUN python3 -m pip cache purge +RUN mkdir /tempmodels + +RUN midori_ai_downloader llava-v1.5-13b-Q6_K.gguf && mv llava-v1.5-13b-Q6_K.gguf /tempmodels/llava-v1.5-13b-Q6_K.gguf +RUN midori_ai_downloader mmproj-model-f16.gguf && mv mmproj-model-f16.gguf /tempmodels/mmproj-model-f16.gguf + RUN apt-get autoclean && apt-get clean RUN echo "Placeholder: USER localai" diff --git a/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-intelf16-gpu b/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-intelf16-gpu index 8e1e829..fe0f901 100644 --- a/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-intelf16-gpu +++ b/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-intelf16-gpu @@ -33,6 +33,11 @@ RUN python3 -m pip install --no-cache-dir cryptography aiohttp tk RUN python3 -m pip install --no-cache-dir -r midori_program_requirments.txt RUN python3 -m pip cache purge +RUN mkdir /tempmodels + +RUN midori_ai_downloader llava-v1.5-13b-Q6_K.gguf && mv llava-v1.5-13b-Q6_K.gguf /tempmodels/llava-v1.5-13b-Q6_K.gguf +RUN midori_ai_downloader mmproj-model-f16.gguf && mv mmproj-model-f16.gguf /tempmodels/mmproj-model-f16.gguf + RUN apt-get autoclean && apt-get clean RUN echo "Placeholder: USER localai" diff --git a/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-intelf32-gpu b/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-intelf32-gpu index 8e1e829..fe0f901 100644 --- a/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-intelf32-gpu +++ b/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-intelf32-gpu @@ -33,6 +33,11 @@ RUN python3 -m pip install --no-cache-dir cryptography aiohttp tk RUN python3 -m pip install --no-cache-dir -r midori_program_requirments.txt RUN python3 -m pip cache purge +RUN mkdir /tempmodels + +RUN midori_ai_downloader llava-v1.5-13b-Q6_K.gguf && mv llava-v1.5-13b-Q6_K.gguf /tempmodels/llava-v1.5-13b-Q6_K.gguf +RUN midori_ai_downloader mmproj-model-f16.gguf && mv mmproj-model-f16.gguf /tempmodels/mmproj-model-f16.gguf + RUN apt-get autoclean && apt-get clean RUN echo "Placeholder: USER localai" diff --git a/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-nvidia-gpu b/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-nvidia-gpu index b84acd0..5f39889 100644 --- a/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-nvidia-gpu +++ b/other_files/subsystem_docker_stuff/localai/localai-subsystem-dockerfile-nvidia-gpu @@ -33,6 +33,11 @@ RUN python3 -m pip install --no-cache-dir cryptography aiohttp tk RUN python3 -m pip install --no-cache-dir -r midori_program_requirments.txt RUN python3 -m pip cache purge +RUN mkdir /tempmodels + +RUN midori_ai_downloader llava-v1.5-13b-Q6_K.gguf && mv llava-v1.5-13b-Q6_K.gguf /tempmodels/llava-v1.5-13b-Q6_K.gguf +RUN midori_ai_downloader mmproj-model-f16.gguf && mv mmproj-model-f16.gguf /tempmodels/mmproj-model-f16.gguf + RUN apt-get autoclean && apt-get clean RUN echo "Placeholder: USER localai" diff --git a/other_files/subsystem_docker_stuff/localai/supporting_files/localai_entrypoint.sh b/other_files/subsystem_docker_stuff/localai/supporting_files/localai_entrypoint.sh index 043e71e..d4f4b08 100644 --- a/other_files/subsystem_docker_stuff/localai/supporting_files/localai_entrypoint.sh +++ b/other_files/subsystem_docker_stuff/localai/supporting_files/localai_entrypoint.sh @@ -162,15 +162,13 @@ if [ ! -f bert-MiniLM-L6-v2q4_0.bin ]; then fi if [ ! -f ggml-model-q4_k.gguf ]; then - #wget --no-check-certificate --no-cache --no-cookies https://huggingface.co/PsiPi/liuhaotian_llava-v1.5-13b-GGUF/resolve/main/llava-v1.5-13b-Q6_K.gguf - #hf-downloader -un PsiPi -r liuhaotian_llava-v1.5-13b-GGUF -m llava-v1.5-13b-Q6_K.gguf - midori_ai_downloader llava-v1.5-13b-Q6_K.gguf - mv llava-v1.5-13b-Q6_K.gguf ggml-model-q4_k.gguf + echo moving model from temp folder + mv /tempmodels/llava-v1.5-13b-Q6_K.gguf ggml-model-q4_k.gguf fi if [ ! -f mmproj-model-f16.gguf ]; then - #wget --no-check-certificate --no-cache --no-cookies https://huggingface.co/PsiPi/liuhaotian_llava-v1.5-13b-GGUF/resolve/main/mmproj-model-f16.gguf - #hf-downloader -un PsiPi -r liuhaotian_llava-v1.5-13b-GGUF -m mmproj-model-f16.gguf - midori_ai_downloader mmproj-model-f16.gguf + echo moving model from temp folder + mv /tempmodels/mmproj-model-f16.gguf mmproj-model-f16.gguf + fi if [ ! -f chat-simple.tmpl ]; then wget --no-check-certificate --no-cache --no-cookies https://github.com/mudler/LocalAI/blob/b8240b4c1839089b9d06a3e2b1c629a294cff87e/examples/configurations/llava/chat-simple.tmpl