Skip to content

Latest commit

 

History

History
31 lines (24 loc) · 1.2 KB

Docker.md

File metadata and controls

31 lines (24 loc) · 1.2 KB

Docker

Prerequisites

  • Docker must be installed and running on your system.
  • Create a folder to store big models & intermediate files (ex. /mnt/models)

Images

There is a Docker image available for our project, you can pull the docker image by:

docker pull approachingai/ktransformers:0.2.1

Notice: In this image, we compile the ktransformers in AVX512 instuction CPUs, if your cpu not support AVX512, it is suggested to recompile and install ktransformer in the /workspace/ktransformers directory within the container.

Building docker image locally

  • Download Dockerfile in there

  • finish, execute

    docker build  -t approachingai/ktransformers:0.2.1 .

Usage

Assuming you have the nvidia-container-toolkit that you can use the GPU in a Docker container.

docker run --gpus all -v /path/to/models:/models --name ktransformers -itd approachingai/ktransformers:0.2.1
docker exec -it ktransformers /bin/bash
python -m ktransformers.local_chat  --gguf_path /models/path/to/gguf_path --model_path /models/path/to/model_path --cpu_infer 33

More operators you can see in the readme