This is my first attempt at using LLMs to write a coding buddy. This was written on an Apple Silicon laptop and requires that the model chosen fits within the memory constraints of your system.
Choose a model. You'll need to choose one that fits within your available
memory, this defaults to llama3:latest which is one of the smallest models I can
find. Set your model in my_model
on line 19.
Setup a virtual env
python3 -m venv .venv
source .venv/bin/activate
Install the required packages
pip install -r requirements.txt
pip install llama-index-readers-github
Configure the server
Read from GitHub repo jeffwelling/giticket
export CODEY_SOURCE="github"
export CODEY_GITHUB_OWNER="jeffwelling"
export CODEY_GITHUB_REPO="giticket"
export GITHUB_TOKEN="SomeSuperSecretTokenGoesHere"
Read from local directory called codey_data
export CODEY_SOURCE="dir"
Start the server
streamlit run codey.py
Feel free to ask questions and file issues, but this is really nothing more than some glue holding together streamlit and llama_index. I'm happy to help but I'm no expert and you may need to ask around those communities for assistance.
This project is under BSD-3-Clause license.
Copyright (c) 2024, Jeff Welling