From a3635763eb0d860e527cf3ab5f268d3b9f46b787 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 6 May 2024 01:38:46 +0000 Subject: [PATCH] build(deps): bump llama-cpp-python from 0.2.65 to 0.2.69 Bumps [llama-cpp-python](https://github.com/abetlen/llama-cpp-python) from 0.2.65 to 0.2.69. - [Release notes](https://github.com/abetlen/llama-cpp-python/releases) - [Changelog](https://github.com/abetlen/llama-cpp-python/blob/main/CHANGELOG.md) - [Commits](https://github.com/abetlen/llama-cpp-python/compare/v0.2.65...v0.2.69) --- updated-dependencies: - dependency-name: llama-cpp-python dependency-type: direct:production update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] --- poetry.lock | 6 +++--- pyproject.toml | 2 +- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/poetry.lock b/poetry.lock index 25e79b7..162eb2b 100644 --- a/poetry.lock +++ b/poetry.lock @@ -143,12 +143,12 @@ i18n = ["Babel (>=2.7)"] [[package]] name = "llama-cpp-python" -version = "0.2.65" +version = "0.2.69" description = "Python bindings for the llama.cpp library" optional = false python-versions = ">=3.8" files = [ - {file = "llama_cpp_python-0.2.65.tar.gz", hash = "sha256:5d8b8bec70576176f213b7abedda267aedafda1d165b4a30b3ebf9d3df597d55"}, + {file = "llama_cpp_python-0.2.69.tar.gz", hash = "sha256:b37e864b4d9f7ac286a3e926d87afab2f136ae9290e11088f7a205b80d3c04a9"}, ] [package.dependencies] @@ -577,4 +577,4 @@ standard = ["colorama (>=0.4)", "httptools (>=0.5.0)", "python-dotenv (>=0.13)", [metadata] lock-version = "2.0" python-versions = "^3.11" -content-hash = "75ca1d3aaad3b5c1fb32721275a082b4b9fe95088cda25007c56187f27868041" +content-hash = "5e13b3eb6120678c68127b08c529ba764a66ef8a9bcd0adfc024563075484c16" diff --git a/pyproject.toml b/pyproject.toml index c0ea59f..09d23d2 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -8,7 +8,7 @@ readme = "README.md" [tool.poetry.dependencies] python = "^3.9" -llama-cpp-python = {extras = ["server"], version = "^0.2.65"} +llama-cpp-python = {extras = ["server"], version = "^0.2.69"} uvicorn = "^0.29.0"