From 717802655fc30d2a105e5cabb852f0a9bde625d8 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 29 Apr 2024 01:20:49 +0000 Subject: [PATCH] build(deps): bump llama-cpp-python from 0.2.63 to 0.2.65 Bumps [llama-cpp-python](https://github.com/abetlen/llama-cpp-python) from 0.2.63 to 0.2.65. - [Release notes](https://github.com/abetlen/llama-cpp-python/releases) - [Changelog](https://github.com/abetlen/llama-cpp-python/blob/main/CHANGELOG.md) - [Commits](https://github.com/abetlen/llama-cpp-python/compare/v0.2.63...v0.2.65) --- updated-dependencies: - dependency-name: llama-cpp-python dependency-type: direct:production update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] --- poetry.lock | 6 +++--- pyproject.toml | 2 +- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/poetry.lock b/poetry.lock index 26b0700..25e79b7 100644 --- a/poetry.lock +++ b/poetry.lock @@ -143,12 +143,12 @@ i18n = ["Babel (>=2.7)"] [[package]] name = "llama-cpp-python" -version = "0.2.63" +version = "0.2.65" description = "Python bindings for the llama.cpp library" optional = false python-versions = ">=3.8" files = [ - {file = "llama_cpp_python-0.2.63.tar.gz", hash = "sha256:26425fd77a8c463bac709c1d4b8194d7c4ddca3d795f82b2ae7ff4246f2f5539"}, + {file = "llama_cpp_python-0.2.65.tar.gz", hash = "sha256:5d8b8bec70576176f213b7abedda267aedafda1d165b4a30b3ebf9d3df597d55"}, ] [package.dependencies] @@ -577,4 +577,4 @@ standard = ["colorama (>=0.4)", "httptools (>=0.5.0)", "python-dotenv (>=0.13)", [metadata] lock-version = "2.0" python-versions = "^3.11" -content-hash = "61ccf7ddff47b21d4afeb46b2b79485b257993961a016d2bf91bdd35f4ff09a3" +content-hash = "75ca1d3aaad3b5c1fb32721275a082b4b9fe95088cda25007c56187f27868041" diff --git a/pyproject.toml b/pyproject.toml index a2ef101..c0ea59f 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -8,7 +8,7 @@ readme = "README.md" [tool.poetry.dependencies] python = "^3.9" -llama-cpp-python = {extras = ["server"], version = "^0.2.63"} +llama-cpp-python = {extras = ["server"], version = "^0.2.65"} uvicorn = "^0.29.0"