From fffc2c028a08f9cc06fcf83ac8ec53e9d5e8ce97 Mon Sep 17 00:00:00 2001 From: Konrad Zawora Date: Fri, 13 Dec 2024 07:26:46 +0100 Subject: [PATCH] Add version inference via setuptools-scm (#58) --- pyproject.toml | 18 ++++++++++++++++++ setup.py | 8 +------- 2 files changed, 19 insertions(+), 7 deletions(-) create mode 100644 pyproject.toml diff --git a/pyproject.toml b/pyproject.toml new file mode 100644 index 00000000..db74425f --- /dev/null +++ b/pyproject.toml @@ -0,0 +1,18 @@ +[build-system] +requires = ["setuptools>=64", "setuptools-scm>=8"] +build-backend = "setuptools.build_meta" + +[project] +name = "vllm-hpu-extension" +dynamic = ["version"] +description = "HPU extension package for vLLM" +authors = [ + {name = "Tomasz Zielinski", email = "tzielinski@habana.ai"}, + {name = "Konrad Zawora", email = "kzawora@habana.ai"}, +] +dependencies = [] +requires-python = "==3.10.*" +readme = "README.md" +license = {text = "Apache 2.0"} + +[tool.setuptools_scm] diff --git a/setup.py b/setup.py index 5c7a91b3..62801513 100644 --- a/setup.py +++ b/setup.py @@ -2,20 +2,14 @@ setup( name="vllm-hpu-extension", - version="0.1", packages=find_packages(), install_requires=[], - author="Tomasz Zielinski", - author_email="tzielinski@habana.ai", - description="HPU extension package for vLLM", long_description="HPU extension package for vLLM. The package contains custom HPU-specific ops. It only works together with vLLM.", long_description_content_type="text/markdown", - url="https://github.com/tzielinski-habana/vllm-hpu-extension", + url="https://github.com/HabanaAI/vllm-hpu-extension", classifiers=[ "Programming Language :: Python :: 3", "License :: OSI Approved :: Apache 2.0", "Operating System :: OS Independent", ], - python_requires=">=3.6", ) -