Merge pull request #350 from abetlen/migrate-to-scikit-build-core

Migrate to scikit-build-core
This commit is contained in:
Andrei 2023-06-09 03:00:01 -04:00 committed by GitHub
commit fb2c5f7fd9
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
10 changed files with 79 additions and 147 deletions

View file

@ -22,14 +22,15 @@ jobs:
- uses: actions/setup-python@v3 - uses: actions/setup-python@v3
- name: Install cibuildwheel - name: Install cibuildwheel
run: python -m pip install cibuildwheel==2.12.1 run: python3 -m pip install cibuildwheel==2.12.1
- name: Install dependencies - name: Install dependencies
run: | run: |
python -m pip install --upgrade pip pytest cmake scikit-build setuptools python3 -m pip install --upgrade pip
python3 -m pip install --verbose --editable .
- name: Build wheels - name: Build wheels
run: python -m cibuildwheel --output-dir wheelhouse run: python3 -m cibuildwheel --output-dir wheelhouse
- uses: actions/upload-artifact@v3 - uses: actions/upload-artifact@v3
with: with:
@ -46,10 +47,11 @@ jobs:
- uses: actions/setup-python@v3 - uses: actions/setup-python@v3
- name: Install dependencies - name: Install dependencies
run: | run: |
python -m pip install --upgrade pip pytest cmake scikit-build setuptools python3 -m pip install --upgrade pip build
python3 -m pip install --verbose --editable .
- name: Build source distribution - name: Build source distribution
run: | run: |
python setup.py sdist python3 -m build --sdist
- uses: actions/upload-artifact@v3 - uses: actions/upload-artifact@v3
with: with:
path: ./dist/*.tar.gz path: ./dist/*.tar.gz

View file

@ -1,39 +0,0 @@
name: Build Docker
on: workflow_dispatch
permissions:
contents: write
packages: write
jobs:
docker:
name: Build and push Docker image
runs-on: ubuntu-latest
steps:
- name: Checkout
uses: actions/checkout@v3
with:
submodules: "true"
- name: Set up QEMU
uses: docker/setup-qemu-action@v2
- name: Set up Docker Buildx
uses: docker/setup-buildx-action@v2
- name: Login to GitHub Container Registry
uses: docker/login-action@v2
with:
registry: ghcr.io
username: ${{ github.repository_owner }}
password: ${{ secrets.GITHUB_TOKEN }}
- name: Build and push
uses: docker/build-push-action@v4
with:
context: .
push: true # push to registry
pull: true # always fetch the latest base images
platforms: linux/amd64,linux/arm64 # build for both amd64 and arm64
tags: ghcr.io/abetlen/llama-cpp-python:latest

View file

@ -1,30 +0,0 @@
# Based on: https://packaging.python.org/en/latest/guides/publishing-package-distribution-releases-using-github-actions-ci-cd-workflows/
name: Publish to TestPyPI
on: workflow_dispatch
jobs:
build-n-publish:
name: Build and publish
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v3
with:
submodules: "true"
- name: Set up Python
uses: actions/setup-python@v4
with:
python-version: "3.8"
- name: Install dependencies
run: |
python -m pip install --upgrade pip pytest cmake scikit-build setuptools
- name: Build source distribution
run: |
python setup.py sdist
- name: Publish to Test PyPI
uses: pypa/gh-action-pypi-publish@release/v1
with:
password: ${{ secrets.TEST_PYPI_API_TOKEN }}
repository-url: https://test.pypi.org/legacy/

View file

@ -19,10 +19,10 @@ jobs:
python-version: "3.8" python-version: "3.8"
- name: Install dependencies - name: Install dependencies
run: | run: |
python -m pip install --upgrade pip pytest cmake scikit-build setuptools python3 -m pip install --upgrade pip build
- name: Build source distribution - name: Build source distribution
run: | run: |
python setup.py sdist python3 -m build --sdist
- name: Publish distribution to PyPI - name: Publish distribution to PyPI
# TODO: move to tag based releases # TODO: move to tag based releases
# if: startsWith(github.ref, 'refs/tags') # if: startsWith(github.ref, 'refs/tags')

View file

@ -26,11 +26,11 @@ jobs:
python-version: ${{ matrix.python-version }} python-version: ${{ matrix.python-version }}
- name: Install dependencies - name: Install dependencies
run: | run: |
python -m pip install --upgrade pip pytest cmake scikit-build setuptools fastapi sse-starlette httpx uvicorn python3 -m pip install --upgrade pip
pip install . -v python3 -m pip install --verbose --editable .[server,test]
- name: Test with pytest - name: Test with pytest
run: | run: |
pytest python3 -m pytest
build-windows: build-windows:
@ -49,11 +49,11 @@ jobs:
python-version: ${{ matrix.python-version }} python-version: ${{ matrix.python-version }}
- name: Install dependencies - name: Install dependencies
run: | run: |
python -m pip install --upgrade pip pytest cmake scikit-build setuptools fastapi sse-starlette httpx uvicorn python3 -m pip install --upgrade pip
pip install . -v python3 -m pip install --verbose --editable .[server,test]
- name: Test with pytest - name: Test with pytest
run: | run: |
pytest python3 -m pytest
build-macos: build-macos:
@ -72,8 +72,8 @@ jobs:
python-version: ${{ matrix.python-version }} python-version: ${{ matrix.python-version }}
- name: Install dependencies - name: Install dependencies
run: | run: |
python -m pip install --upgrade pip pytest cmake scikit-build setuptools fastapi sse-starlette httpx uvicorn python3 -m pip install --upgrade pip
pip install . -v python3 -m pip install --verbose --editable .[server,test]
- name: Test with pytest - name: Test with pytest
run: | run: |
pytest python3 -m pytest

View file

@ -7,6 +7,8 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
## [Unreleased] ## [Unreleased]
- (build-system) Migrate from scikit-build to scikit-build-core
## [v0.1.59] ## [v0.1.59]
### Added ### Added

View file

@ -18,16 +18,16 @@ if (UNIX AND NOT FORCE_CMAKE)
) )
install( install(
FILES ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/libllama.so FILES ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/libllama.so
DESTINATION llama_cpp DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp
) )
else() else()
set(BUILD_SHARED_LIBS "On") set(BUILD_SHARED_LIBS "On")
add_subdirectory(vendor/llama.cpp) add_subdirectory(vendor/llama.cpp)
install( install(
TARGETS llama TARGETS llama
LIBRARY DESTINATION llama_cpp LIBRARY DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp
RUNTIME DESTINATION llama_cpp RUNTIME DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp
ARCHIVE DESTINATION llama_cpp ARCHIVE DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp
FRAMEWORK DESTINATION llama_cpp FRAMEWORK DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp
) )
endif() endif()

View file

@ -1,30 +1,38 @@
update: update:
poetry install poetry install
python3 -m pip install --upgrade pip
git submodule update --init --recursive git submodule update --init --recursive
update.vendor: update.vendor:
cd vendor/llama.cpp && git pull origin master cd vendor/llama.cpp && git pull origin master
build: build:
python3 setup.py develop python3 -m pip install --upgrade pip
python3 -m pip install --verbose --editable .
build.cuda: build.cuda:
CMAKE_ARGS="-DLLAMA_CUBLAS=on" FORCE_CMAKE=1 python3 setup.py develop python3 -m pip install --upgrade pip
CMAKE_ARGS="-DLLAMA_CUBLAS=on" FORCE_CMAKE=1 python3 -m pip install --verbose --editable .
build.opencl: build.opencl:
CMAKE_ARGS="-DLLAMA_CLBLAST=on" FORCE_CMAKE=1 python3 setup.py develop python3 -m pip install --upgrade pip
CMAKE_ARGS="-DLLAMA_CLBLAST=on" FORCE_CMAKE=1 python3 -m pip install --verbose --editable .
build.openblas: build.openblas:
CMAKE_ARGS="-DLLAMA_OPENBLAS=on" FORCE_CMAKE=1 python3 setup.py develop python3 -m pip install --upgrade pip
CMAKE_ARGS="-DLLAMA_OPENBLAS=on" FORCE_CMAKE=1 python3 -m pip install --verbose --editable .
build.blis: build.blis:
CMAKE_ARGS="-DLLAMA_OPENBLAS=on -DLLAMA_OPENBLAS_VENDOR=blis" FORCE_CMAKE=1 python3 setup.py develop python3 -m pip install --upgrade pip
CMAKE_ARGS="-DLLAMA_OPENBLAS=on -DLLAMA_OPENBLAS_VENDOR=blis" FORCE_CMAKE=1 python3 -m pip install --verbose --editable .
build.metal: build.metal:
CMAKE_ARGS="-DLLAMA_METAL=on" FORCE_CMAKE=1 python3 setup.py develop python3 -m pip install --upgrade pip
CMAKE_ARGS="-DLLAMA_METAL=on" FORCE_CMAKE=1 python3 -m pip install --verbose --editable .
build.sdist: build.sdist:
python3 setup.py sdist python3 -m pip install --upgrade pip build
python3 -m build --sdist
deploy.pypi: deploy.pypi:
python3 -m twine upload dist/* python3 -m twine upload dist/*
@ -36,7 +44,6 @@ deploy.gh-docs:
clean: clean:
- cd vendor/llama.cpp && make clean - cd vendor/llama.cpp && make clean
- cd vendor/llama.cpp && rm libllama.so - cd vendor/llama.cpp && rm libllama.so
- rm -rf _skbuild
- rm llama_cpp/*.so - rm llama_cpp/*.so
- rm llama_cpp/*.dylib - rm llama_cpp/*.dylib
- rm llama_cpp/*.dll - rm llama_cpp/*.dll

View file

@ -1,16 +1,47 @@
[tool.poetry] [build-system]
requires = [
"scikit-build-core>=0.4.4",
"cmake>=3.18",
"ninja",
]
build-backend = "scikit_build_core.build"
[project]
name = "llama_cpp_python" name = "llama_cpp_python"
version = "0.1.59" version = "0.1.59"
description = "Python bindings for the llama.cpp library" description = "Python bindings for the llama.cpp library"
authors = ["Andrei Betlen <abetlen@gmail.com>"]
license = "MIT"
readme = "README.md" readme = "README.md"
homepage = "https://github.com/abetlen/llama-cpp-python" license = { text = "MIT" }
repository = "https://github.com/abetlen/llama-cpp-python" authors = [
packages = [{include = "llama_cpp"}] { name = "Andrei Betlen", email = "abetlen@gmail.com" },
include = [
"LICENSE.md",
] ]
requires-python = ">=3.7"
dependencies = [
"typing-extensions>=4.5.0",
"numpy>=1.20.0",
"diskcache>=5.6.1",
]
classifiers = [
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.7",
"Programming Language :: Python :: 3.8",
"Programming Language :: Python :: 3.9",
"Programming Language :: Python :: 3.10",
"Programming Language :: Python :: 3.11",
]
[tool.scikit-build]
wheel.packages = ["llama_cpp", "llama_cpp.server"]
wheel.expand-macos-universal-tags = true
cmake.verbose = true
[project.optional-dependencies]
server = [
"uvicorn",
"fastapi",
"sse-starlette",
]
test = ["pytest", "httpx"]
[tool.poetry.dependencies] [tool.poetry.dependencies]
python = "^3.8.1" python = "^3.8.1"
@ -33,12 +64,3 @@ scikit-build = "0.17.6"
[tool.poetry.extras] [tool.poetry.extras]
server = ["uvicorn", "fastapi", "sse-starlette"] server = ["uvicorn", "fastapi", "sse-starlette"]
[build-system]
requires = [
"setuptools>=42",
"scikit-build>=0.13",
"cmake>=3.18",
"ninja",
]
build-backend = "setuptools.build_meta"

View file

@ -1,32 +0,0 @@
from skbuild import setup
from pathlib import Path
this_directory = Path(__file__).parent
long_description = (this_directory / "README.md").read_text(encoding="utf-8")
setup(
name="llama_cpp_python",
description="A Python wrapper for llama.cpp",
long_description=long_description,
long_description_content_type="text/markdown",
version="0.1.59",
author="Andrei Betlen",
author_email="abetlen@gmail.com",
license="MIT",
package_dir={"llama_cpp": "llama_cpp", "llama_cpp.server": "llama_cpp/server"},
packages=["llama_cpp", "llama_cpp.server"],
install_requires=["typing-extensions>=4.5.0", "numpy>=1.20.0", "diskcache>=5.6.1"],
extras_require={
"server": ["uvicorn>=0.21.1", "fastapi>=0.95.0", "sse-starlette>=1.3.3"],
},
python_requires=">=3.7",
classifiers=[
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.7",
"Programming Language :: Python :: 3.8",
"Programming Language :: Python :: 3.9",
"Programming Language :: Python :: 3.10",
"Programming Language :: Python :: 3.11",
],
)