-
Notifications
You must be signed in to change notification settings - Fork 65
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Merge pull request #102 from acon96/release/v0.2.11
Release v0.2.11
- Loading branch information
Showing
14 changed files
with
612 additions
and
165 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,88 @@ | ||
name: Create Release | ||
|
||
on: | ||
workflow_dispatch: | ||
inputs: | ||
release_notes: | ||
description: "Release Notes" | ||
required: true | ||
type: string | ||
|
||
permissions: | ||
contents: write | ||
|
||
jobs: | ||
build_wheels: | ||
name: Build wheels on ${{ matrix.arch }} (HA ${{ matrix.home_assistant_version }}) | ||
runs-on: ubuntu-latest | ||
strategy: | ||
fail-fast: false | ||
matrix: | ||
home_assistant_version: ["2023.12.4", "2024.2.1"] | ||
arch: ["aarch64", "armhf", "amd64", "i386"] | ||
|
||
steps: | ||
- name: Checkout code | ||
uses: actions/checkout@v4 | ||
|
||
- name: Read llama-cpp-python version | ||
run: cat custom_components/llama_conversation/const.py | grep "EMBEDDED_LLAMA_CPP_PYTHON_VERSION" | tr -d ' ' | tr -d '"' >> $GITHUB_ENV | ||
|
||
- name: Build artifact | ||
uses: uraimo/run-on-arch-action@v2 | ||
id: build | ||
with: | ||
arch: none | ||
distro: none | ||
base_image: homeassistant/${{ matrix.arch }}-homeassistant:${{ matrix.home_assistant_version }} | ||
|
||
# Create an artifacts directory | ||
setup: | | ||
mkdir -p "${PWD}/artifacts" | ||
# Mount the artifacts directory as /artifacts in the container | ||
dockerRunArgs: | | ||
--volume "${PWD}/artifacts:/artifacts" | ||
# The shell to run commands with in the container | ||
shell: /bin/bash | ||
|
||
# Produce a binary artifact and place it in the mounted volume | ||
run: | | ||
apk update | ||
apk add build-base python3-dev cmake | ||
pip3 install build | ||
cd /tmp | ||
git clone --quiet --recurse-submodules https://github.com/abetlen/llama-cpp-python --branch "v${{ env.EMBEDDED_LLAMA_CPP_PYTHON_VERSION }}" | ||
cd llama-cpp-python | ||
export CMAKE_ARGS="-DLLAVA_BUILD=OFF" | ||
python3 -m build --wheel | ||
cp -f ./dist/*.whl /artifacts/ | ||
- name: Upload artifacts | ||
uses: actions/upload-artifact@v4 | ||
with: | ||
path: ./artifacts/*.whl | ||
name: artifact_${{ matrix.arch }}_${{ matrix.home_assistant_version }} | ||
|
||
release: | ||
name: Create Release | ||
needs: [ build_wheels ] | ||
runs-on: ubuntu-latest | ||
if: "startsWith(github.event.ref, 'refs/tags/v')" # only create a release if this was run on a tag | ||
|
||
steps: | ||
- name: Download artifacts | ||
uses: actions/download-artifact@v4 | ||
with: | ||
path: dist | ||
merge-multiple: true | ||
|
||
- name: Create GitHub release | ||
uses: softprops/action-gh-release@v2 | ||
with: | ||
files: dist/* | ||
body: ${{ inputs.release_notes }} | ||
make_latest: true |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,4 +1,4 @@ | ||
# text-generation-webui - Home Assistant Addon | ||
NOTE: This is super experimental and may or may not work on a Raspberry Pi | ||
Installs text-generation-webui into a docker container using CPU only mode (llama.cpp) | ||
|
||
Installs text-generation-webui into a docker container using CPU only mode (llama.cpp) | ||
NOTE: This addon is not the preferred way to run LLama.cpp as part of Home Assistant and will not be updated. |
Oops, something went wrong.