diff --git a/.github/workflows/build-publish.yaml b/.github/workflows/build-publish.yaml index e755415..8632b92 100644 --- a/.github/workflows/build-publish.yaml +++ b/.github/workflows/build-publish.yaml @@ -53,13 +53,6 @@ jobs: sudo apt-get update sudo apt-get install -y qemu-user-static - - name: Clone llama-cpp-python - run: | - git config --global user.email "podman-deskop@redhat.com" - git config --global user.name "Podman Desktop team" - cd chat - ./setup.sh - - name: Build Playground Image id: build-image uses: redhat-actions/buildah-build@v2 diff --git a/.github/workflows/pr-check.yaml b/.github/workflows/pr-check.yaml index 1e658f2..e990f45 100644 --- a/.github/workflows/pr-check.yaml +++ b/.github/workflows/pr-check.yaml @@ -49,13 +49,6 @@ jobs: sudo apt-get update sudo apt-get install -y qemu-user-static - - name: Clone llama-cpp-python - run: | - git config --global user.email "podman-deskop@redhat.com" - git config --global user.name "Podman Desktop team" - cd chat - ./setup.sh - - name: Build Playground Image id: build-image uses: redhat-actions/buildah-build@v2 diff --git a/.github/workflows/release.yaml b/.github/workflows/release.yaml index d4177a5..2f8d840 100644 --- a/.github/workflows/release.yaml +++ b/.github/workflows/release.yaml @@ -89,13 +89,6 @@ jobs: sudo apt-get update sudo apt-get install -y qemu-user-static - - name: Clone llama-cpp-python - run: | - git config --global user.email "podman-deskop@redhat.com" - git config --global user.name "Podman Desktop team" - cd chat - ./setup.sh - - name: Build Playground Image id: build-image uses: redhat-actions/buildah-build@v2 @@ -137,4 +130,4 @@ jobs: with: tag: ${{ needs.tag.outputs.githubTag }} name: ${{ needs.tag.outputs.githubTag }} - + \ No newline at end of file diff --git a/.gitignore b/.gitignore deleted file mode 100644 index 8a64383..0000000 --- a/.gitignore +++ /dev/null @@ -1 +0,0 @@ -llama-cpp-python diff --git a/chat/base/Containerfile b/chat/base/Containerfile index a4b2597..131cf56 100644 --- a/chat/base/Containerfile +++ b/chat/base/Containerfile @@ -21,7 +21,6 @@ RUN microdnf install -y python3 python3-pip gcc g++ shadow-utils && microdnf cle RUN useradd -r -g root -m -d /home/default -s /bin/bash default WORKDIR /home/default COPY requirements.txt requirements.txt -COPY llama-cpp-python llama-cpp-python RUN pip install --no-cache-dir --upgrade -r requirements.txt USER default COPY run.sh run.sh diff --git a/chat/cuda/amd64/Containerfile b/chat/cuda/amd64/Containerfile index bf14595..86d94c3 100644 --- a/chat/cuda/amd64/Containerfile +++ b/chat/cuda/amd64/Containerfile @@ -20,7 +20,6 @@ RUN dnf install -y gcc-toolset-13-gcc gcc-toolset-13-gcc-c++ USER 1001 WORKDIR /locallm COPY requirements.txt . -COPY --chown=1001 llama-cpp-python llama-cpp-python COPY run.sh . ENV CMAKE_ARGS="-DGGML_CUDA=on -DLLAMA_AVX2=OFF -DLLAMA_FMA=OFF -DLLAMA_F16C=OFF" ENV FORCE_CMAKE=1 diff --git a/chat/requirements.txt b/chat/requirements.txt index c9f91c6..3b71459 100644 --- a/chat/requirements.txt +++ b/chat/requirements.txt @@ -1,2 +1,2 @@ -./llama-cpp-python[server] +llama-cpp-python[server]==0.3.7 pip==25.0.1 diff --git a/chat/setup.sh b/chat/setup.sh deleted file mode 100755 index 19cdb12..0000000 --- a/chat/setup.sh +++ /dev/null @@ -1,6 +0,0 @@ -git clone https://github.com/abetlen/llama-cpp-python --no-checkout --recurse-submodules -cd llama-cpp-python -git checkout v0.3.2 -git submodule update -git fetch origin pull/1552/head -git rebase ffc47e5a95cf31fe6d9d6d7953d5eb17fdcc8513 diff --git a/chat/vulkan/amd64/Containerfile b/chat/vulkan/amd64/Containerfile index da1ed4f..08c88cb 100644 --- a/chat/vulkan/amd64/Containerfile +++ b/chat/vulkan/amd64/Containerfile @@ -32,7 +32,6 @@ RUN dnf install -y python3-dnf-plugin-versionlock && \ WORKDIR /locallm COPY requirements.txt ./ -COPY llama-cpp-python llama-cpp-python RUN pip install --upgrade pip ENV CMAKE_ARGS="-DGGML_VULKAN=on" ENV FORCE_CMAKE=1 diff --git a/chat/vulkan/arm64/Containerfile b/chat/vulkan/arm64/Containerfile index 5f66182..fb6aedf 100644 --- a/chat/vulkan/arm64/Containerfile +++ b/chat/vulkan/arm64/Containerfile @@ -30,7 +30,6 @@ https://dl.fedoraproject.org/pub/epel/epel-release-latest-9.noarch.rpm && \ dnf install -y glslc && \ dnf clean all COPY requirements.txt ./ -COPY llama-cpp-python llama-cpp-python RUN pip install --upgrade pip ENV CMAKE_ARGS="-DGGML_VULKAN=on" ENV FORCE_CMAKE=1