|
13 | 13 | MODEL_NAME: "codellama-7b.Q2_K.gguf"
|
14 | 14 | jobs:
|
15 | 15 |
|
16 |
| - build-linux-cuda: |
17 |
| - name: Build Linux x86-64 CUDA12 |
18 |
| - runs-on: ubuntu-latest |
19 |
| - steps: |
20 |
| - - uses: actions/checkout@v4 |
21 |
| - - name: Build libraries |
22 |
| - shell: bash |
23 |
| - run: | |
24 |
| - .github/dockcross/dockcross-manylinux_2_28-x64 .github/build_cuda_linux.sh "-DOS_NAME=Linux -DOS_ARCH=x86_64" |
25 |
| - - name: Upload artifacts |
26 |
| - uses: actions/upload-artifact@v4 |
27 |
| - with: |
28 |
| - name: linux-libraries-cuda |
29 |
| - path: ${{ github.workspace }}/src/main/resources_linux_cuda/de/kherud/llama/ |
| 16 | +# todo: doesn't work with the newest llama.cpp version |
| 17 | +# build-linux-cuda: |
| 18 | +# name: Build Linux x86-64 CUDA12 |
| 19 | +# runs-on: ubuntu-latest |
| 20 | +# steps: |
| 21 | +# - uses: actions/checkout@v4 |
| 22 | +# - name: Build libraries |
| 23 | +# shell: bash |
| 24 | +# run: | |
| 25 | +# .github/dockcross/dockcross-manylinux_2_28-x64 .github/build_cuda_linux.sh "-DOS_NAME=Linux -DOS_ARCH=x86_64" |
| 26 | +# - name: Upload artifacts |
| 27 | +# uses: actions/upload-artifact@v4 |
| 28 | +# with: |
| 29 | +# name: linux-libraries-cuda |
| 30 | +# path: ${{ github.workspace }}/src/main/resources_linux_cuda/de/kherud/llama/ |
30 | 31 |
|
31 | 32 | build-linux-docker:
|
32 | 33 | name: Build ${{ matrix.target.os }}-${{ matrix.target.arch }}
|
@@ -194,7 +195,7 @@ jobs:
|
194 | 195 |
|
195 | 196 | publish:
|
196 | 197 | if: ${{ github.event_name != 'workflow_dispatch' || github.event.inputs.build_only == 'no' }}
|
197 |
| - needs: [ test-linux,build-macos-native,build-win-native,build-linux-cuda ] |
| 198 | + needs: [ test-linux,build-macos-native,build-win-native ] #,build-linux-cuda |
198 | 199 | runs-on: ubuntu-latest
|
199 | 200 | steps:
|
200 | 201 | - uses: actions/checkout@v4
|
|
0 commit comments