-
Notifications
You must be signed in to change notification settings - Fork 0
49 lines (43 loc) · 1.42 KB
/
publish-docker.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
name: Publish Docker images
on:
release:
types: [published]
workflow_dispatch:
permissions:
packages: write
contents: read
jobs:
build-llama-cpp:
uses: localagi/ai-dedicated-workflows/.github/workflows/operation-docker-build-publish.yml@v3
with:
dockerfile: Dockerfile
context-repository: abetlen/llama-cpp-python
context-repository-ref: ${{ github.ref_name }}
registry-repo-name: llama-cpp-python
tags: |
type=schedule
type=ref,event=branch
type=semver,pattern={{raw}}
flavor: |
suffix=${{ matrix.suffix }}
build-args: |
FROM_IMAGE=${{ matrix.from }}
${{ matrix.build-arg }}
# platforms: linux/amd64,linux/arm64/v8
secrets: inherit
strategy:
fail-fast: false
matrix:
# see https://hub.docker.com/r/nvidia/cuda/tags
include:
- blas: CUBLAS
build-arg: "CMAKE_ARGS=-DLLAMA_CUBLAS=on -DCMAKE_CUDA_ARCHITECTURES=61;86"
suffix: "-cublas-11.7.1"
from: nvidia/cuda:11.7.1-devel-ubuntu22.04
- blas: CUBLAS
build-arg: "CMAKE_ARGS=-DLLAMA_CUBLAS=on -DCMAKE_CUDA_ARCHITECTURES=61;86;89"
suffix: "-cublas-12.1.1"
from: nvidia/cuda:12.1.1-devel-ubuntu22.04
- blas: OPENBLAS
from: python:3.10-slim-bullseye
build-arg: "CMAKE_ARGS=-DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS"