diff --git a/.github/workflows/on-push.yml b/.github/workflows/on-push.yml index d1e7de9..4e84ebe 100644 --- a/.github/workflows/on-push.yml +++ b/.github/workflows/on-push.yml @@ -13,14 +13,14 @@ jobs: - run: docker login --username "${{ github.actor }}" --password ${{ secrets.GITHUB_TOKEN }} ghcr.io - run: docker buildx create --use # 7B - - run: docker buildx build --platform linux/amd64,linux/arm64 -f api/Dockerfile --tag ghcr.io/getumbrel/llama-gpt-api-llama-2-7b-chat:${{ github.sha }} --push . - - run: docker buildx build --platform linux/amd64,linux/arm64 -f api/Dockerfile --tag ghcr.io/getumbrel/llama-gpt-api-llama-2-7b-chat:latest --push . + - run: docker buildx build --platform linux/amd64,linux/arm64 -f ./api/Dockerfile --tag ghcr.io/getumbrel/llama-gpt-api-llama-2-7b-chat:${{ github.sha }} --push ./api + - run: docker buildx build --platform linux/amd64,linux/arm64 -f ./api/Dockerfile --tag ghcr.io/getumbrel/llama-gpt-api-llama-2-7b-chat:latest --push ./api # 13B - - run: docker buildx build --platform linux/amd64,linux/arm64 -f api/13B.Dockerfile --tag ghcr.io/getumbrel/llama-gpt-api-llama-2-13b-chat:${{ github.sha }} --push . - - run: docker buildx build --platform linux/amd64,linux/arm64 -f api/13B.Dockerfile --tag ghcr.io/getumbrel/llama-gpt-api-llama-2-13b-chat:latest --push . + - run: docker buildx build --platform linux/amd64,linux/arm64 -f ./api/13B.Dockerfile --tag ghcr.io/getumbrel/llama-gpt-api-llama-2-13b-chat:${{ github.sha }} --push ./api + - run: docker buildx build --platform linux/amd64,linux/arm64 -f ./api/13B.Dockerfile --tag ghcr.io/getumbrel/llama-gpt-api-llama-2-13b-chat:latest --push ./api # 70B - # - run: docker buildx build --platform linux/amd64,linux/arm64 -f api/70B.Dockerfile --tag ghcr.io/getumbrel/llama-gpt-api-llama-2-70b-chat:${{ github.sha }} --push . - # - run: docker buildx build --platform linux/amd64,linux/arm64 -f api/70B.Dockerfile --tag ghcr.io/getumbrel/llama-gpt-api-llama-2-70b-chat:latest --push . + # - run: docker buildx build --platform linux/amd64,linux/arm64 -f ./api/70B.Dockerfile --tag ghcr.io/getumbrel/llama-gpt-api-llama-2-70b-chat:${{ github.sha }} --push ./api + # - run: docker buildx build --platform linux/amd64,linux/arm64 -f ./api/70B.Dockerfile --tag ghcr.io/getumbrel/llama-gpt-api-llama-2-70b-chat:latest --push ./api build_ui: runs-on: ubuntu-latest @@ -28,5 +28,5 @@ jobs: - uses: actions/checkout@c85c95e3d7251135ab7dc9ce3241c5835cc595a9 # v3.5.3 - run: docker login --username "${{ github.actor }}" --password ${{ secrets.GITHUB_TOKEN }} ghcr.io - run: docker buildx create --use - - run: docker buildx build --platform linux/amd64,linux/arm64 -f ui/Dockerfile --tag ghcr.io/getumbrel/llama-gpt-ui:${{ github.sha }} --push . - - run: docker buildx build --platform linux/amd64,linux/arm64 -f ui/Dockerfile --tag ghcr.io/getumbrel/llama-gpt-ui:latest --push . + - run: docker buildx build --platform linux/amd64,linux/arm64 -f ./ui/Dockerfile --tag ghcr.io/getumbrel/llama-gpt-ui:${{ github.sha }} --push ./ui + - run: docker buildx build --platform linux/amd64,linux/arm64 -f ./ui/Dockerfile --tag ghcr.io/getumbrel/llama-gpt-ui:latest --push ./ui diff --git a/docker-compose.yml b/docker-compose.yml index e4d8abd..da087f4 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -2,10 +2,11 @@ version: '3.6' services: llama-gpt-api: - # image: 'ghcr.io/getumbrel/llama-gpt-api-llama-2-7b-chat:latest' - build: - context: ./api - dockerfile: Dockerfile + image: 'ghcr.io/getumbrel/llama-gpt-api-llama-2-7b-chat:latest' + # build: + # context: ./api + # dockerfile: Dockerfile + command: /bin/sh -c "make build && python3 -m llama_cpp.server --n_ctx 4096 --n_threads 8 --n_gpu_layers 0 --n_batch 2096" environment: MODEL: '/models/llama-2-7b-chat.bin'