mirror of
https://github.com/mudler/LocalAI.git
synced 2024-12-18 20:27:57 +00:00
Drop fat images, will document how to consume models
This commit is contained in:
parent
650a22aef1
commit
c61b023bc8
12
Earthfile
12
Earthfile
@ -11,11 +11,6 @@ go-deps:
|
|||||||
SAVE ARTIFACT go.mod AS LOCAL go.mod
|
SAVE ARTIFACT go.mod AS LOCAL go.mod
|
||||||
SAVE ARTIFACT go.sum AS LOCAL go.sum
|
SAVE ARTIFACT go.sum AS LOCAL go.sum
|
||||||
|
|
||||||
model-image:
|
|
||||||
ARG MODEL_IMAGE=quay.io/go-skynet/models:ggml2-alpaca-7b-v0.2
|
|
||||||
FROM $MODEL_IMAGE
|
|
||||||
SAVE ARTIFACT /models/model.bin
|
|
||||||
|
|
||||||
build:
|
build:
|
||||||
FROM +go-deps
|
FROM +go-deps
|
||||||
WORKDIR /build
|
WORKDIR /build
|
||||||
@ -34,14 +29,13 @@ image:
|
|||||||
ENTRYPOINT [ "/llama-cli" ]
|
ENTRYPOINT [ "/llama-cli" ]
|
||||||
SAVE IMAGE --push $IMAGE
|
SAVE IMAGE --push $IMAGE
|
||||||
|
|
||||||
lite-image:
|
image:
|
||||||
FROM +go-deps
|
FROM +go-deps
|
||||||
ARG IMAGE=alpaca-cli-nomodel
|
ARG IMAGE=alpaca-cli-nomodel
|
||||||
COPY +build/llama-cli /llama-cli
|
COPY +build/llama-cli /llama-cli
|
||||||
ENV MODEL_PATH=/model.bin
|
ENV MODEL_PATH=/model.bin
|
||||||
ENTRYPOINT [ "/llama-cli" ]
|
ENTRYPOINT [ "/llama-cli" ]
|
||||||
SAVE IMAGE --push $IMAGE-lite
|
SAVE IMAGE --push $IMAGE
|
||||||
|
|
||||||
image-all:
|
image-all:
|
||||||
#BUILD --platform=linux/amd64 --platform=linux/arm64 +image
|
BUILD --platform=linux/amd64 --platform=linux/arm64 +image
|
||||||
BUILD --platform=linux/amd64 --platform=linux/arm64 +lite-image
|
|
||||||
|
Loading…
Reference in New Issue
Block a user