LocalAI/pkg
Ettore Di Giacinto 530bec9c64
feat(llama.cpp): do not specify backends to autoload and add llama.cpp variants (#2232)
* feat(initializer): do not specify backends to autoload

We can simply try to autoload the backends extracted in the asset dir.
This will allow to build variants of the same backend (for e.g. with different instructions sets),
so to have a single binary for all the variants.

Signed-off-by: mudler <mudler@localai.io>

* refactor(prepare): refactor out llama.cpp prepare steps

Make it so are idempotent and that we can re-build

Signed-off-by: mudler <mudler@localai.io>

* [TEST] feat(build): build noavx version along

Signed-off-by: mudler <mudler@localai.io>

* build: make build parallel

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* build: do not override CMAKE_ARGS

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* build: add fallback variant

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* Fixups

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* fix(huggingface-langchain): fail if no token is set

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* fix(huggingface-langchain): rename

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* fix: do not autoload local-store

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* fix: give priority between the listed backends

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

---------

Signed-off-by: mudler <mudler@localai.io>
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-05-04 17:56:12 +02:00
..
assets fix: security scanner warning noise: error handlers part 1 (#2141) 2024-04-26 10:34:31 +02:00
downloader fix: reduce chmod permissions for created files and directories (#2137) 2024-04-26 00:47:06 +02:00
functions fix: security scanner warning noise: error handlers part 2 (#2145) 2024-04-29 15:11:42 +02:00
gallery feat(gallery): support model deletion (#2173) 2024-04-28 23:42:46 +02:00
grpc refactor(application): introduce application global state (#2072) 2024-04-29 17:42:37 +00:00
langchain feat(llama.cpp): do not specify backends to autoload and add llama.cpp variants (#2232) 2024-05-04 17:56:12 +02:00
model feat(llama.cpp): do not specify backends to autoload and add llama.cpp variants (#2232) 2024-05-04 17:56:12 +02:00
stablediffusion feat: support upscaled image generation with esrgan (#509) 2023-06-05 17:21:38 +02:00
startup feat: Galleries UI (#2104) 2024-04-23 09:22:58 +02:00
store feat(stores): Vector store backend (#1795) 2024-03-22 21:14:04 +01:00
templates fix: reduce chmod permissions for created files and directories (#2137) 2024-04-26 00:47:06 +02:00
tinydream feat: add tiny dream stable diffusion support (#1283) 2023-12-24 19:27:24 +00:00
utils refactor(application): introduce application global state (#2072) 2024-04-29 17:42:37 +00:00
xsync feat(gallery): display job status also during navigation (#2151) 2024-04-27 09:08:33 +02:00