mirror of
https://github.com/mudler/LocalAI.git
synced 2025-03-14 00:06:53 +00:00
* feat(aio): update AIO image defaults cpu: - text-to-text: llama3.1 - embeddings: granite-embeddings - vision: moonream2 gpu/intel: - text-to-text: localai-functioncall-qwen2.5-7b-v0.5 - embeddings: granite-embeddings - vision: minicpm Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * feat(aio): use minicpm as moondream2 stopped working https://github.com/ggml-org/llama.cpp/pull/12322#issuecomment-2717483759 Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
AIO CPU size
Use this image with CPU-only.
Please keep using only C++ backends so the base image is as small as possible (without CUDA, cuDNN, python, etc).