LocalAI/pkg
Ettore Di Giacinto 8814b31805
chore: drop gpt4all.cpp (#3106)
chore: drop gpt4all

gpt4all is already supported in llama.cpp - the backend was kept for
keeping compatibility with old gpt4all models (prior to gguf format).

It is good time now to clean up and remove it to slim the compilation
process.

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-07 23:35:55 +02:00
..
assets chore: fix go.mod module (#2635) 2024-06-23 08:24:36 +00:00
concurrency groundwork: add pkg/concurrency and the associated test file (#2745) 2024-07-18 23:29:21 +00:00
downloader fix: be consistent in downloading files, check for scanner errors (#3108) 2024-08-02 20:06:25 +02:00
functions feat(openai): add json_schema format type and strict mode (#3193) 2024-08-07 15:27:02 -04:00
grpc chore: fix go.mod module (#2635) 2024-06-23 08:24:36 +00:00
langchain feat(llama.cpp): do not specify backends to autoload and add llama.cpp variants (#2232) 2024-05-04 17:56:12 +02:00
library rf: centralize base64 image handling (#2595) 2024-06-24 08:34:36 +02:00
model chore: drop gpt4all.cpp (#3106) 2024-08-07 23:35:55 +02:00
oci chore: fix go.mod module (#2635) 2024-06-23 08:24:36 +00:00
stablediffusion feat: support upscaled image generation with esrgan (#509) 2023-06-05 17:21:38 +02:00
startup fix: be consistent in downloading files, check for scanner errors (#3108) 2024-08-02 20:06:25 +02:00
store chore: fix go.mod module (#2635) 2024-06-23 08:24:36 +00:00
templates chore: fix go.mod module (#2635) 2024-06-23 08:24:36 +00:00
tinydream feat: add tiny dream stable diffusion support (#1283) 2023-12-24 19:27:24 +00:00
utils fix(gallery): do not attempt to delete duplicate files (#3031) 2024-07-28 10:27:56 +02:00
xsync chore: fix go.mod module (#2635) 2024-06-23 08:24:36 +00:00
xsysinfo feat(default): use number of physical cores as default (#2483) 2024-06-04 15:23:29 +02:00