LocalAI/pkg
Ettore Di Giacinto 092bb0bd6b
fix(base-grpc): close channel in base grpc server (#3734)
If the LLM does not implement any logic for PredictStream, we close the
channel immediately to not leave the process hanging.

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-10-05 15:14:27 +02:00
..
assets chore: fix go.mod module (#2635) 2024-06-23 08:24:36 +00:00
concurrency groundwork: add pkg/concurrency and the associated test file (#2745) 2024-07-18 23:29:21 +00:00
downloader test: preliminary tests and merge fix for authv2 (#3584) 2024-09-24 09:32:48 +02:00
functions feat(openai): add json_schema format type and strict mode (#3193) 2024-08-07 15:27:02 -04:00
grpc fix(base-grpc): close channel in base grpc server (#3734) 2024-10-05 15:14:27 +02:00
langchain feat(llama.cpp): do not specify backends to autoload and add llama.cpp variants (#2232) 2024-05-04 17:56:12 +02:00
library rf: centralize base64 image handling (#2595) 2024-06-24 08:34:36 +02:00
model feat(shutdown): allow force shutdown of backends (#3733) 2024-10-05 10:41:35 +02:00
oci chore: fix go.mod module (#2635) 2024-06-23 08:24:36 +00:00
stablediffusion feat: support upscaled image generation with esrgan (#509) 2023-06-05 17:21:38 +02:00
startup fix: be consistent in downloading files, check for scanner errors (#3108) 2024-08-02 20:06:25 +02:00
store chore: fix go.mod module (#2635) 2024-06-23 08:24:36 +00:00
templates feat(multimodal): allow to template placeholders (#3728) 2024-10-04 18:32:29 +02:00
tinydream feat: add tiny dream stable diffusion support (#1283) 2023-12-24 19:27:24 +00:00
utils feat(api): allow to pass videos to backends (#3601) 2024-09-19 11:21:59 +02:00
xsync chore: fix go.mod module (#2635) 2024-06-23 08:24:36 +00:00
xsysinfo feat(default): use number of physical cores as default (#2483) 2024-06-04 15:23:29 +02:00