LocalAI/backend/python/parler-tts
Ettore Di Giacinto 11d960b2a6
chore(cli): be consistent between workers and expose ExtraLLamaCPPArgs to both (#3428)
* chore(cli): be consistent between workers and expose ExtraLLamaCPPArgs to both

Fixes: https://github.com/mudler/LocalAI/issues/3427

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* bump grpcio

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

---------

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-30 00:10:17 +02:00
..
backend.py feat: create bash library to handle install/run/test of python backends (#2286) 2024-05-11 18:32:46 +02:00
install.sh fix: add missing openvino/optimum/etc libraries for Intel, fixes #2289 (#2292) 2024-05-12 09:01:45 +02:00
Makefile feat: create bash library to handle install/run/test of python backends (#2286) 2024-05-11 18:32:46 +02:00
requirements-after.txt fix(parler-tts): pin numba 2024-08-21 13:09:12 +02:00
requirements-cpu.txt fix(python): move accelerate and GPU-specific libs to build-type (#3194) 2024-08-07 17:02:32 +02:00
requirements-cublas11.txt fix(python): move accelerate and GPU-specific libs to build-type (#3194) 2024-08-07 17:02:32 +02:00
requirements-cublas12.txt fix(python): move accelerate and GPU-specific libs to build-type (#3194) 2024-08-07 17:02:32 +02:00
requirements-hipblas.txt fix(parler-tts): pin torchaudio and torch for hipblas 2024-08-21 18:27:20 +02:00
requirements-intel.txt fix(python): move accelerate and GPU-specific libs to build-type (#3194) 2024-08-07 17:02:32 +02:00
requirements.txt chore(cli): be consistent between workers and expose ExtraLLamaCPPArgs to both (#3428) 2024-08-30 00:10:17 +02:00
run.sh feat: create bash library to handle install/run/test of python backends (#2286) 2024-05-11 18:32:46 +02:00
test.py feat: create bash library to handle install/run/test of python backends (#2286) 2024-05-11 18:32:46 +02:00
test.sh feat: create bash library to handle install/run/test of python backends (#2286) 2024-05-11 18:32:46 +02:00