LocalAI/backend/python/mamba
Ettore Di Giacinto 11d960b2a6
chore(cli): be consistent between workers and expose ExtraLLamaCPPArgs to both (#3428)
* chore(cli): be consistent between workers and expose ExtraLLamaCPPArgs to both

Fixes: https://github.com/mudler/LocalAI/issues/3427

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* bump grpcio

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

---------

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2024-08-30 00:10:17 +02:00
..
backend.py feat: create bash library to handle install/run/test of python backends (#2286) 2024-05-11 18:32:46 +02:00
install.sh feat: create bash library to handle install/run/test of python backends (#2286) 2024-05-11 18:32:46 +02:00
Makefile feat: create bash library to handle install/run/test of python backends (#2286) 2024-05-11 18:32:46 +02:00
README.md feat: 🐍 add mamba support (#1589) 2024-01-19 23:42:50 +01:00
requirements-after.txt fix: ensure correct version of torch is always installed based on BUILD_TYPE(#2890) 2024-08-05 16:38:33 +00:00
requirements-cpu.txt fix(python): move accelerate and GPU-specific libs to build-type (#3194) 2024-08-07 17:02:32 +02:00
requirements-cublas11.txt fix(python): move accelerate and GPU-specific libs to build-type (#3194) 2024-08-07 17:02:32 +02:00
requirements-cublas12.txt fix(python): move accelerate and GPU-specific libs to build-type (#3194) 2024-08-07 17:02:32 +02:00
requirements-install.txt fix: ensure correct version of torch is always installed based on BUILD_TYPE(#2890) 2024-08-05 16:38:33 +00:00
requirements.txt chore(cli): be consistent between workers and expose ExtraLLamaCPPArgs to both (#3428) 2024-08-30 00:10:17 +02:00
run.sh feat: create bash library to handle install/run/test of python backends (#2286) 2024-05-11 18:32:46 +02:00
test.py feat: create bash library to handle install/run/test of python backends (#2286) 2024-05-11 18:32:46 +02:00
test.sh feat: create bash library to handle install/run/test of python backends (#2286) 2024-05-11 18:32:46 +02:00

Creating a separate environment for the mamba project

make mamba