LocalAI/backend/python/exllama2
Ettore Di Giacinto e19d7226f8
feat: more embedded models, coqui fixes, add model usage and description (#1556)
* feat: add model descriptions and usage

* remove default model gallery

* models: add embeddings and tts

* docs: update table

* docs: updates

* images: cleanup pip cache after install

* images: always run apt-get clean

* ux: improve gRPC connection errors

* ux: improve some messages

* fix: fix coqui when no AudioPath is passed by

* embedded: add more models

* Add usage

* Reorder table
2024-01-08 00:37:02 +01:00
..
backend_pb2_grpc.py exllama(v2): fix exllamav1, add exllamav2 (#1384) 2023-12-05 08:15:37 +01:00
backend_pb2.py feat(diffusers): update, add autopipeline, controlnet (#1432) 2023-12-13 19:20:22 +01:00
exllama2_backend.py fix: exllama2 backend (#1484) 2023-12-24 08:32:12 +00:00
exllama2.yml exllama(v2): fix exllamav1, add exllamav2 (#1384) 2023-12-05 08:15:37 +01:00
install.sh feat: more embedded models, coqui fixes, add model usage and description (#1556) 2024-01-08 00:37:02 +01:00
Makefile deps(conda): use transformers-env with vllm,exllama(2) (#1554) 2024-01-06 13:32:28 +01:00
run.sh deps(conda): use transformers-env with vllm,exllama(2) (#1554) 2024-01-06 13:32:28 +01:00