LocalAI/backend/python/exllama2
Ludovic Leroux 12c0d9443e
feat: use tokenizer.apply_chat_template() in vLLM (#1990)
Use tokenizer.apply_chat_template() in vLLM

Signed-off-by: Ludovic LEROUX <ludovic@inpher.io>
2024-04-11 19:20:22 +02:00
..
backend_pb2_grpc.py feat: use tokenizer.apply_chat_template() in vLLM (#1990) 2024-04-11 19:20:22 +02:00
backend_pb2.py feat: use tokenizer.apply_chat_template() in vLLM (#1990) 2024-04-11 19:20:22 +02:00
exllama2_backend.py fix: exllama2 backend (#1484) 2023-12-24 08:32:12 +00:00
exllama2.yml exllama(v2): fix exllamav1, add exllamav2 (#1384) 2023-12-05 08:15:37 +01:00
install.sh feat(intel): add diffusers/transformers support (#1746) 2024-03-07 14:37:45 +01:00
Makefile deps(conda): use transformers-env with vllm,exllama(2) (#1554) 2024-01-06 13:32:28 +01:00
run.sh deps(conda): use transformers-env with vllm,exllama(2) (#1554) 2024-01-06 13:32:28 +01:00