mirror of
https://github.com/mudler/LocalAI.git
synced 2024-12-28 08:28:51 +00:00
06cd9ef98d
* feat(extra-backends): Improvements vllm: add max_tokens, wire up stream event mamba: fixups, adding examples for mamba-chat * examples(mamba-chat): add * docs: update
28 lines
1.1 KiB
YAML
28 lines
1.1 KiB
YAML
name: mamba-chat
|
|
backend: mamba
|
|
parameters:
|
|
model: "havenhq/mamba-chat"
|
|
|
|
trimsuffix:
|
|
- <|endoftext|>
|
|
|
|
# https://huggingface.co/HuggingFaceH4/zephyr-7b-beta/blob/main/tokenizer_config.json
|
|
# "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
|
|
template:
|
|
chat_message: |
|
|
{{if eq .RoleName "assistant"}}<|assistant|>{{else if eq .RoleName "system"}}<|system|>{{else if eq .RoleName "user"}}<|user|>{{end}}
|
|
{{if .Content}}{{.Content}}{{end}}
|
|
</s>
|
|
|
|
chat: |
|
|
{{.Input}}
|
|
<|assistant|>
|
|
|
|
completion: |
|
|
{{.Input}}
|
|
usage: |
|
|
curl http://localhost:8080/v1/chat/completions -H "Content-Type: application/json" -d '{
|
|
"model": "mamba-chat",
|
|
"messages": [{"role": "user", "content": "how are you doing"}],
|
|
"temperature": 0.7
|
|
}' |