LocalAI/gallery/vicuna-chat.yaml

27 lines
442 B
YAML
Raw Normal View History

---
name: "vicuna-chat"
description: |
Vicuna chat
license: "LLaMA"
config_file: |
backend: llama-cpp
context_size: 4096
roles:
user: "User: "
system: "System: "
assistant: "Assistant: "
f16: true
stopwords:
- <|end|>
- <|endoftext|>
- <eos>
template:
completion: |
Complete the following sentence: {{.Input}}
chat: |
{{.Input}}
ASSISTANT: