---
name: "llava"

config_file: |
  backend: llama-cpp
  context_size: 4096
  f16: true

  mmap: true
  roles:
    user: "USER:"
    assistant: "ASSISTANT:"
    system: "SYSTEM:"

  template:
    chat: |
      A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions.
      {{.Input}}
      ASSISTANT: