--- name: "llava" config_file: | backend: llama-cpp context_size: 4096 f16: true mmap: true roles: user: "USER:" assistant: "ASSISTANT:" system: "SYSTEM:" template: chat: | A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions. {{.Input}} ASSISTANT: