--- name: "vicuna-chat" description: | Vicuna chat license: "LLaMA" config_file: | backend: llama-cpp context_size: 4096 roles: user: "User: " system: "System: " assistant: "Assistant: " f16: true template: completion: | Complete the following sentence: {{.Input}} chat: | {{.Input}} ASSISTANT: