name: "codellama" config_file: | backend: llama-cpp context_size: 4096 f16: true mmap: true