2024-01-25 23:13:21 +00:00
|
|
|
name: codellama-7b-gguf
|
|
|
|
backend: transformers
|
|
|
|
parameters:
|
|
|
|
model: huggingface://TheBloke/CodeLlama-7B-GGUF/codellama-7b.Q4_K_M.gguf
|
2024-03-11 18:49:03 +00:00
|
|
|
temperature: 0.5
|
2024-01-25 23:13:21 +00:00
|
|
|
top_k: 40
|
|
|
|
seed: -1
|
|
|
|
top_p: 0.95
|
2024-03-11 18:49:03 +00:00
|
|
|
mirostat: 2
|
|
|
|
mirostat_eta: 1.0
|
|
|
|
mirostat_tau: 1.0
|
|
|
|
|
2024-01-25 23:13:21 +00:00
|
|
|
context_size: 4096
|
|
|
|
f16: true
|
|
|
|
gpu_layers: 90
|
|
|
|
usage: |
|
|
|
|
curl http://localhost:8080/v1/completions -H "Content-Type: application/json" -d '{
|
|
|
|
"model": "codellama-7b-gguf",
|
|
|
|
"prompt": "import socket\n\ndef ping_exponential_backoff(host: str):"
|
|
|
|
}'
|