LocalAI/examples/configurations/phi-2.yaml

30 lines
686 B
YAML
Raw Normal View History

2024-01-01 09:51:47 +00:00
name: phi-2
context_size: 2048
f16: true
gpu_layers: 90
mmap: true
2024-01-01 13:39:42 +00:00
trimsuffix:
- "\n"
2024-01-01 09:51:47 +00:00
parameters:
model: huggingface://TheBloke/phi-2-GGUF/phi-2.Q8_0.gguf
temperature: 0.2
top_k: 40
top_p: 0.95
seed: -1
mirostat: 2
mirostat_eta: 1.0
mirostat_tau: 1.0
2024-01-01 09:51:47 +00:00
template:
chat: &template |-
2024-01-01 09:51:47 +00:00
Instruct: {{.Input}}
Output:
completion: *template
usage: |
To use this model, interact with the API (in another terminal) with curl for instance:
curl http://localhost:8080/v1/chat/completions -H "Content-Type: application/json" -d '{
"model": "phi-2",
"messages": [{"role": "user", "content": "How are you doing?", "temperature": 0.1}]
}'