LocalAI/api/backend
Ettore Di Giacinto 8ccf5b2044
feat(speculative-sampling): allow to specify a draft model in the model config (#1052)
**Description**

This PR fixes #1013.

It adds `draft_model` and `n_draft` to the model YAML config in order to
load models with speculative sampling. This should be compatible as well
with grammars.

example:

```yaml
backend: llama                                                                                                                                                                   
context_size: 1024                                                                                                                                                                        
name: my-model-name
parameters:
  model: foo-bar
n_draft: 16                                                                                                                                                                      
draft_model: model-name
```

---------

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2023-09-14 17:44:16 +02:00
..
embeddings.go feat: add --single-active-backend to allow only one backend active at the time (#925) 2023-08-19 01:49:33 +02:00
image.go feat(diffusers): Add lora (#965) 2023-08-27 10:11:16 +02:00
llm.go fix(utf8): prevent multi-byte utf8 characters from being mangled (#981) 2023-08-30 23:56:59 +00:00
options.go feat(speculative-sampling): allow to specify a draft model in the model config (#1052) 2023-09-14 17:44:16 +02:00
transcript.go fix: drop racy code, refactor and group API schema (#931) 2023-08-20 14:04:45 +02:00
tts.go feat: add --single-active-backend to allow only one backend active at the time (#925) 2023-08-19 01:49:33 +02:00