LocalAI/api
Ettore Di Giacinto 1120847f72
feat: bump llama.cpp, add gguf support (#943)
**Description**

This PR syncs up the `llama` backend to use `gguf`
(https://github.com/go-skynet/go-llama.cpp/pull/180). It also adds
`llama-stable` to the targets so we can still load ggml. It adapts the
current tests to use the `llama-backend` for ggml and uses a `gguf`
model to run tests on the new backend.

In order to consume the new version of go-llama.cpp, it also bump go to
1.21 (images, pipelines, etc)

---------

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2023-08-24 01:18:58 +02:00
..
backend Feat: rwkv improvements: (#937) 2023-08-22 18:48:06 +02:00
config Feat: rwkv improvements: (#937) 2023-08-22 18:48:06 +02:00
localai feat: backend monitor shutdown endpoint, process based (#938) 2023-08-23 18:38:37 +02:00
openai fix: drop racy code, refactor and group API schema (#931) 2023-08-20 14:04:45 +02:00
options feat: add --single-active-backend to allow only one backend active at the time (#925) 2023-08-19 01:49:33 +02:00
schema fix: drop racy code, refactor and group API schema (#931) 2023-08-20 14:04:45 +02:00
api_test.go feat: bump llama.cpp, add gguf support (#943) 2023-08-24 01:18:58 +02:00
api.go feat: backend monitor shutdown endpoint, process based (#938) 2023-08-23 18:38:37 +02:00
apt_suite_test.go feat: add CI/tests (#58) 2023-04-22 00:44:52 +02:00