LocalAI/backend
Ettore Di Giacinto c56b6ddb1c
fix(llama.cpp): disable infinite context shifting (#1704)
Infinite context loop might as well trigger an infinite loop of context
shifting if the model hallucinates and does not stop answering.
This has the unpleasant effect that the predicion never terminates,
which is the case especially on small models which tends to hallucinate.

Workarounds https://github.com/mudler/LocalAI/issues/1333 by removing
context-shifting.

See also upstream issue: https://github.com/ggerganov/llama.cpp/issues/3969
2024-02-13 21:17:21 +01:00
..
cpp fix(llama.cpp): disable infinite context shifting (#1704) 2024-02-13 21:17:21 +01:00
go fix: drop unused code (#1697) 2024-02-11 11:28:59 +01:00
python fix(vall-e-x): Fix voice cloning (#1696) 2024-02-11 11:20:00 +01:00
backend_grpc.pb.go transformers: correctly load automodels (#1643) 2024-01-26 00:13:21 +01:00
backend.proto transformers: correctly load automodels (#1643) 2024-01-26 00:13:21 +01:00