LocalAI/.github/workflows
Ettore Di Giacinto c89271b2e4
feat(llama.cpp): add distributed llama.cpp inferencing (#2324)
* feat(llama.cpp): support distributed llama.cpp

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* feat: let tweak how chat messages are merged together

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* refactor

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* Makefile: register to ALL_GRPC_BACKENDS

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* refactoring, allow disable auto-detection of backends

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

* minor fixups

Signed-off-by: mudler <mudler@localai.io>

* feat: add cmd to start rpc-server from llama.cpp

Signed-off-by: mudler <mudler@localai.io>

* ci: add ccache

Signed-off-by: mudler <mudler@localai.io>

---------

Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Signed-off-by: mudler <mudler@localai.io>
2024-05-15 01:17:02 +02:00
..
disabled feat(conda): conda environments (#1144) 2023-11-04 15:30:32 +01:00
bump_deps.yaml build(deps): bump peter-evans/create-pull-request from 5 to 6 (#2005) 2024-04-11 13:58:04 +00:00
bump_docs.yaml build(deps): bump peter-evans/create-pull-request from 5 to 6 (#2005) 2024-04-11 13:58:04 +00:00
checksum_checker.yaml ci: get latest git version 2024-05-09 11:33:16 +02:00
dependabot_auto.yml build(deps): bump dependabot/fetch-metadata from 2.0.0 to 2.1.0 (#2186) 2024-04-29 21:12:37 +00:00
generate_grpc_cache.yaml fix: bring everything onto the same GRPC version to fix tests (#2199) 2024-04-30 19:12:15 +00:00
image_build.yml fix: bring everything onto the same GRPC version to fix tests (#2199) 2024-04-30 19:12:15 +00:00
image-pr.yml feat: update ROCM and use smaller image (#2196) 2024-05-03 18:46:49 +02:00
image.yml feat: update ROCM and use smaller image (#2196) 2024-05-03 18:46:49 +02:00
labeler.yml fix(seed): generate random seed per-request if -1 is set (#1952) 2024-04-03 22:25:47 +02:00
localaibot_automerge.yml fix - correct checkout versions (#2029) 2024-04-13 19:01:17 +02:00
release.yaml feat(llama.cpp): add distributed llama.cpp inferencing (#2324) 2024-05-15 01:17:02 +02:00
secscan.yaml build(deps): bump github/codeql-action from 2 to 3 (#2041) 2024-04-15 22:02:44 +00:00
test-extra.yml feat: migrate python backends from conda to uv (#2215) 2024-05-10 15:08:08 +02:00
test.yml feat: auto select llama-cpp cuda runtime (#2306) 2024-05-14 19:40:18 +02:00
update_swagger.yaml ci: add swagger pipeline 2024-04-25 15:11:01 +02:00
yaml-check.yml Update yaml-check.yml 2024-04-25 16:05:02 +02:00