28a36e20aa
[![Mend Renovate](https://app.renovatebot.com/images/banner.svg)](https://renovatebot.com) This PR contains the following updates: | Package | Type | Update | Change | |---|---|---|---| | [google.golang.org/grpc](https://togithub.com/grpc/grpc-go) | require | minor | `v1.57.0` -> `v1.58.1` | --- ### Release Notes <details> <summary>grpc/grpc-go (google.golang.org/grpc)</summary> ### [`v1.58.1`](https://togithub.com/grpc/grpc-go/releases/tag/v1.58.1): Release 1.58.1 [Compare Source](https://togithub.com/grpc/grpc-go/compare/v1.58.0...v1.58.1) ### Bug Fixes - grpc: fix a bug that was decrementing active RPC count too early for streaming RPCs; leading to channel moving to IDLE even though it had open streams - grpc: fix a bug where transports were not being closed upon channel entering IDLE ### [`v1.58.0`](https://togithub.com/grpc/grpc-go/releases/tag/v1.58.0): Release 1.58.0 [Compare Source](https://togithub.com/grpc/grpc-go/compare/v1.57.0...v1.58.0) ### API Changes See [#​6472](https://togithub.com/grpc/grpc-go/issues/6472) for details about these changes. - balancer: add `StateListener` to `NewSubConnOptions` for `SubConn` state updates and deprecate `Balancer.UpdateSubConnState` ([#​6481](https://togithub.com/grpc/grpc-go/issues/6481)) - `UpdateSubConnState` will be deleted in the future. - balancer: add `SubConn.Shutdown` and deprecate `Balancer.RemoveSubConn` ([#​6493](https://togithub.com/grpc/grpc-go/issues/6493)) - `RemoveSubConn` will be deleted in the future. - resolver: remove deprecated `AddressType` ([#​6451](https://togithub.com/grpc/grpc-go/issues/6451)) - This was previously used as a signal to enable the "grpclb" load balancing policy, and to pass LB addresses to the policy. Instead, `balancer/grpclb/state.Set()` should be used to add these addresses to the name resolver's output. The built-in "dns" name resolver already does this. - resolver: add new field `Endpoints` to `State` and deprecate `Addresses` ([#​6471](https://togithub.com/grpc/grpc-go/issues/6471)) - `Addresses` will be deleted in the future. ### New Features - balancer/leastrequest: Add experimental support for least request LB policy and least request configured as a custom xDS policy ([#​6510](https://togithub.com/grpc/grpc-go/issues/6510), [#​6517](https://togithub.com/grpc/grpc-go/issues/6517)) - Set `GRPC_EXPERIMENTAL_ENABLE_LEAST_REQUEST=true` to enable - stats: Add an RPC event for blocking caused by the LB policy's picker ([#​6422](https://togithub.com/grpc/grpc-go/issues/6422)) ### Bug Fixes - clusterresolver: fix deadlock when dns resolver responds inline with update or error at build time ([#​6563](https://togithub.com/grpc/grpc-go/issues/6563)) - grpc: fix a bug where the channel could erroneously report `TRANSIENT_FAILURE` when actually moving to `IDLE` ([#​6497](https://togithub.com/grpc/grpc-go/issues/6497)) - balancergroup: do not cache closed sub-balancers by default; affects `rls`, `weightedtarget` and `clustermanager` LB policies ([#​6523](https://togithub.com/grpc/grpc-go/issues/6523)) - client: fix a bug that prevented detection of RPC status in trailers-only RPC responses when using `ClientStream.Header()`, and prevented retry of the RPC ([#​6557](https://togithub.com/grpc/grpc-go/issues/6557)) ### Performance Improvements - client & server: Add experimental `[With]SharedWriteBuffer` to improve performance by reducing allocations when sending RPC messages. (Disabled by default.) ([#​6309](https://togithub.com/grpc/grpc-go/issues/6309)) - Special Thanks: [@​s-matyukevich](https://togithub.com/s-matyukevich) </details> --- ### Configuration 📅 **Schedule**: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 **Automerge**: Disabled by config. Please merge this manually once you are satisfied. ♻ **Rebasing**: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 🔕 **Ignore**: Close this PR and you won't be reminded about this update again. --- - [ ] <!-- rebase-check -->If you want to rebase/retry this PR, check this box --- This PR has been generated by [Mend Renovate](https://www.mend.io/free-developer-tools/renovate/). View repository job log [here](https://developer.mend.io/github/go-skynet/LocalAI). <!--renovate-debug:eyJjcmVhdGVkSW5WZXIiOiIzNi44My4wIiwidXBkYXRlZEluVmVyIjoiMzYuODMuMCIsInRhcmdldEJyYW5jaCI6Im1hc3RlciJ9--> Co-authored-by: renovate[bot] <29139614+renovate[bot]@users.noreply.github.com> |
||
---|---|---|
.github | ||
.vscode | ||
api | ||
cmd/grpc | ||
examples | ||
extra | ||
internal | ||
models | ||
pkg | ||
prompt-templates | ||
tests | ||
.dockerignore | ||
.env | ||
.gitattributes | ||
.gitignore | ||
assets.go | ||
docker-compose.yaml | ||
Dockerfile | ||
Earthfile | ||
entrypoint.sh | ||
go.mod | ||
go.sum | ||
LICENSE | ||
main.go | ||
Makefile | ||
README.md | ||
renovate.json |
LocalAI
💡 Get help - ❓FAQ 💭Discussions 💬 Discord 📖 Documentation website
LocalAI is a drop-in replacement REST API that's compatible with OpenAI API specifications for local inferencing. It allows you to run LLMs (and not only) locally or on-prem with consumer grade hardware, supporting multiple model families that are compatible with the ggml format, pytorch and more. Does not require GPU.
Follow LocalAI
Connect with the Creator
Share LocalAI Repository
In a nutshell:
- Local, OpenAI drop-in alternative REST API. You own your data.
- NO GPU required. NO Internet access is required either
- Optional, GPU Acceleration is available in
llama.cpp
-compatible LLMs. See also the build section.
- Optional, GPU Acceleration is available in
- Supports multiple models
- 🏃 Once loaded the first time, it keep models loaded in memory for faster inference
- ⚡ Doesn't shell-out, but uses C++ bindings for a faster inference and better performance.
LocalAI was created by Ettore Di Giacinto and is a community-driven project, focused on making the AI accessible to anyone. Any contribution, feedback and PR is welcome!
Note that this started just as a fun weekend project in order to try to create the necessary pieces for a full AI assistant like ChatGPT
: the community is growing fast and we are working hard to make it better and more stable. If you want to help, please consider contributing (see below)!
🔥🔥 Hot topics / Roadmap
🚀 Features
- 📖 Text generation with GPTs (
llama.cpp
,gpt4all.cpp
, ... 📖 and more) - 🗣 Text to Audio
- 🔈 Audio to Text (Audio transcription with
whisper.cpp
) - 🎨 Image generation with stable diffusion
- 🔥 OpenAI functions 🆕
- 🧠 Embeddings generation for vector databases
- ✍️ Constrained grammars
- 🖼️ Download Models directly from Huggingface
📖 🎥 Media, Blogs, Social
- Create a slackbot for teams and OSS projects that answer to documentation
- LocalAI meets k8sgpt
- Question Answering on Documents locally with LangChain, LocalAI, Chroma, and GPT4All
- Tutorial to use k8sgpt with LocalAI
💻 Usage
Check out the Getting started section in our documentation.
💡 Example: Use GPT4ALL-J model
See the documentation
🔗 Resources
Citation
If you utilize this repository, data in a downstream project, please consider citing it with:
@misc{localai,
author = {Ettore Di Giacinto},
title = {LocalAI: The free, Open source OpenAI alternative},
year = {2023},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/go-skynet/LocalAI}},
❤️ Sponsors
Do you find LocalAI useful?
Support the project by becoming a backer or sponsor. Your logo will show up here with a link to your website.
A huge thank you to our generous sponsors who support this project:
Spectro Cloud |
Spectro Cloud kindly supports LocalAI by providing GPU and computing resources to run tests on lamdalabs! |
And a huge shout-out to individuals sponsoring the project by donating hardware or backing the project.
- Sponsor list
- JDAM00 (donating HW for the CI)
🌟 Star history
📖 License
LocalAI is a community-driven project created by Ettore Di Giacinto.
MIT - Author Ettore Di Giacinto
🙇 Acknowledgements
LocalAI couldn't have been built without the help of great software already available from the community. Thank you!
- llama.cpp
- https://github.com/tatsu-lab/stanford_alpaca
- https://github.com/cornelk/llama-go for the initial ideas
- https://github.com/antimatter15/alpaca.cpp
- https://github.com/EdVince/Stable-Diffusion-NCNN
- https://github.com/ggerganov/whisper.cpp
- https://github.com/saharNooby/rwkv.cpp
- https://github.com/rhasspy/piper
- https://github.com/cmp-nct/ggllm.cpp
🤗 Contributors
This is a community project, a special thanks to our contributors! 🤗