🤖 The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs ggml, gguf, GPTQ, onnx, TF compatible models: llama, llama2, rwkv, whisper, vicuna, koala, cerebras, falcon, dolly, starcoder, and many others
Go to file
Ettore Di Giacinto a8e91345e2 fixup: create piper libdir also when not built
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
2023-11-12 22:17:11 +01:00
.github Merge branch 'master' into enable_gpu 2023-11-11 19:20:36 +01:00
.vscode feat: Add more test-cases and remove dev container (#433) 2023-05-30 13:01:55 +02:00
api feat(llama.cpp): support lora with scale and yarn (#1277) 2023-11-11 18:40:48 +01:00
backend/cpp feat(llama.cpp): support lora with scale and yarn (#1277) 2023-11-11 18:40:48 +01:00
cmd/grpc cleanup: drop bloomz and ggllm as now supported by llama.cpp (#1217) 2023-10-26 07:43:31 +02:00
custom-ca-certs feat(certificates): add support for custom CA certificates (#880) 2023-11-01 20:10:14 +01:00
examples 🔥 add LaVA support and GPT vision API, Multiple requests for llama.cpp, return JSON types (#1254) 2023-11-11 13:14:59 +01:00
extra feat(llama.cpp): support lora with scale and yarn (#1277) 2023-11-11 18:40:48 +01:00
internal feat: cleanups, small enhancements 2023-07-04 18:58:19 +02:00
metrics feat(metrics): Adding initial support for prometheus metrics (#1176) 2023-10-17 18:22:53 +02:00
models Add docker-compose 2023-04-13 01:13:14 +02:00
pkg feat(llama.cpp): support lora with scale and yarn (#1277) 2023-11-11 18:40:48 +01:00
prompt-templates Requested Changes from GPT4ALL to Luna-AI-Llama2 (#1092) 2023-09-22 11:22:17 +02:00
tests ci: add GPU tests (#1095) 2023-10-19 13:50:40 +02:00
.dockerignore Remove .git from .dockerignore 2023-07-06 21:25:10 +02:00
.env 🔥 add LaVA support and GPT vision API, Multiple requests for llama.cpp, return JSON types (#1254) 2023-11-11 13:14:59 +01:00
.gitattributes Create .gitattributes to force git clone to keep the LF line endings on .sh files (#838) 2023-07-30 15:27:43 +02:00
.gitignore Update .gitignore for backend/llama.cpp (#1235) 2023-11-01 09:52:02 +01:00
CONTRIBUTING.md Add the CONTRIBUTING.md (#1098) 2023-09-24 14:54:55 +02:00
Dockerfile fixup: create piper libdir also when not built 2023-11-12 22:17:11 +01:00
Earthfile Rename project to LocalAI (#35) 2023-04-19 18:43:10 +02:00
LICENSE docs: update docs/license(clarification) and point to new website (#415) 2023-05-29 23:09:19 +02:00
Makefile deps(go-piper): update to 2023.11.6-3 (#1257) 2023-11-11 18:40:26 +01:00
README.md Requested Changes from GPT4ALL to Luna-AI-Llama2 (#1092) 2023-09-22 11:22:17 +02:00
assets.go feat: Update gpt4all, support multiple implementations in runtime (#472) 2023-06-01 23:38:52 +02:00
docker-compose.yaml fix: update docker-compose.yaml (#1131) 2023-10-05 22:13:18 +02:00
entrypoint.sh deps(go-piper): update to 2023.11.6-3 (#1257) 2023-11-11 18:40:26 +01:00
go.mod fix(deps): update module github.com/onsi/gomega to v1.28.1 (#1205) 2023-10-24 09:16:02 +02:00
go.sum fix(deps): update module github.com/onsi/gomega to v1.28.1 (#1205) 2023-10-24 09:16:02 +02:00
main.go feat(metrics): Adding initial support for prometheus metrics (#1176) 2023-10-17 18:22:53 +02:00
renovate.json ci: manually update deps 2023-05-04 15:01:29 +02:00

README.md



LocalAI

LocalAI forks LocalAI stars LocalAI pull-requests

💡 Get help - FAQ 💭Discussions 💬 Discord 📖 Documentation website

💻 Quickstart 📣 News 🛫 Examples 🖼️ Models

testsBuild and Releasebuild container imagesBump dependenciesArtifact Hub

LocalAI is a drop-in replacement REST API that's compatible with OpenAI API specifications for local inferencing. It allows you to run LLMs (and not only) locally or on-prem with consumer grade hardware, supporting multiple model families that are compatible with the ggml format, pytorch and more. Does not require GPU.

Follow LocalAI

Follow LocalAI_API Join LocalAI Discord Community

Connect with the Creator

Follow mudler_it Follow on Github

Share LocalAI Repository

Follow _LocalAI Share on Telegram Share on Reddit Buy Me A Coffee


In a nutshell:

  • Local, OpenAI drop-in alternative REST API. You own your data.
  • NO GPU required. NO Internet access is required either
    • Optional, GPU Acceleration is available in llama.cpp-compatible LLMs. See also the build section.
  • Supports multiple models
  • 🏃 Once loaded the first time, it keep models loaded in memory for faster inference
  • Doesn't shell-out, but uses C++ bindings for a faster inference and better performance.

LocalAI was created by Ettore Di Giacinto and is a community-driven project, focused on making the AI accessible to anyone. Any contribution, feedback and PR is welcome!

Note that this started just as a fun weekend project in order to try to create the necessary pieces for a full AI assistant like ChatGPT: the community is growing fast and we are working hard to make it better and more stable. If you want to help, please consider contributing (see below)!

🔥🔥 Hot topics / Roadmap

🚀 Features

📖 🎥 Media, Blogs, Social

💻 Usage

Check out the Getting started section in our documentation.

💡 Example: Use Luna-AI Llama model

See the documentation

🔗 Resources

Citation

If you utilize this repository, data in a downstream project, please consider citing it with:

@misc{localai,
  author = {Ettore Di Giacinto},
  title = {LocalAI: The free, Open source OpenAI alternative},
  year = {2023},
  publisher = {GitHub},
  journal = {GitHub repository},
  howpublished = {\url{https://github.com/go-skynet/LocalAI}},

❤️ Sponsors

Do you find LocalAI useful?

Support the project by becoming a backer or sponsor. Your logo will show up here with a link to your website.

A huge thank you to our generous sponsors who support this project:

Spectro Cloud logo_600x600px_transparent bg
Spectro Cloud
Spectro Cloud kindly supports LocalAI by providing GPU and computing resources to run tests on lamdalabs!

And a huge shout-out to individuals sponsoring the project by donating hardware or backing the project.

🌟 Star history

LocalAI Star history Chart

📖 License

LocalAI is a community-driven project created by Ettore Di Giacinto.

MIT - Author Ettore Di Giacinto

🙇 Acknowledgements

LocalAI couldn't have been built without the help of great software already available from the community. Thank you!

🤗 Contributors

This is a community project, a special thanks to our contributors! 🤗