LocalAI/examples/autoGPT
James Braza e34b5f0119
Cleaning up `examples/` models and starter `.env` files (#1124)
Closes https://github.com/go-skynet/LocalAI/issues/1066 and
https://github.com/go-skynet/LocalAI/issues/1065

Standardizes all `examples/`:
- Models in one place (other than `rwkv`, which was one-offy)
- Env files as `.env.example` with `cp`
    - Also standardizes comments and links docs
2023-10-02 18:14:10 +02:00
..
.env.example Cleaning up `examples/` models and starter `.env` files (#1124) 2023-10-02 18:14:10 +02:00
README.md Cleaning up `examples/` models and starter `.env` files (#1124) 2023-10-02 18:14:10 +02:00
docker-compose.yaml docs(examples): add AutoGPT (#397) 2023-05-27 19:42:24 +02:00

README.md

AutoGPT

Example of integration with AutoGPT.

Run

# Clone LocalAI
git clone https://github.com/go-skynet/LocalAI

cd LocalAI/examples/autoGPT

cp -rfv .env.example .env

# Edit the .env file to set a different model by editing `PRELOAD_MODELS`.
vim .env

docker-compose run --rm auto-gpt

Note: The example automatically downloads the gpt4all model as it is under a permissive license. The GPT4All model does not seem to be enough to run AutoGPT. WizardLM-7b-uncensored seems to perform better (with f16: true).

Without docker

Run AutoGPT with OPENAI_API_BASE pointing to the LocalAI endpoint. If you run it locally for instance:

OPENAI_API_BASE=http://localhost:8080 python ...

Note: you need a model named gpt-3.5-turbo and text-embedding-ada-002. You can preload those in LocalAI at start by setting in the env:

PRELOAD_MODELS=[{"url": "github:go-skynet/model-gallery/gpt4all-j.yaml", "name": "gpt-3.5-turbo"}, { "url": "github:go-skynet/model-gallery/bert-embeddings.yaml", "name": "text-embedding-ada-002"}]