mirror of
https://github.com/mudler/LocalAI.git
synced 2024-12-19 04:37:53 +00:00
e34b5f0119
Closes https://github.com/go-skynet/LocalAI/issues/1066 and https://github.com/go-skynet/LocalAI/issues/1065 Standardizes all `examples/`: - Models in one place (other than `rwkv`, which was one-offy) - Env files as `.env.example` with `cp` - Also standardizes comments and links docs
36 lines
1.1 KiB
Markdown
36 lines
1.1 KiB
Markdown
# AutoGPT
|
|
|
|
Example of integration with [AutoGPT](https://github.com/Significant-Gravitas/Auto-GPT).
|
|
|
|
## Run
|
|
|
|
```bash
|
|
# Clone LocalAI
|
|
git clone https://github.com/go-skynet/LocalAI
|
|
|
|
cd LocalAI/examples/autoGPT
|
|
|
|
cp -rfv .env.example .env
|
|
|
|
# Edit the .env file to set a different model by editing `PRELOAD_MODELS`.
|
|
vim .env
|
|
|
|
docker-compose run --rm auto-gpt
|
|
```
|
|
|
|
Note: The example automatically downloads the `gpt4all` model as it is under a permissive license. The GPT4All model does not seem to be enough to run AutoGPT. WizardLM-7b-uncensored seems to perform better (with `f16: true`).
|
|
|
|
|
|
## Without docker
|
|
|
|
Run AutoGPT with `OPENAI_API_BASE` pointing to the LocalAI endpoint. If you run it locally for instance:
|
|
|
|
```
|
|
OPENAI_API_BASE=http://localhost:8080 python ...
|
|
```
|
|
|
|
Note: you need a model named `gpt-3.5-turbo` and `text-embedding-ada-002`. You can preload those in LocalAI at start by setting in the env:
|
|
|
|
```
|
|
PRELOAD_MODELS=[{"url": "github:go-skynet/model-gallery/gpt4all-j.yaml", "name": "gpt-3.5-turbo"}, { "url": "github:go-skynet/model-gallery/bert-embeddings.yaml", "name": "text-embedding-ada-002"}]
|
|
``` |