mirror of
https://github.com/mudler/LocalAI.git
synced 2024-12-20 05:07:54 +00:00
Update README.md
Add short-term roadmap and mention webui
This commit is contained in:
parent
bf85a31f9e
commit
be682e6c2f
10
README.md
10
README.md
@ -90,7 +90,7 @@ The API takes takes the following:
|
|||||||
| gpt4all | GPT4ALL | false | Set to true for gpt4all models. |
|
| gpt4all | GPT4ALL | false | Set to true for gpt4all models. |
|
||||||
|
|
||||||
|
|
||||||
Once the server is running, you can make requests to it using HTTP. For example, to generate text based on an instruction, you can send a POST request to the `/predict` endpoint with the instruction as the request body:
|
Once the server is running, you can start making requests to it using HTTP. For example, to generate text based on an instruction, you can send a POST request to the `/predict` endpoint with the instruction as the request body:
|
||||||
|
|
||||||
```
|
```
|
||||||
curl --location --request POST 'http://localhost:8080/predict' --header 'Content-Type: application/json' --data-raw '{
|
curl --location --request POST 'http://localhost:8080/predict' --header 'Content-Type: application/json' --data-raw '{
|
||||||
@ -102,6 +102,8 @@ curl --location --request POST 'http://localhost:8080/predict' --header 'Content
|
|||||||
}'
|
}'
|
||||||
```
|
```
|
||||||
|
|
||||||
|
There is also available a simple web interface (for instance, http://localhost:8080/) which can be used as a playground.
|
||||||
|
|
||||||
Note: The API doesn't inject a template for talking to the instance, while the CLI does. You have to use a prompt similar to what's described in the standford-alpaca docs: https://github.com/tatsu-lab/stanford_alpaca#data-release, for instance:
|
Note: The API doesn't inject a template for talking to the instance, while the CLI does. You have to use a prompt similar to what's described in the standford-alpaca docs: https://github.com/tatsu-lab/stanford_alpaca#data-release, for instance:
|
||||||
|
|
||||||
```
|
```
|
||||||
@ -197,6 +199,12 @@ docker run --privileged -v /var/run/docker.sock:/var/run/docker.sock --rm -t -v
|
|||||||
./llama-cli --instruction "What's an alpaca?"
|
./llama-cli --instruction "What's an alpaca?"
|
||||||
```
|
```
|
||||||
|
|
||||||
|
## Short-term roadmap
|
||||||
|
|
||||||
|
- Mimic OpenAI API (https://github.com/go-skynet/llama-cli/issues/10)
|
||||||
|
- Binary releases (https://github.com/go-skynet/llama-cli/issues/6)
|
||||||
|
- Upstream our golang bindings to llama.cpp (https://github.com/ggerganov/llama.cpp/issues/351)
|
||||||
|
|
||||||
## License
|
## License
|
||||||
|
|
||||||
MIT
|
MIT
|
||||||
|
Loading…
Reference in New Issue
Block a user