mirror of
https://github.com/mudler/LocalAI.git
synced 2024-12-20 13:13:11 +00:00
Update README.md
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
This commit is contained in:
parent
1a9299a7c0
commit
cdcfb2617c
13
README.md
13
README.md
@ -66,6 +66,19 @@ docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-aio-cpu
|
|||||||
# docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-gpu-nvidia-cuda-12
|
# docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-gpu-nvidia-cuda-12
|
||||||
```
|
```
|
||||||
|
|
||||||
|
To load models:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Start LocalAI with the phi-2 model
|
||||||
|
local-ai run huggingface://TheBloke/phi-2-GGUF/phi-2.Q8_0.gguf
|
||||||
|
# Install and run a model from the Ollama OCI registry
|
||||||
|
local-ai run ollama://gemma:2b
|
||||||
|
# Run a model from a configuration file
|
||||||
|
local-ai run https://gist.githubusercontent.com/.../phi-2.yaml
|
||||||
|
# Install and run a model from a standard OCI registry (e.g., Docker Hub)
|
||||||
|
local-ai run oci://localai/phi-2:latest
|
||||||
|
```
|
||||||
|
|
||||||
[💻 Getting started](https://localai.io/basics/getting_started/index.html)
|
[💻 Getting started](https://localai.io/basics/getting_started/index.html)
|
||||||
|
|
||||||
## 📰 Latest project news
|
## 📰 Latest project news
|
||||||
|
Loading…
Reference in New Issue
Block a user