Latest fixups
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
10
README.md
@ -74,19 +74,19 @@
|
||||
|
||||
| Talk Interface | Generate Audio |
|
||||
| --- | --- |
|
||||
|  |  |
|
||||
|  |  |
|
||||
|
||||
| Models Overview | Generate Images |
|
||||
| --- | --- |
|
||||
|  |  |
|
||||
|  |  |
|
||||
|
||||
| Chat Interface | API Overview |
|
||||
| Chat Interface | Home |
|
||||
| --- | --- |
|
||||
|  |  |
|
||||
|  |  |
|
||||
|
||||
| Login | Swarm |
|
||||
| --- | --- |
|
||||
| |  |
|
||||
| |  |
|
||||
|
||||
## 💻 Quickstart
|
||||
|
||||
|
BIN
docs/assets/images/imagen.png
Normal file
After Width: | Height: | Size: 506 KiB |
171
docs/assets/images/logos/logo.svg
Normal file
After Width: | Height: | Size: 108 KiB |
BIN
docs/assets/images/screenshots/screenshot_chat.png
Normal file
After Width: | Height: | Size: 132 KiB |
BIN
docs/assets/images/screenshots/screenshot_gallery.png
Normal file
After Width: | Height: | Size: 284 KiB |
BIN
docs/assets/images/screenshots/screenshot_home.png
Normal file
After Width: | Height: | Size: 287 KiB |
BIN
docs/assets/images/screenshots/screenshot_image.png
Normal file
After Width: | Height: | Size: 506 KiB |
BIN
docs/assets/images/screenshots/screenshot_login.png
Normal file
After Width: | Height: | Size: 225 KiB |
BIN
docs/assets/images/screenshots/screenshot_p2p.png
Normal file
After Width: | Height: | Size: 418 KiB |
BIN
docs/assets/images/screenshots/screenshot_talk.png
Normal file
After Width: | Height: | Size: 246 KiB |
BIN
docs/assets/images/screenshots/screenshot_tts.png
Normal file
After Width: | Height: | Size: 213 KiB |
@ -50,7 +50,7 @@ defaultContentLanguage = 'en'
|
||||
|
||||
logo = "https://raw.githubusercontent.com/mudler/LocalAI/53652d5b8f8049e95ef71f0fed39671eb39c211c/core/http/static/logo.png"
|
||||
logo_text = ""
|
||||
title = "LocalAI documentation" # default html title for documentation pages/sections
|
||||
title = "LocalAI" # default html title for documentation pages/sections
|
||||
|
||||
pathName = "docs" # path name for documentation site | default "docs"
|
||||
|
||||
|
@ -13,6 +13,8 @@ LocalAI supports two modes of distributed inferencing via p2p:
|
||||
- **Federated Mode**: Requests are shared between the cluster and routed to a single worker node in the network based on the load balancer's decision.
|
||||
- **Worker Mode** (aka "model sharding" or "splitting weights"): Requests are processed by all the workers which contributes to the final inference result (by sharing the model weights).
|
||||
|
||||
A list of global instances shared by the community is available at [explorer.localai.io](https://explorer.localai.io).
|
||||
|
||||
## Usage
|
||||
|
||||
Starting LocalAI with `--p2p` generates a shared token for connecting multiple instances: and that's all you need to create AI clusters, eliminating the need for intricate network setups.
|
||||
|
@ -18,14 +18,45 @@ To access the WebUI with an API_KEY, browser extensions such as [Requestly](http
|
||||
|
||||
{{% /alert %}}
|
||||
|
||||
## Using the Bash Installer
|
||||
## Quickstart
|
||||
|
||||
Install LocalAI easily using the bash installer with the following command:
|
||||
|
||||
```sh
|
||||
### Using the Bash Installer
|
||||
```bash
|
||||
curl https://localai.io/install.sh | sh
|
||||
```
|
||||
|
||||
### Run with docker:
|
||||
```bash
|
||||
# CPU only image:
|
||||
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-cpu
|
||||
|
||||
# Nvidia GPU:
|
||||
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-gpu-nvidia-cuda-12
|
||||
|
||||
# CPU and GPU image (bigger size):
|
||||
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest
|
||||
|
||||
# AIO images (it will pre-download a set of models ready for use, see https://localai.io/basics/container/)
|
||||
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-aio-cpu
|
||||
```
|
||||
|
||||
### Load models:
|
||||
|
||||
```bash
|
||||
# From the model gallery (see available models with `local-ai models list`, in the WebUI from the model tab, or visiting https://models.localai.io)
|
||||
local-ai run llama-3.2-1b-instruct:q4_k_m
|
||||
# Start LocalAI with the phi-2 model directly from huggingface
|
||||
local-ai run huggingface://TheBloke/phi-2-GGUF/phi-2.Q8_0.gguf
|
||||
# Install and run a model from the Ollama OCI registry
|
||||
local-ai run ollama://gemma:2b
|
||||
# Run a model from a configuration file
|
||||
local-ai run https://gist.githubusercontent.com/.../phi-2.yaml
|
||||
# Install and run a model from a standard OCI registry (e.g., Docker Hub)
|
||||
local-ai run oci://localai/phi-2:latest
|
||||
```
|
||||
|
||||
|
||||
For a full list of options, refer to the [Installer Options]({{% relref "docs/advanced/installer" %}}) documentation.
|
||||
|
||||
Binaries can also be [manually downloaded]({{% relref "docs/reference/binaries" %}}).
|
||||
|
@ -1,4 +1,3 @@
|
||||
|
||||
+++
|
||||
title = "Overview"
|
||||
weight = 1
|
||||
@ -10,157 +9,93 @@ author = "Ettore Di Giacinto"
|
||||
icon = "info"
|
||||
+++
|
||||
|
||||
<p align="center">
|
||||
<a href="https://localai.io"><img width=512 src="https://github.com/go-skynet/LocalAI/assets/2420543/0966aa2a-166e-4f99-a3e5-6c915fc997dd"></a>
|
||||
</p >
|
||||
# Welcome to LocalAI
|
||||
|
||||
<p align="center">
|
||||
<a href="https://github.com/go-skynet/LocalAI/fork" target="blank">
|
||||
<img src="https://img.shields.io/github/forks/go-skynet/LocalAI?style=for-the-badge" alt="LocalAI forks"/>
|
||||
</a>
|
||||
<a href="https://github.com/go-skynet/LocalAI/stargazers" target="blank">
|
||||
<img src="https://img.shields.io/github/stars/go-skynet/LocalAI?style=for-the-badge" alt="LocalAI stars"/>
|
||||
</a>
|
||||
<a href="https://github.com/go-skynet/LocalAI/pulls" target="blank">
|
||||
<img src="https://img.shields.io/github/issues-pr/go-skynet/LocalAI?style=for-the-badge" alt="LocalAI pull-requests"/>
|
||||
</a>
|
||||
<a href='https://github.com/go-skynet/LocalAI/releases'>
|
||||
<img src='https://img.shields.io/github/release/go-skynet/LocalAI?&label=Latest&style=for-the-badge'>
|
||||
</a>
|
||||
</p>
|
||||
LocalAI is your complete AI stack for running AI models locally. It's designed to be simple, efficient, and accessible, providing a drop-in replacement for OpenAI's API while keeping your data private and secure.
|
||||
|
||||
<p align="center">
|
||||
<a href="https://hub.docker.com/r/localai/localai" target="blank">
|
||||
<img src="https://img.shields.io/badge/dockerhub-images-important.svg?logo=Docker" alt="LocalAI Docker hub"/>
|
||||
</a>
|
||||
<a href="https://quay.io/repository/go-skynet/local-ai?tab=tags&tag=latest" target="blank">
|
||||
<img src="https://img.shields.io/badge/quay.io-images-important.svg?" alt="LocalAI Quay.io"/>
|
||||
</a>
|
||||
</p>
|
||||
## Why LocalAI?
|
||||
|
||||
<p align="center">
|
||||
<a href="https://trendshift.io/repositories/5539" target="_blank"><img src="https://trendshift.io/api/badge/repositories/5539" alt="mudler%2FLocalAI | Trendshift" style="width: 250px; height: 55px;" width="250" height="55"/></a>
|
||||
</p>
|
||||
In today's AI landscape, privacy, control, and flexibility are paramount. LocalAI addresses these needs by:
|
||||
|
||||
<p align="center">
|
||||
<a href="https://twitter.com/LocalAI_API" target="blank">
|
||||
<img src="https://img.shields.io/twitter/follow/LocalAI_API?label=Follow: LocalAI_API&style=social" alt="Follow LocalAI_API"/>
|
||||
</a>
|
||||
<a href="https://discord.gg/uJAeKSAGDy" target="blank">
|
||||
<img src="https://dcbadge.vercel.app/api/server/uJAeKSAGDy?style=flat-square&theme=default-inverted" alt="Join LocalAI Discord Community"/>
|
||||
</a>
|
||||
</p>
|
||||
- **Privacy First**: Your data never leaves your machine
|
||||
- **Complete Control**: Run models on your terms, with your hardware
|
||||
- **Open Source**: MIT licensed and community-driven
|
||||
- **Flexible Deployment**: From laptops to servers, with or without GPUs
|
||||
- **Extensible**: Add new models and features as needed
|
||||
|
||||
> 💡 Get help - [❓FAQ](https://localai.io/faq/) [💭Discussions](https://github.com/go-skynet/LocalAI/discussions) [💭Discord](https://discord.gg/uJAeKSAGDy)
|
||||
>
|
||||
> [💻 Quickstart](https://localai.io/basics/getting_started/) [🖼️ Models](https://models.localai.io/) [🚀 Roadmap](https://github.com/mudler/LocalAI/issues?q=is%3Aissue+is%3Aopen+label%3Aroadmap) [🥽 Demo](https://demo.localai.io) [🌍 Explorer](https://explorer.localai.io) [🛫 Examples](https://github.com/go-skynet/LocalAI/tree/master/examples/)
|
||||
## Core Components
|
||||
|
||||
LocalAI is more than just a single tool - it's a complete ecosystem:
|
||||
|
||||
**LocalAI** is the free, Open Source OpenAI alternative. LocalAI act as a drop-in replacement REST API that's compatible with OpenAI API specifications for local inferencing. It allows you to run LLMs, generate images, audio (and not only) locally or on-prem with consumer grade hardware, supporting multiple model families and architectures. Does not require GPU. It is created and maintained by [Ettore Di Giacinto](https://github.com/mudler).
|
||||
1. **[LocalAI Core](https://github.com/mudler/LocalAI)**
|
||||
- OpenAI-compatible API
|
||||
- Multiple model support (LLMs, image, audio)
|
||||
- No GPU required
|
||||
- Fast inference with native bindings
|
||||
- [Github repository](https://github.com/mudler/LocalAI)
|
||||
|
||||
2. **[LocalAGI](https://github.com/mudler/LocalAGI)**
|
||||
- Autonomous AI agents
|
||||
- No coding required
|
||||
- WebUI and REST API support
|
||||
- Extensible agent framework
|
||||
- [Github repository](https://github.com/mudler/LocalAGI)
|
||||
|
||||
## Start LocalAI
|
||||
3. **[LocalRecall](https://github.com/mudler/LocalRecall)**
|
||||
- Semantic search
|
||||
- Memory management
|
||||
- Vector database
|
||||
- Perfect for AI applications
|
||||
- [Github repository](https://github.com/mudler/LocalRecall)
|
||||
|
||||
Start the image with Docker to have a functional clone of OpenAI! 🚀:
|
||||
## Getting Started
|
||||
|
||||
```bash
|
||||
docker run -p 8080:8080 --name local-ai -ti localai/localai:latest-aio-cpu
|
||||
# Do you have a Nvidia GPUs? Use this instead
|
||||
# CUDA 11
|
||||
# docker run -p 8080:8080 --gpus all --name local-ai -ti localai/localai:latest-aio-gpu-nvidia-cuda-11
|
||||
# CUDA 12
|
||||
# docker run -p 8080:8080 --gpus all --name local-ai -ti localai/localai:latest-aio-gpu-nvidia-cuda-12
|
||||
```
|
||||
|
||||
Or just use the bash installer:
|
||||
The fastest way to get started is with our one-line installer:
|
||||
|
||||
```bash
|
||||
curl https://localai.io/install.sh | sh
|
||||
```
|
||||
|
||||
See the [💻 Quickstart](https://localai.io/basics/getting_started/) for all the options and way you can run LocalAI!
|
||||
Or use Docker for a quick start:
|
||||
|
||||
## What is LocalAI?
|
||||
```bash
|
||||
docker run -p 8080:8080 --name local-ai -ti localai/localai:latest-aio-cpu
|
||||
```
|
||||
|
||||
In a nutshell:
|
||||
For more detailed installation options and configurations, see our [Getting Started guide](/basics/getting_started/).
|
||||
|
||||
- Local, OpenAI drop-in alternative REST API. You own your data.
|
||||
- NO GPU required. NO Internet access is required either
|
||||
- Optional, GPU Acceleration is available. See also the [build section](https://localai.io/basics/build/index.html).
|
||||
- Supports multiple models
|
||||
- 🏃 Once loaded the first time, it keep models loaded in memory for faster inference
|
||||
- ⚡ Doesn't shell-out, but uses bindings for a faster inference and better performance.
|
||||
## Key Features
|
||||
|
||||
LocalAI is focused on making the AI accessible to anyone. Any contribution, feedback and PR is welcome!
|
||||
- **Text Generation**: Run various LLMs locally
|
||||
- **Image Generation**: Create images with stable diffusion
|
||||
- **Audio Processing**: Text-to-speech and speech-to-text
|
||||
- **Vision API**: Image understanding and analysis
|
||||
- **Embeddings**: Vector database support
|
||||
- **Functions**: OpenAI-compatible function calling
|
||||
- **P2P**: Distributed inference capabilities
|
||||
|
||||
Note that this started just as a fun weekend project by [mudler](https://github.com/mudler) in order to try to create the necessary pieces for a full AI assistant like `ChatGPT`: the community is growing fast and we are working hard to make it better and more stable. If you want to help, please consider contributing (see below)!
|
||||
## Community and Support
|
||||
|
||||
### 🚀 Features
|
||||
LocalAI is a community-driven project. You can:
|
||||
|
||||
- 📖 [Text generation with GPTs](https://localai.io/features/text-generation/) (`llama.cpp`, `gpt4all.cpp`, ... [:book: and more](https://localai.io/model-compatibility/index.html#model-compatibility-table))
|
||||
- 🗣 [Text to Audio](https://localai.io/features/text-to-audio/)
|
||||
- 🔈 [Audio to Text](https://localai.io/features/audio-to-text/) (Audio transcription with `whisper.cpp`)
|
||||
- 🎨 [Image generation with stable diffusion](https://localai.io/features/image-generation)
|
||||
- 🔥 [OpenAI functions](https://localai.io/features/openai-functions/) 🆕
|
||||
- 🧠 [Embeddings generation for vector databases](https://localai.io/features/embeddings/)
|
||||
- ✍️ [Constrained grammars](https://localai.io/features/constrained_grammars/)
|
||||
- 🖼️ [Download Models directly from Huggingface ](https://localai.io/models/)
|
||||
- 🥽 [Vision API](https://localai.io/features/gpt-vision/)
|
||||
- 💾 [Stores](https://localai.io/stores)
|
||||
- 📈 [Reranker](https://localai.io/features/reranker/)
|
||||
- 🆕🖧 [P2P Inferencing](https://localai.io/features/distribute/)
|
||||
- Join our [Discord community](https://discord.gg/uJAeKSAGDy)
|
||||
- Check out our [GitHub repository](https://github.com/mudler/LocalAI)
|
||||
- Contribute to the project
|
||||
- Share your use cases and examples
|
||||
|
||||
## Contribute and help
|
||||
## Next Steps
|
||||
|
||||
To help the project you can:
|
||||
Ready to dive in? Here are some recommended next steps:
|
||||
|
||||
- If you have technological skills and want to contribute to development, have a look at the open issues. If you are new you can have a look at the [good-first-issue](https://github.com/go-skynet/LocalAI/issues?q=is%3Aissue+is%3Aopen+label%3A%22good+first+issue%22) and [help-wanted](https://github.com/go-skynet/LocalAI/issues?q=is%3Aissue+is%3Aopen+label%3A%22help+wanted%22) labels.
|
||||
1. [Install LocalAI](/basics/getting_started/)
|
||||
2. [Explore available models](https://models.localai.io)
|
||||
3. [Model compatibility](/model-compatibility/)
|
||||
4. [Try out examples](https://github.com/mudler/LocalAI-examples)
|
||||
5. [Join the community](https://discord.gg/uJAeKSAGDy)
|
||||
6. [Check the LocalAI Github repository](https://github.com/mudler/LocalAI)
|
||||
7. [Check the LocalAGI Github repository](https://github.com/mudler/LocalAGI)
|
||||
|
||||
- If you don't have technological skills you can still help improving documentation or [add examples](https://github.com/go-skynet/LocalAI/tree/master/examples) or share your user-stories with our community, any help and contribution is welcome!
|
||||
|
||||
## 🌟 Star history
|
||||
## License
|
||||
|
||||
[](https://star-history.com/#mudler/LocalAI&Date)
|
||||
|
||||
## ❤️ Sponsors
|
||||
|
||||
> Do you find LocalAI useful?
|
||||
|
||||
Support the project by becoming [a backer or sponsor](https://github.com/sponsors/mudler). Your logo will show up here with a link to your website.
|
||||
|
||||
A huge thank you to our generous sponsors who support this project covering CI expenses, and our [Sponsor list](https://github.com/sponsors/mudler):
|
||||
|
||||
<p align="center">
|
||||
<a href="https://www.spectrocloud.com/" target="blank">
|
||||
<img width=200 src="https://github.com/user-attachments/assets/72eab1dd-8b93-4fc0-9ade-84db49f24962">
|
||||
</a>
|
||||
<a href="https://www.premai.io/" target="blank">
|
||||
<img width=200 src="https://github.com/mudler/LocalAI/assets/2420543/42e4ca83-661e-4f79-8e46-ae43689683d6"> <br>
|
||||
</a>
|
||||
</p>
|
||||
|
||||
## 📖 License
|
||||
|
||||
LocalAI is a community-driven project created by [Ettore Di Giacinto](https://github.com/mudler/).
|
||||
|
||||
MIT - Author Ettore Di Giacinto
|
||||
|
||||
## 🙇 Acknowledgements
|
||||
|
||||
LocalAI couldn't have been built without the help of great software already available from the community. Thank you!
|
||||
|
||||
- [llama.cpp](https://github.com/ggerganov/llama.cpp)
|
||||
- https://github.com/tatsu-lab/stanford_alpaca
|
||||
- https://github.com/cornelk/llama-go for the initial ideas
|
||||
- https://github.com/antimatter15/alpaca.cpp
|
||||
- https://github.com/EdVince/Stable-Diffusion-NCNN
|
||||
- https://github.com/ggerganov/whisper.cpp
|
||||
- https://github.com/saharNooby/rwkv.cpp
|
||||
- https://github.com/rhasspy/piper
|
||||
|
||||
## 🤗 Contributors
|
||||
|
||||
This is a community project, a special thanks to our contributors! 🤗
|
||||
<a href="https://github.com/go-skynet/LocalAI/graphs/contributors">
|
||||
<img src="https://contrib.rocks/image?repo=go-skynet/LocalAI" />
|
||||
</a>
|
||||
LocalAI is MIT licensed, created and maintained by [Ettore Di Giacinto](https://github.com/mudler).
|
||||
|
@ -25,7 +25,10 @@ hero:
|
||||
height: 540px
|
||||
|
||||
title: ""
|
||||
subtitle: "**The free, OpenAI, Anthropic, et al alternative. Your All-in-One Complete AI Stack** - Run powerful language models, autonomous agents, and document intelligence **locally** on your hardware. **No cloud, no limits, no compromise.**"
|
||||
subtitle: |
|
||||
**The free, OpenAI, Anthropic alternative. Your All-in-One Complete AI Stack** - Run powerful language models, autonomous agents, and document intelligence **locally** on your hardware.
|
||||
|
||||
**No cloud, no limits, no compromise.**
|
||||
|
||||
image:
|
||||
path: "images"
|
||||
@ -44,7 +47,10 @@ hero:
|
||||
url: "https://github.com/mudler/LocalAI"
|
||||
|
||||
info: |
|
||||
**Drop-in replacement for OpenAI API** - modular suite of tools that work seamlessly together or independently. Start with LocalAI's OpenAI-compatible API, extend with autonomous agents, and enhance with semantic search - all running locally on your hardware.
|
||||
**Drop-in replacement for OpenAI API** - modular suite of tools that work seamlessly together or independently.
|
||||
|
||||
Start with **[LocalAI](https://localai.io)**'s OpenAI-compatible API, extend with **[LocalAGI](https://github.com/mudler/LocalAGI)**'s autonomous agents, and enhance with **[LocalRecall](https://github.com/mudler/LocalRecall)**'s semantic search - all running locally on your hardware.
|
||||
|
||||
**Open Source** MIT Licensed.
|
||||
|
||||
# Feature Grid
|
||||
@ -59,21 +65,22 @@ featureGrid:
|
||||
|
||||
items:
|
||||
- title: LLM Inferencing
|
||||
icon: api
|
||||
icon: memory_alt
|
||||
description: LocalAI is a free, **Open Source** OpenAI alternative. Run **LLMs**, generate **images**, **audio** and more **locally** with consumer grade hardware.
|
||||
ctaLink:
|
||||
text: learn more
|
||||
url: /basics/getting_started/
|
||||
- title: Agentic-first
|
||||
icon: ai
|
||||
icon: smart_toy
|
||||
description: |
|
||||
Extend LocalAI with LocalAGI, an autonomous AI agent platform that runs locally. Build and deploy autonomous agents with ease. Interact with REST APIs or use the WebUI.
|
||||
Extend LocalAI with LocalAGI, an autonomous AI agent platform that runs locally, no coding required.
|
||||
Build and deploy autonomous agents with ease. Interact with REST APIs or use the WebUI.
|
||||
ctaLink:
|
||||
text: learn more
|
||||
url: https://github.com/mudler/LocalAGI
|
||||
|
||||
- title: Memory and Knowledge base
|
||||
icon: memory
|
||||
icon: psychology
|
||||
description:
|
||||
Extend LocalAI with LocalRecall, A local rest api for semantic search and memory management. Perfect for AI applications.
|
||||
ctaLink:
|
||||
@ -99,21 +106,21 @@ featureGrid:
|
||||
description: |
|
||||
Support for various model families including LLMs, image generation, and audio models.
|
||||
Supports multiple backends for inferencing, including vLLM, llama.cpp, and more.
|
||||
You can switch between them as needed.
|
||||
You can switch between them as needed and install them from the Web interface or the CLI.
|
||||
ctaLink:
|
||||
text: learn more
|
||||
url: /model-compatibility
|
||||
|
||||
- title: Privacy Focused
|
||||
icon: lock
|
||||
icon: security
|
||||
description: Keep your data local. No data leaves your machine, ensuring complete privacy.
|
||||
ctaLink:
|
||||
text: learn more
|
||||
url: /basics/getting_started/
|
||||
url: /basics/container/
|
||||
|
||||
- title: Easy Setup
|
||||
icon: settings
|
||||
description: Simple installation and configuration. Get started in minutes with Docker or local installation.
|
||||
description: Simple installation and configuration. Get started in minutes with Binaries installation, Docker, Podman, Kubernetes or local installation.
|
||||
ctaLink:
|
||||
text: learn more
|
||||
url: /basics/getting_started/
|
||||
@ -125,12 +132,7 @@ featureGrid:
|
||||
text: learn more
|
||||
url: https://github.com/mudler/LocalAI
|
||||
|
||||
- title: Flexible Deployment
|
||||
icon: deploy
|
||||
description: Deploy anywhere - from your laptop to your server. Supports Docker, Kubernetes, and more.
|
||||
ctaLink:
|
||||
text: learn more
|
||||
url: /basics/container/
|
||||
|
||||
|
||||
- title: Extensible
|
||||
icon: extension
|
||||
@ -140,7 +142,7 @@ featureGrid:
|
||||
url: /docs/integrations/
|
||||
|
||||
- title: Peer 2 Peer
|
||||
icon: network
|
||||
icon: hub
|
||||
description: |
|
||||
LocalAI is designed to be a decentralized LLM inference, powered by a peer-to-peer system based on libp2p.
|
||||
It is designed to be used in a local or remote network, and is compatible with any LLM model.
|
||||
@ -162,7 +164,8 @@ imageText:
|
||||
template: image text
|
||||
|
||||
title: Run AI models locally with ease
|
||||
subtitle: LocalAI makes it simple to run various AI models on your own hardware, from text generation to image creation.
|
||||
subtitle: |
|
||||
LocalAI makes it simple to run various AI models on your own hardware. From text generation to image creation, autonomous agents to semantic search - all orchestrated through a unified API.
|
||||
|
||||
list:
|
||||
- text: OpenAI API compatibility
|
||||
@ -171,13 +174,40 @@ imageText:
|
||||
- text: Multiple model support
|
||||
icon: hub
|
||||
|
||||
- text: Image understanding
|
||||
icon: image
|
||||
|
||||
- text: Image generation
|
||||
icon: image
|
||||
|
||||
- text: Audio generation
|
||||
icon: music_note
|
||||
|
||||
- text: Voice activity detection
|
||||
icon: mic
|
||||
|
||||
- text: Speech recognition
|
||||
icon: mic
|
||||
|
||||
- text: Video generation
|
||||
icon: movie
|
||||
|
||||
- text: Privacy focused
|
||||
icon: lock
|
||||
icon: security
|
||||
|
||||
- text: Autonomous agents with [LocalAGI](https://github.com/mudler/LocalAGI)
|
||||
icon: smart_toy
|
||||
|
||||
- text: Semantic search with [LocalRecall](https://github.com/mudler/LocalRecall)
|
||||
icon: psychology
|
||||
|
||||
- text: Agent orchestration
|
||||
icon: hub
|
||||
|
||||
image:
|
||||
path: "images"
|
||||
filename: "localai_architecture.png"
|
||||
alt: "LocalAI Architecture"
|
||||
filename: "imagen.png"
|
||||
alt: "LocalAI Image generation"
|
||||
|
||||
imgOrder:
|
||||
desktop: 2
|
||||
@ -189,7 +219,7 @@ imageText:
|
||||
|
||||
# Image compare
|
||||
imageCompare:
|
||||
enable: true
|
||||
enable: false
|
||||
weight: 30
|
||||
template: image compare
|
||||
|
||||
@ -204,8 +234,8 @@ imageCompare:
|
||||
addCircleBlur: false,
|
||||
showLabels: true,
|
||||
labelOptions: {
|
||||
before: 'Input',
|
||||
after: 'Output',
|
||||
before: 'Dark',
|
||||
after: 'Light',
|
||||
onHover: false
|
||||
}
|
||||
}
|
||||
@ -217,17 +247,17 @@ imageCompare:
|
||||
config: {
|
||||
startingPoint: 50,
|
||||
addCircle: true,
|
||||
addCircleBlur: false,
|
||||
addCircleBlur: true,
|
||||
showLabels: true,
|
||||
labelOptions: {
|
||||
before: 'Prompt',
|
||||
after: 'Result',
|
||||
onHover: false
|
||||
onHover: true
|
||||
}
|
||||
}
|
||||
imagePath: "images/screenshots"
|
||||
imageBefore: "image_generation_prompt.webp"
|
||||
imageAfter: "image_generation_result.webp"
|
||||
imageBefore: "imagen_before.webp"
|
||||
imageAfter: "imagen_after.webp"
|
||||
|
||||
- title: Audio Generation
|
||||
config: {
|
||||
|
57
docs/layouts/partials/header.html
Normal file
@ -0,0 +1,57 @@
|
||||
<!-- Navbar Start -->
|
||||
<header id="topnav">
|
||||
<div class="container d-flex justify-content-between align-items-center">
|
||||
<!-- Logo container-->
|
||||
<a class="logo" aria-label="Home" href='{{ relLangURL "" }}'>
|
||||
|
||||
</a>
|
||||
<!-- End Logo container-->
|
||||
|
||||
<div class="d-flex align-items-center">
|
||||
|
||||
<div id="navigation">
|
||||
<!-- Navigation Menu -->
|
||||
<ul class="navigation-menu nav-right">
|
||||
{{- range .Site.Menus.primary }}
|
||||
<li><a href="{{ relLangURL .URL }}">{{ .Name }}</a></li>
|
||||
{{ end }}
|
||||
</ul><!--end navigation menu-->
|
||||
</div><!--end navigation-->
|
||||
|
||||
<!-- Social Links Start -->
|
||||
{{ with $.Scratch.Get "social_list" }}
|
||||
<ul class="social-link d-flex list-inline mb-0">
|
||||
{{ range . }}
|
||||
{{ $path := printf "images/social/%s.%s" . "svg" }}
|
||||
<li class="list-inline-item mb-0">
|
||||
<a href="{{ if eq . `rss` }} {{ `index.xml` | absURL }} {{ else if eq . `bluesky` }} https://bsky.app/profile/{{ index site.Params.social . }} {{ else }} https://{{ . }}.com/{{ index site.Params.social . }} {{ end }}" alt="{{ . }}" rel="noopener noreferrer" target="_blank">
|
||||
<div class="btn btn-icon btn-landing border-0">
|
||||
{{ with resources.Get $path }}
|
||||
{{ .Content | safeHTML }}
|
||||
{{ end }}
|
||||
</div>
|
||||
</a>
|
||||
</li>
|
||||
{{ end }}
|
||||
</ul>
|
||||
{{ end }}
|
||||
<!-- Social Links End -->
|
||||
|
||||
<div class="menu-extras ms-3 me-2">
|
||||
<div class="menu-item">
|
||||
<!-- Mobile menu toggle-->
|
||||
<button class="navbar-toggle btn btn-icon btn-soft-light" id="isToggle" aria-label="toggleMenu" onclick="toggleMenu()">
|
||||
<div class="lines">
|
||||
<span></span>
|
||||
<span></span>
|
||||
<span></span>
|
||||
</div>
|
||||
</button>
|
||||
<!-- End mobile menu toggle-->
|
||||
</div>
|
||||
</div>
|
||||
|
||||
</div>
|
||||
</div><!--end container-->
|
||||
</header><!--end header-->
|
||||
<!-- Navbar End -->
|
BIN
docs/static/android-chrome-192x192.png
vendored
Before Width: | Height: | Size: 57 KiB After Width: | Height: | Size: 16 KiB |
BIN
docs/static/android-chrome-512x512.png
vendored
Before Width: | Height: | Size: 359 KiB After Width: | Height: | Size: 30 KiB |
BIN
docs/static/apple-touch-icon.png
vendored
Before Width: | Height: | Size: 52 KiB After Width: | Height: | Size: 14 KiB |
BIN
docs/static/favicon-16x16.png
vendored
Before Width: | Height: | Size: 769 B After Width: | Height: | Size: 711 B |
BIN
docs/static/favicon-32x32.png
vendored
Before Width: | Height: | Size: 2.3 KiB After Width: | Height: | Size: 1.7 KiB |
BIN
docs/static/favicon.ico
vendored
Before Width: | Height: | Size: 15 KiB After Width: | Height: | Size: 15 KiB |
171
docs/static/favicon.svg
vendored
Normal file
After Width: | Height: | Size: 108 KiB |
1
docs/static/site.webmanifest
vendored
Normal file
@ -0,0 +1 @@
|
||||
{"name":"","short_name":"","icons":[{"src":"/android-chrome-192x192.png","sizes":"192x192","type":"image/png"},{"src":"/android-chrome-512x512.png","sizes":"512x512","type":"image/png"}],"theme_color":"#ffffff","background_color":"#ffffff","display":"standalone"}
|