mirror of
https://github.com/mudler/LocalAI.git
synced 2026-04-02 06:04:09 -04:00
* feat: add distributed mode (experimental) Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fix data races, mutexes, transactions Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * refactorings Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fixups Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fix events and tool stream in agent chat Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * use ginkgo Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * refactoring and consolidation Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * refactoring and consolidation Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * refactoring and consolidation Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * refactoring and consolidation Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * refactoring and consolidation Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * refactoring and consolidation Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * refactoring and consolidation Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * refactoring and consolidation Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fix(cron): compute correctly time boundaries avoiding re-triggering Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * enhancements, refactorings Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * do not flood of healthy checks Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * do not list obvious backends as text backends Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * tests fixups Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * refactoring and consolidation Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Drop redundant healthcheck Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * enhancements, refactorings Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
93 lines
4.1 KiB
Markdown
93 lines
4.1 KiB
Markdown
+++
|
|
title = "Overview"
|
|
weight = 1
|
|
toc = true
|
|
description = "What is LocalAI?"
|
|
tags = ["Beginners"]
|
|
categories = [""]
|
|
url = "/docs/overview"
|
|
author = "Ettore Di Giacinto"
|
|
icon = "info"
|
|
+++
|
|
|
|
|
|
LocalAI is your complete AI stack for running AI models locally. It's designed to be simple, efficient, and accessible, providing a drop-in replacement for OpenAI's API while keeping your data private and secure.
|
|
|
|
## Why LocalAI?
|
|
|
|
In today's AI landscape, privacy, control, and flexibility are paramount. LocalAI addresses these needs by:
|
|
|
|
- **Privacy First**: Your data never leaves your machine
|
|
- **Complete Control**: Run models on your terms, with your hardware
|
|
- **Open Source**: MIT licensed and community-driven
|
|
- **Flexible Deployment**: From laptops to servers, with or without GPUs
|
|
- **Extensible**: Add new models and features as needed
|
|
|
|
## What's Included
|
|
|
|
LocalAI is a single binary (or container) that gives you everything you need:
|
|
|
|
- **OpenAI-compatible API** — Drop-in replacement for OpenAI, Anthropic, and Open Responses APIs
|
|
- **Built-in Web Interface** — Chat, model management, agent creation, image generation, and system monitoring
|
|
- **AI Agents** — Create autonomous agents with MCP (Model Context Protocol) tool support, directly from the UI
|
|
- **Multiple Model Support** — LLMs, image generation, text-to-speech, speech-to-text, vision, embeddings, and more
|
|
- **GPU Acceleration** — Automatic detection and support for NVIDIA, AMD, Intel, and Vulkan GPUs
|
|
- **Distributed Mode** — Scale horizontally with worker nodes, P2P federation, and model sharding
|
|
- **No GPU Required** — Runs on CPU with consumer-grade hardware
|
|
|
|
LocalAI integrates [LocalAGI](https://github.com/mudler/LocalAGI) (agent platform) and [LocalRecall](https://github.com/mudler/LocalRecall) (semantic memory) as built-in libraries — no separate installation needed.
|
|
|
|
## Getting Started
|
|
|
|
LocalAI can be installed in several ways. **Docker is the recommended installation method** for most users as it provides the easiest setup and works across all platforms.
|
|
|
|
### Recommended: Docker Installation
|
|
|
|
The quickest way to get started with LocalAI is using Docker:
|
|
|
|
```bash
|
|
docker run -p 8080:8080 --name local-ai -ti localai/localai:latest-cpu
|
|
```
|
|
|
|
Then open **http://localhost:8080** to access the web interface, install models, and start chatting.
|
|
|
|
For GPU support, see the [Container images reference]({{% relref "getting-started/container-images" %}}) or the [Quickstart guide]({{% relref "getting-started/quickstart" %}}).
|
|
|
|
For complete installation instructions including Docker, macOS, Linux, Kubernetes, and building from source, see the [Installation guide](/installation/).
|
|
|
|
## Key Features
|
|
|
|
- **Text Generation**: Run various LLMs locally (llama.cpp, transformers, vLLM, and more)
|
|
- **Image Generation**: Create images with Stable Diffusion, Flux, and other models
|
|
- **Audio Processing**: Text-to-speech and speech-to-text
|
|
- **Vision API**: Image understanding and analysis
|
|
- **Embeddings**: Vector representations for search and retrieval
|
|
- **Function Calling**: OpenAI-compatible tool use
|
|
- **AI Agents**: Autonomous agents with MCP tool support
|
|
- **MCP Apps**: Interactive tool UIs in the web interface
|
|
- **P2P & Distributed**: Federated inference and model sharding across machines
|
|
|
|
## Community and Support
|
|
|
|
LocalAI is a community-driven project. You can:
|
|
|
|
- Join our [Discord community](https://discord.gg/uJAeKSAGDy)
|
|
- Check out our [GitHub repository](https://github.com/mudler/LocalAI)
|
|
- Contribute to the project
|
|
- Share your use cases and examples
|
|
|
|
## Next Steps
|
|
|
|
Ready to dive in? Here are some recommended next steps:
|
|
|
|
1. **[Install LocalAI](/installation/)** - Start with [Docker installation](/installation/docker/) (recommended) or choose another method
|
|
2. **[Quickstart guide]({{% relref "getting-started/quickstart" %}})** - Get up and running in minutes
|
|
3. [Explore available models](https://models.localai.io)
|
|
4. [Model compatibility](/model-compatibility/)
|
|
5. [Try out examples]({{% relref "getting-started/try-it-out" %}})
|
|
6. [Join the community](https://discord.gg/uJAeKSAGDy)
|
|
|
|
## License
|
|
|
|
LocalAI is MIT licensed, created and maintained by [Ettore Di Giacinto](https://github.com/mudler).
|