## π Introduction
With OpenLLM, you can run inference with any open-source large-language models,
deploy to the cloud or on-premises, and build powerful AI apps.
π **State-of-the-art LLMs**: built-in supports a wide range of open-source LLMs
and model runtime, including Llama 2οΌStableLM, Falcon, Dolly, Flan-T5, ChatGLM,
StarCoder and more.
π₯ **Flexible APIs**: serve LLMs over RESTful API or gRPC with one command,
query via WebUI, CLI, our Python/Javascript client, or any HTTP client.
βοΈ **Freedom To Build**: First-class support for LangChain, BentoML and Hugging
Face that allows you to easily create your own AI apps by composing LLMs with
other models and services.
π― **Streamline Deployment**: Automatically generate your LLM server Docker
Images or deploy as serverless endpoint via
[βοΈ BentoCloud](https://l.bentoml.com/bento-cloud).
π€οΈ **Bring your own LLM**: Fine-tune any LLM to suit your needs with
`LLM.tuning()`. (Coming soon)
