From 5eea40a5995bac4bbfaf389f8a284968ecec8db1 Mon Sep 17 00:00:00 2001
From: Aaron <29749331+aarnphm@users.noreply.github.com>
Date: Mon, 4 Sep 2023 15:02:01 -0400
Subject: [PATCH] chore(readme): update README for release [generated] [skip
ci]
Signed-off-by: Aaron <29749331+aarnphm@users.noreply.github.com>
---
openllm-python/README.md | 21 +++++++++++++--------
1 file changed, 13 insertions(+), 8 deletions(-)
diff --git a/openllm-python/README.md b/openllm-python/README.md
index cf7afaab..8ba82b92 100644
--- a/openllm-python/README.md
+++ b/openllm-python/README.md
@@ -22,7 +22,7 @@
-
+
@@ -407,17 +407,19 @@ pip install "openllm[baichuan]"
### Runtime Implementations (Experimental)
Different LLMs may have multiple runtime implementations. For instance, they
-might use Pytorch (`pt`), Tensorflow (`tf`), or Flax (`flax`).
+might use Pytorch (`pt`), Tensorflow (`tf`), Flax (`flax`) or vLLM (`vllm`).
If you wish to specify a particular runtime for a model, you can do so by
-setting the `OPENLLM_{MODEL_NAME}_FRAMEWORK={runtime}` environment variable
+setting the `OPENLLM_BACKEND={runtime}` environment variable
before running `openllm start`.
For example, if you want to use the Tensorflow (`tf`) implementation for the
`flan-t5` model, you can use the following command:
```bash
-OPENLLM_FLAN_T5_FRAMEWORK=tf openllm start flan-t5
+OPENLLM_BACKEND=tf openllm start flan-t5
+
+openllm start flan-t5 --backend tf
```
> [!NOTE]
@@ -425,6 +427,9 @@ OPENLLM_FLAN_T5_FRAMEWORK=tf openllm start flan-t5
> [Jax's installation](https://github.com/google/jax#pip-installation-gpu-cuda-installed-via-pip-easier)
> to make sure that you have Jax support for the corresponding CUDA version.
+> [!IMPORTANT]
+> To use vLLM backend, at least a GPU with Ampere or newer architecture and CUDA 11.8 is required.
+
### Quantisation
OpenLLM supports quantisation with
@@ -543,10 +548,10 @@ client.embed("I like to eat apples")
The following UIs are currently available for OpenLLM:
-| UI | Owner | Type | Progress |
-|-------------------------------------------------------------------------------------------|----------------------------------------------|----------------------|----------|
-| [Clojure](https://github.com/bentoml/OpenLLM/blob/main/openllm-contrib/clojure/README.md) | [@GutZuFusss](https://github.com/GutZuFusss) | Community-maintained | 🔧 |
-| TS | BentoML Team | | 🚧 |
+| UI | Owner | Type | Progress |
+|-----------------------------------------------------------------------------------|-----------------------------------------------|----------------------|----------|
+| [Clojure](https://github.com/bentoml/OpenLLM/blob/main/contrib/clojure/README.md) | [@GutZuFusss](https://github.com/GutZuFusss) | Community-maintained | 🔧 |
+| TS | BentoML Team | | 🚧 |
## ⚙️ Integrations