mirror of
https://github.com/bentoml/OpenLLM.git
synced 2026-01-24 23:37:48 -05:00
chore(deps): lock vLLM to 0.2.4
Signed-off-by: Aaron <29749331+aarnphm@users.noreply.github.com>
This commit is contained in:
@@ -39,7 +39,7 @@ classifiers = [
|
||||
]
|
||||
dependencies = [
|
||||
"bentoml[io]>=1.1.10",
|
||||
"transformers[torch,tokenizers]>=4.35.0",
|
||||
"transformers[torch,tokenizers]>=4.36.0",
|
||||
"openllm-client>=0.4.35",
|
||||
"openllm-core>=0.4.35",
|
||||
"safetensors",
|
||||
@@ -100,7 +100,7 @@ Tracker = "https://github.com/bentoml/OpenLLM/issues"
|
||||
Twitter = "https://twitter.com/bentomlai"
|
||||
|
||||
[project.optional-dependencies]
|
||||
agents = ["transformers[agents]>=4.35.0", "diffusers", "soundfile"]
|
||||
agents = ["transformers[agents]>=4.36.0", "diffusers", "soundfile"]
|
||||
all = ["openllm[full]"]
|
||||
awq = ["autoawq"]
|
||||
baichuan = ["cpm-kernels"]
|
||||
@@ -119,7 +119,7 @@ openai = ["openai[datalib]>=1", "tiktoken"]
|
||||
playground = ["jupyter", "notebook", "ipython", "jupytext", "nbformat"]
|
||||
qwen = ["cpm-kernels", "tiktoken"]
|
||||
starcoder = ["bitsandbytes"]
|
||||
vllm = ["vllm>=0.2.2"]
|
||||
vllm = ["vllm>=0.2.4"]
|
||||
|
||||
[tool.hatch.version]
|
||||
fallback-version = "0.0.0"
|
||||
|
||||
@@ -31,7 +31,7 @@ def build_editable(path, package='openllm'):
|
||||
raise RuntimeError('Please install OpenLLM from PyPI or built it from Git source.')
|
||||
def construct_python_options(llm, llm_fs, extra_dependencies=None, adapter_map=None):
|
||||
from . import RefResolver
|
||||
packages = ['scipy', 'bentoml[tracing]>=1.1.10', 'vllm==0.2.2', 'ray==2.6.0', f'openllm>={RefResolver.from_strategy("release").version}'] # apparently bnb misses this one
|
||||
packages = ['scipy', 'bentoml[tracing]>=1.1.10', 'vllm==0.2.4', 'ray==2.6.0', f'openllm>={RefResolver.from_strategy("release").version}'] # apparently bnb misses this one
|
||||
if adapter_map is not None: packages += ['openllm[fine-tune]']
|
||||
if extra_dependencies is not None: packages += [f'openllm[{k}]' for k in extra_dependencies]
|
||||
if llm.config['requirements'] is not None: packages.extend(llm.config['requirements'])
|
||||
@@ -53,9 +53,9 @@ def construct_docker_options(llm, _, quantize, adapter_map, dockerfile_template,
|
||||
return DockerOptions(cuda_version='12.1', env=environ, dockerfile_template=dockerfile_template)
|
||||
@inject
|
||||
def create_bento(
|
||||
bento_tag, llm_fs, llm,
|
||||
quantize, dockerfile_template,
|
||||
adapter_map=None, extra_dependencies=None, serialisation=None,
|
||||
bento_tag, llm_fs, llm, #
|
||||
quantize, dockerfile_template, #
|
||||
adapter_map=None, extra_dependencies=None, serialisation=None, #
|
||||
_bento_store=Provide[BentoMLContainer.bento_store], _model_store=Provide[BentoMLContainer.model_store],
|
||||
):
|
||||
_serialisation = openllm_core.utils.first_not_none(serialisation, default=llm.config['serialisation'])
|
||||
|
||||
Reference in New Issue
Block a user