mirror of
https://github.com/bentoml/OpenLLM.git
synced 2026-04-19 14:40:46 -04:00
chore(deps): lock vLLM to 0.2.4
Signed-off-by: Aaron <29749331+aarnphm@users.noreply.github.com>
This commit is contained in:
@@ -39,7 +39,7 @@ classifiers = [
|
||||
]
|
||||
dependencies = [
|
||||
"bentoml[io]>=1.1.10",
|
||||
"transformers[torch,tokenizers]>=4.35.0",
|
||||
"transformers[torch,tokenizers]>=4.36.0",
|
||||
"openllm-client>=0.4.35",
|
||||
"openllm-core>=0.4.35",
|
||||
"safetensors",
|
||||
@@ -100,7 +100,7 @@ Tracker = "https://github.com/bentoml/OpenLLM/issues"
|
||||
Twitter = "https://twitter.com/bentomlai"
|
||||
|
||||
[project.optional-dependencies]
|
||||
agents = ["transformers[agents]>=4.35.0", "diffusers", "soundfile"]
|
||||
agents = ["transformers[agents]>=4.36.0", "diffusers", "soundfile"]
|
||||
all = ["openllm[full]"]
|
||||
awq = ["autoawq"]
|
||||
baichuan = ["cpm-kernels"]
|
||||
@@ -119,7 +119,7 @@ openai = ["openai[datalib]>=1", "tiktoken"]
|
||||
playground = ["jupyter", "notebook", "ipython", "jupytext", "nbformat"]
|
||||
qwen = ["cpm-kernels", "tiktoken"]
|
||||
starcoder = ["bitsandbytes"]
|
||||
vllm = ["vllm>=0.2.2"]
|
||||
vllm = ["vllm>=0.2.4"]
|
||||
|
||||
[tool.hatch.version]
|
||||
fallback-version = "0.0.0"
|
||||
|
||||
@@ -31,7 +31,7 @@ def build_editable(path, package='openllm'):
|
||||
raise RuntimeError('Please install OpenLLM from PyPI or built it from Git source.')
|
||||
def construct_python_options(llm, llm_fs, extra_dependencies=None, adapter_map=None):
|
||||
from . import RefResolver
|
||||
packages = ['scipy', 'bentoml[tracing]>=1.1.10', 'vllm==0.2.2', 'ray==2.6.0', f'openllm>={RefResolver.from_strategy("release").version}'] # apparently bnb misses this one
|
||||
packages = ['scipy', 'bentoml[tracing]>=1.1.10', 'vllm==0.2.4', 'ray==2.6.0', f'openllm>={RefResolver.from_strategy("release").version}'] # apparently bnb misses this one
|
||||
if adapter_map is not None: packages += ['openllm[fine-tune]']
|
||||
if extra_dependencies is not None: packages += [f'openllm[{k}]' for k in extra_dependencies]
|
||||
if llm.config['requirements'] is not None: packages.extend(llm.config['requirements'])
|
||||
@@ -53,9 +53,9 @@ def construct_docker_options(llm, _, quantize, adapter_map, dockerfile_template,
|
||||
return DockerOptions(cuda_version='12.1', env=environ, dockerfile_template=dockerfile_template)
|
||||
@inject
|
||||
def create_bento(
|
||||
bento_tag, llm_fs, llm,
|
||||
quantize, dockerfile_template,
|
||||
adapter_map=None, extra_dependencies=None, serialisation=None,
|
||||
bento_tag, llm_fs, llm, #
|
||||
quantize, dockerfile_template, #
|
||||
adapter_map=None, extra_dependencies=None, serialisation=None, #
|
||||
_bento_store=Provide[BentoMLContainer.bento_store], _model_store=Provide[BentoMLContainer.model_store],
|
||||
):
|
||||
_serialisation = openllm_core.utils.first_not_none(serialisation, default=llm.config['serialisation'])
|
||||
|
||||
@@ -144,7 +144,7 @@ class Dependencies:
|
||||
_LOWER_BENTOML_CONSTRAINT = '1.1.10'
|
||||
_BENTOML_EXT = ['io']
|
||||
_TRANSFORMERS_EXT = ['torch', 'tokenizers']
|
||||
_TRANSFORMERS_CONSTRAINTS = '4.35.0'
|
||||
_TRANSFORMERS_CONSTRAINTS = '4.36.0'
|
||||
|
||||
FINE_TUNE_DEPS = ['peft>=0.6.0', 'datasets', 'trl', 'huggingface-hub']
|
||||
GRPC_DEPS = [f'bentoml[grpc]>={_LOWER_BENTOML_CONSTRAINT}']
|
||||
@@ -155,7 +155,7 @@ GGML_DEPS = ['ctransformers']
|
||||
CTRANSLATE_DEPS = ['ctranslate2>=3.22.0']
|
||||
AWQ_DEPS = ['autoawq']
|
||||
GPTQ_DEPS = ['auto-gptq[triton]>=0.4.2']
|
||||
VLLM_DEPS = ['vllm>=0.2.2']
|
||||
VLLM_DEPS = ['vllm>=0.2.4']
|
||||
|
||||
_base_requirements: dict[str, t.Any] = {
|
||||
inflection.dasherize(name): config_cls.__openllm_requirements__
|
||||
|
||||
Reference in New Issue
Block a user