mirror of
https://github.com/bentoml/OpenLLM.git
synced 2026-03-07 08:38:20 -05:00
fix(serialisation): vLLM safetensors support (#324)
* fix(serilisation): vllm support for safetensors Signed-off-by: aarnphm-ec2-dev <29749331+aarnphm@users.noreply.github.com> * chore: running tools Signed-off-by: Aaron <29749331+aarnphm@users.noreply.github.com> * chore: generalize one shot generation Signed-off-by: Aaron <29749331+aarnphm@users.noreply.github.com> * chore: add changelog [skip ci] Signed-off-by: paperspace <29749331+aarnphm@users.noreply.github.com> --------- Signed-off-by: aarnphm-ec2-dev <29749331+aarnphm@users.noreply.github.com> Signed-off-by: Aaron <29749331+aarnphm@users.noreply.github.com> Signed-off-by: paperspace <29749331+aarnphm@users.noreply.github.com>
This commit is contained in:
@@ -19,8 +19,8 @@ class GPTNeoX(metaclass=_DummyMetaclass):
|
||||
_backends=["torch"]
|
||||
def __init__(self,*param_decls:_t.Any,**attrs: _t.Any):_require_backends(self,["torch"])
|
||||
class Llama(metaclass=_DummyMetaclass):
|
||||
_backends=["torch","fairscale","sentencepiece"]
|
||||
def __init__(self,*param_decls:_t.Any,**attrs: _t.Any):_require_backends(self,["torch","fairscale","sentencepiece"])
|
||||
_backends=["torch","fairscale","sentencepiece","scipy"]
|
||||
def __init__(self,*param_decls:_t.Any,**attrs: _t.Any):_require_backends(self,["torch","fairscale","sentencepiece","scipy"])
|
||||
class MPT(metaclass=_DummyMetaclass):
|
||||
_backends=["torch","triton","einops"]
|
||||
def __init__(self,*param_decls:_t.Any,**attrs: _t.Any):_require_backends(self,["torch","triton","einops"])
|
||||
|
||||
@@ -28,8 +28,8 @@ class VLLMStarCoder(metaclass=_DummyMetaclass):
|
||||
_backends=["vllm","bitsandbytes"]
|
||||
def __init__(self,*param_decls:_t.Any,**attrs: _t.Any):_require_backends(self,["vllm","bitsandbytes"])
|
||||
class VLLMLlama(metaclass=_DummyMetaclass):
|
||||
_backends=["vllm","fairscale","sentencepiece"]
|
||||
def __init__(self,*param_decls:_t.Any,**attrs: _t.Any):_require_backends(self,["vllm","fairscale","sentencepiece"])
|
||||
_backends=["vllm","fairscale","sentencepiece","scipy"]
|
||||
def __init__(self,*param_decls:_t.Any,**attrs: _t.Any):_require_backends(self,["vllm","fairscale","sentencepiece","scipy"])
|
||||
class AutoVLLM(metaclass=_DummyMetaclass):
|
||||
_backends=["vllm"]
|
||||
def __init__(self,*param_decls:_t.Any,**attrs: _t.Any):_require_backends(self,["vllm"])
|
||||
|
||||
Reference in New Issue
Block a user