mirror of
https://github.com/bentoml/OpenLLM.git
synced 2026-03-11 03:33:44 -04:00
fix(dolly): remove loading as pipeline
Signed-off-by: aarnphm-ec2-dev <29749331+aarnphm@users.noreply.github.com>
This commit is contained in:
@@ -237,9 +237,6 @@ class DollyV2(openllm.LLM["transformers.Pipeline", "transformers.PreTrainedToken
|
||||
model = transformers.AutoModelForCausalLM.from_pretrained(_ref.path, **attrs)
|
||||
|
||||
kwds: dict[str, t.Any] = {}
|
||||
if self.bettertransformer:
|
||||
# This is a pipeline, provide a accelerator args
|
||||
kwds["accelerator"] = "bettertransformer"
|
||||
return InstructionTextGenerationPipeline(model=model, tokenizer=_ref.custom_objects["tokenizer"], **kwds)
|
||||
|
||||
def sanitize_parameters(
|
||||
|
||||
Reference in New Issue
Block a user