From fc103be6e86981f839fcf44d0d85232a82fa2e42 Mon Sep 17 00:00:00 2001 From: aarnphm-ec2-dev <29749331+aarnphm@users.noreply.github.com> Date: Mon, 26 Jun 2023 22:05:40 +0000 Subject: [PATCH] fix(dolly): remove loading as pipeline Signed-off-by: aarnphm-ec2-dev <29749331+aarnphm@users.noreply.github.com> --- src/openllm/models/dolly_v2/modeling_dolly_v2.py | 3 --- 1 file changed, 3 deletions(-) diff --git a/src/openllm/models/dolly_v2/modeling_dolly_v2.py b/src/openllm/models/dolly_v2/modeling_dolly_v2.py index d74a294d..4bed285d 100644 --- a/src/openllm/models/dolly_v2/modeling_dolly_v2.py +++ b/src/openllm/models/dolly_v2/modeling_dolly_v2.py @@ -237,9 +237,6 @@ class DollyV2(openllm.LLM["transformers.Pipeline", "transformers.PreTrainedToken model = transformers.AutoModelForCausalLM.from_pretrained(_ref.path, **attrs) kwds: dict[str, t.Any] = {} - if self.bettertransformer: - # This is a pipeline, provide a accelerator args - kwds["accelerator"] = "bettertransformer" return InstructionTextGenerationPipeline(model=model, tokenizer=_ref.custom_objects["tokenizer"], **kwds) def sanitize_parameters(