From a32cf324d844cb12b9a6dbbff34badd777b60ed6 Mon Sep 17 00:00:00 2001 From: Aaron Pham <29749331+aarnphm@users.noreply.github.com> Date: Thu, 14 Sep 2023 03:42:28 -0400 Subject: [PATCH] fix(prompt): correct export extra objects items (#351) Signed-off-by: Aaron <29749331+aarnphm@users.noreply.github.com> --- openllm-python/src/openllm/models/auto/__init__.py | 7 ++++++- openllm-python/src/openllm/models/baichuan/__init__.py | 9 ++++++++- openllm-python/src/openllm/models/chatglm/__init__.py | 7 ++++++- openllm-python/src/openllm/models/dolly_v2/__init__.py | 7 ++++++- openllm-python/src/openllm/models/falcon/__init__.py | 7 ++++++- openllm-python/src/openllm/models/gpt_neox/__init__.py | 7 ++++++- openllm-python/src/openllm/models/llama/__init__.py | 10 +++++++++- openllm-python/src/openllm/models/mpt/__init__.py | 10 +++++++++- openllm-python/src/openllm/models/opt/__init__.py | 7 ++++++- openllm-python/src/openllm/models/stablelm/__init__.py | 9 ++++++++- .../src/openllm/models/starcoder/__init__.py | 9 ++++++++- 11 files changed, 78 insertions(+), 11 deletions(-) diff --git a/openllm-python/src/openllm/models/auto/__init__.py b/openllm-python/src/openllm/models/auto/__init__.py index 5e6f1f43..bf61359b 100644 --- a/openllm-python/src/openllm/models/auto/__init__.py +++ b/openllm-python/src/openllm/models/auto/__init__.py @@ -53,7 +53,12 @@ else: _import_structure['modeling_tf_auto'].extend(['AutoTFLLM', 'MODEL_TF_MAPPING']) if t.TYPE_CHECKING: from .modeling_tf_auto import MODEL_TF_MAPPING as MODEL_TF_MAPPING, AutoTFLLM as AutoTFLLM -__lazy = LazyModule(__name__, os.path.abspath('__file__'), _import_structure) +__lazy = LazyModule(__name__, + os.path.abspath('__file__'), + _import_structure, + extra_objects={ + 'CONFIG_MAPPING': CONFIG_MAPPING, 'CONFIG_MAPPING_NAMES': CONFIG_MAPPING_NAMES, 'AutoConfig': AutoConfig, + }) __all__ = __lazy.__all__ __dir__ = __lazy.__dir__ __getattr__ = __lazy.__getattr__ diff --git a/openllm-python/src/openllm/models/baichuan/__init__.py b/openllm-python/src/openllm/models/baichuan/__init__.py index 16c1ee2f..2682ddf3 100644 --- a/openllm-python/src/openllm/models/baichuan/__init__.py +++ b/openllm-python/src/openllm/models/baichuan/__init__.py @@ -27,4 +27,11 @@ else: _import_structure['modeling_vllm_baichuan'] = ['VLLMBaichuan'] if t.TYPE_CHECKING: from .modeling_vllm_baichuan import VLLMBaichuan as VLLMBaichuan -sys.modules[__name__] = LazyModule(__name__, globals()['__file__'], _import_structure) +sys.modules[__name__] = LazyModule(__name__, + globals()['__file__'], + _import_structure, + extra_objects={ + 'DEFAULT_PROMPT_TEMPLATE': DEFAULT_PROMPT_TEMPLATE, + 'START_BAICHUAN_COMMAND_DOCSTRING': START_BAICHUAN_COMMAND_DOCSTRING, + 'BaichuanConfig': BaichuanConfig + }) diff --git a/openllm-python/src/openllm/models/chatglm/__init__.py b/openllm-python/src/openllm/models/chatglm/__init__.py index 3bc256f3..9dcafb02 100644 --- a/openllm-python/src/openllm/models/chatglm/__init__.py +++ b/openllm-python/src/openllm/models/chatglm/__init__.py @@ -19,4 +19,9 @@ else: _import_structure['modeling_chatglm'] = ['ChatGLM'] if t.TYPE_CHECKING: from .modeling_chatglm import ChatGLM as ChatGLM -sys.modules[__name__] = LazyModule(__name__, globals()['__file__'], _import_structure) +sys.modules[__name__] = LazyModule(__name__, + globals()['__file__'], + _import_structure, + extra_objects={ + 'DEFAULT_PROMPT_TEMPLATE': DEFAULT_PROMPT_TEMPLATE, 'START_CHATGLM_COMMAND_DOCSTRING': START_CHATGLM_COMMAND_DOCSTRING, 'ChatGLMConfig': ChatGLMConfig + }) diff --git a/openllm-python/src/openllm/models/dolly_v2/__init__.py b/openllm-python/src/openllm/models/dolly_v2/__init__.py index b9dbce05..c48243d3 100644 --- a/openllm-python/src/openllm/models/dolly_v2/__init__.py +++ b/openllm-python/src/openllm/models/dolly_v2/__init__.py @@ -26,4 +26,9 @@ else: _import_structure['modeling_vllm_dolly_v2'] = ['VLLMDollyV2'] if t.TYPE_CHECKING: from .modeling_vllm_dolly_v2 import VLLMDollyV2 as VLLMDollyV2 -sys.modules[__name__] = LazyModule(__name__, globals()['__file__'], _import_structure) +sys.modules[__name__] = LazyModule(__name__, + globals()['__file__'], + _import_structure, + extra_objects={ + 'DEFAULT_PROMPT_TEMPLATE': DEFAULT_PROMPT_TEMPLATE, 'START_DOLLY_V2_COMMAND_DOCSTRING': START_DOLLY_V2_COMMAND_DOCSTRING, 'DollyV2Config': DollyV2Config + }) diff --git a/openllm-python/src/openllm/models/falcon/__init__.py b/openllm-python/src/openllm/models/falcon/__init__.py index 7587914e..d553c4a7 100644 --- a/openllm-python/src/openllm/models/falcon/__init__.py +++ b/openllm-python/src/openllm/models/falcon/__init__.py @@ -26,4 +26,9 @@ else: _import_structure['modeling_vllm_falcon'] = ['VLLMFalcon'] if t.TYPE_CHECKING: from .modeling_vllm_falcon import VLLMFalcon as VLLMFalcon -sys.modules[__name__] = LazyModule(__name__, globals()['__file__'], _import_structure) +sys.modules[__name__] = LazyModule(__name__, + globals()['__file__'], + _import_structure, + extra_objects={ + 'DEFAULT_PROMPT_TEMPLATE': DEFAULT_PROMPT_TEMPLATE, 'START_FALCON_COMMAND_DOCSTRING': START_FALCON_COMMAND_DOCSTRING, 'FalconConfig': FalconConfig + }) diff --git a/openllm-python/src/openllm/models/gpt_neox/__init__.py b/openllm-python/src/openllm/models/gpt_neox/__init__.py index f9b7de52..437645e1 100644 --- a/openllm-python/src/openllm/models/gpt_neox/__init__.py +++ b/openllm-python/src/openllm/models/gpt_neox/__init__.py @@ -26,4 +26,9 @@ else: _import_structure['modeling_vllm_gpt_neox'] = ['VLLMGPTNeoX'] if t.TYPE_CHECKING: from .modeling_vllm_gpt_neox import VLLMGPTNeoX as VLLMGPTNeoX -sys.modules[__name__] = LazyModule(__name__, globals()['__file__'], _import_structure) +sys.modules[__name__] = LazyModule(__name__, + globals()['__file__'], + _import_structure, + extra_objects={ + 'DEFAULT_PROMPT_TEMPLATE': DEFAULT_PROMPT_TEMPLATE, 'START_GPT_NEOX_COMMAND_DOCSTRING': START_GPT_NEOX_COMMAND_DOCSTRING, 'GPTNeoXConfig': GPTNeoXConfig + }) diff --git a/openllm-python/src/openllm/models/llama/__init__.py b/openllm-python/src/openllm/models/llama/__init__.py index 5e66177a..764e0c6b 100644 --- a/openllm-python/src/openllm/models/llama/__init__.py +++ b/openllm-python/src/openllm/models/llama/__init__.py @@ -27,4 +27,12 @@ else: _import_structure['modeling_llama'] = ['Llama'] if t.TYPE_CHECKING: from .modeling_llama import Llama as Llama -sys.modules[__name__] = LazyModule(__name__, globals()['__file__'], _import_structure) +sys.modules[__name__] = LazyModule(__name__, + globals()['__file__'], + _import_structure, + extra_objects={ + 'DEFAULT_PROMPT_TEMPLATE': DEFAULT_PROMPT_TEMPLATE, + 'START_LLAMA_COMMAND_DOCSTRING': START_LLAMA_COMMAND_DOCSTRING, + 'LlamaConfig': LlamaConfig, + 'PROMPT_MAPPING': PROMPT_MAPPING + }) diff --git a/openllm-python/src/openllm/models/mpt/__init__.py b/openllm-python/src/openllm/models/mpt/__init__.py index 1d578d82..e5782f2e 100644 --- a/openllm-python/src/openllm/models/mpt/__init__.py +++ b/openllm-python/src/openllm/models/mpt/__init__.py @@ -27,4 +27,12 @@ else: _import_structure['modeling_vllm_mpt'] = ['VLLMMPT'] if t.TYPE_CHECKING: from .modeling_vllm_mpt import VLLMMPT as VLLMMPT -sys.modules[__name__] = LazyModule(__name__, globals()['__file__'], _import_structure) +sys.modules[__name__] = LazyModule(__name__, + globals()['__file__'], + _import_structure, + extra_objects={ + 'DEFAULT_PROMPT_TEMPLATE': DEFAULT_PROMPT_TEMPLATE, + 'START_MPT_COMMAND_DOCSTRING': START_MPT_COMMAND_DOCSTRING, + 'MPTConfig': MPTConfig, + 'PROMPT_MAPPING': PROMPT_MAPPING + }) diff --git a/openllm-python/src/openllm/models/opt/__init__.py b/openllm-python/src/openllm/models/opt/__init__.py index 673c874b..2f818789 100644 --- a/openllm-python/src/openllm/models/opt/__init__.py +++ b/openllm-python/src/openllm/models/opt/__init__.py @@ -42,4 +42,9 @@ else: _import_structure['modeling_tf_opt'] = ['TFOPT'] if t.TYPE_CHECKING: from .modeling_tf_opt import TFOPT as TFOPT -sys.modules[__name__] = LazyModule(__name__, globals()['__file__'], _import_structure) +sys.modules[__name__] = LazyModule(__name__, + globals()['__file__'], + _import_structure, + extra_objects={ + 'DEFAULT_PROMPT_TEMPLATE': DEFAULT_PROMPT_TEMPLATE, 'START_OPT_COMMAND_DOCSTRING': START_OPT_COMMAND_DOCSTRING, 'OPTConfig': OPTConfig, + }) diff --git a/openllm-python/src/openllm/models/stablelm/__init__.py b/openllm-python/src/openllm/models/stablelm/__init__.py index 0f9cc385..fedce01b 100644 --- a/openllm-python/src/openllm/models/stablelm/__init__.py +++ b/openllm-python/src/openllm/models/stablelm/__init__.py @@ -26,4 +26,11 @@ else: _import_structure['modeling_vllm_stablelm'] = ['VLLMStableLM'] if t.TYPE_CHECKING: from .modeling_vllm_stablelm import VLLMStableLM as VLLMStableLM -sys.modules[__name__] = LazyModule(__name__, globals()['__file__'], _import_structure) +sys.modules[__name__] = LazyModule(__name__, + globals()['__file__'], + _import_structure, + extra_objects={ + 'DEFAULT_PROMPT_TEMPLATE': DEFAULT_PROMPT_TEMPLATE, + 'START_STABLELM_COMMAND_DOCSTRING': START_STABLELM_COMMAND_DOCSTRING, + 'StableLMConfig': StableLMConfig, + }) diff --git a/openllm-python/src/openllm/models/starcoder/__init__.py b/openllm-python/src/openllm/models/starcoder/__init__.py index 4a546102..ec067fbe 100644 --- a/openllm-python/src/openllm/models/starcoder/__init__.py +++ b/openllm-python/src/openllm/models/starcoder/__init__.py @@ -26,4 +26,11 @@ else: _import_structure['modeling_vllm_starcoder'] = ['VLLMStarCoder'] if t.TYPE_CHECKING: from .modeling_vllm_starcoder import VLLMStarCoder as VLLMStarCoder -sys.modules[__name__] = LazyModule(__name__, globals()['__file__'], _import_structure) +sys.modules[__name__] = LazyModule(__name__, + globals()['__file__'], + _import_structure, + extra_objects={ + 'DEFAULT_PROMPT_TEMPLATE': DEFAULT_PROMPT_TEMPLATE, + 'START_STARCODER_COMMAND_DOCSTRING': START_STARCODER_COMMAND_DOCSTRING, + 'StarCoderConfig': StarCoderConfig, + })