mirror of
https://github.com/bentoml/OpenLLM.git
synced 2026-02-19 23:24:12 -05:00
Generally correctly format it with ruff format and manual style Signed-off-by: Aaron <29749331+aarnphm@users.noreply.github.com>
32 lines
2.1 KiB
Python
32 lines
2.1 KiB
Python
"""OpenLLM.
|
|
===========
|
|
|
|
An open platform for operating large language models in production.
|
|
Fine-tune, serve, deploy, and monitor any LLMs with ease.
|
|
|
|
* Built-in support for Mistral, Llama 2, Yi, StableLM, Dolly, Flan-T5, Vicuna
|
|
* Option to bring your own fine-tuned LLMs
|
|
* Online Serving with HTTP, gRPC, SSE or custom API
|
|
* Native integration with BentoML, LangChain, OpenAI compatible endpoints, LlamaIndex for custom LLM apps
|
|
"""
|
|
|
|
# update-config-stubs.py: import stubs start
|
|
from openlm_core.config import CONFIG_MAPPING as CONFIG_MAPPING, CONFIG_MAPPING_NAMES as CONFIG_MAPPING_NAMES, AutoConfig as AutoConfig, BaichuanConfig as BaichuanConfig, ChatGLMConfig as ChatGLMConfig, DollyV2Config as DollyV2Config, FalconConfig as FalconConfig, FlanT5Config as FlanT5Config, GPTNeoXConfig as GPTNeoXConfig, LlamaConfig as LlamaConfig, MistralConfig as MistralConfig, MixtralConfig as MixtralConfig, MPTConfig as MPTConfig, OPTConfig as OPTConfig, PhiConfig as PhiConfig, QwenConfig as QwenConfig, StableLMConfig as StableLMConfig, StarCoderConfig as StarCoderConfig, YiConfig as YiConfig
|
|
# update-config-stubs.py: import stubs stop
|
|
|
|
from openllm_cli._sdk import build as build, import_model as import_model, list_models as list_models, start as start
|
|
from openllm_core._configuration import GenerationConfig as GenerationConfig, LLMConfig as LLMConfig, SamplingParams as SamplingParams
|
|
from openllm_core._schemas import GenerationInput as GenerationInput, GenerationOutput as GenerationOutput, MetadataOutput as MetadataOutput
|
|
|
|
from . import bundle as bundle, client as client, exceptions as exceptions, serialisation as serialisation, utils as utils
|
|
from ._deprecated import Runner as Runner
|
|
from ._llm import LLM as LLM
|
|
from ._quantisation import infer_quantisation_config as infer_quantisation_config
|
|
from ._strategies import CascadingResourceStrategy as CascadingResourceStrategy, get_resource as get_resource
|
|
from .client import AsyncHTTPClient as AsyncHTTPClient, HTTPClient as HTTPClient
|
|
from .entrypoints import mount_entrypoints as mount_entrypoints
|
|
from .protocol import openai as openai
|
|
from .serialisation import ggml as ggml, transformers as transformers
|
|
|
|
COMPILED: bool = ...
|