Files
OpenLLM/typings/vllm/entrypoints/api_server.pyi
aarnphm-ec2-dev 820b4991fa chore(stubs): add generated for auto-gptq and vllm [skip ci]
This is to help with working on CPU machine

Signed-off-by: aarnphm-ec2-dev <29749331+aarnphm@users.noreply.github.com>
2023-08-03 02:28:24 +00:00

14 lines
499 B
Python
Generated

from _typeshed import Incomplete
from fastapi import Request as Request
from fastapi.responses import Response
from vllm.engine.arg_utils import AsyncEngineArgs as AsyncEngineArgs
from vllm.engine.async_llm_engine import AsyncLLMEngine as AsyncLLMEngine
from vllm.sampling_params import SamplingParams as SamplingParams
from vllm.utils import random_uuid as random_uuid
TIMEOUT_KEEP_ALIVE: int
TIMEOUT_TO_PREVENT_DEADLOCK: int
app: Incomplete
async def generate(request: Request) -> Response: ...