revert: configuration not to dump flatten (#597)

Signed-off-by: Aaron Pham <29749331+aarnphm@users.noreply.github.com>
This commit is contained in:
Aaron Pham
2023-11-09 14:48:23 -05:00
committed by GitHub
parent d60f2fb909
commit 665a41940e

View File

@@ -56,7 +56,7 @@ _Metadata = openllm.MetadataOutput(
model_name=llm_config['model_name'],
backend=llm.__llm_backend__,
model_id=llm.model_id,
configuration=llm_config.model_dump_json(flatten=True).decode(),
configuration=llm_config.model_dump_json().decode(),
prompt_template=llm.runner.prompt_template,
system_message=llm.runner.system_message,
)
@@ -67,6 +67,5 @@ def metadata_v1(_: str) -> openllm.MetadataOutput:
return _Metadata
openllm.mount_entrypoints(
svc, llm
) # HACK: This must always be the last line in this file, as we will do some MK for OpenAPI schema.
# HACK: This must always be the last line in this file, as we will do some MK for OpenAPI schema.
openllm.mount_entrypoints(svc, llm)