From 665a41940e550a6e72a4bba9d8095c028d9e35b5 Mon Sep 17 00:00:00 2001 From: Aaron Pham <29749331+aarnphm@users.noreply.github.com> Date: Thu, 9 Nov 2023 14:48:23 -0500 Subject: [PATCH] revert: configuration not to dump flatten (#597) Signed-off-by: Aaron Pham <29749331+aarnphm@users.noreply.github.com> --- openllm-python/src/openllm/_service.py | 7 +++---- 1 file changed, 3 insertions(+), 4 deletions(-) diff --git a/openllm-python/src/openllm/_service.py b/openllm-python/src/openllm/_service.py index 11fe5351..37ed9bb0 100644 --- a/openllm-python/src/openllm/_service.py +++ b/openllm-python/src/openllm/_service.py @@ -56,7 +56,7 @@ _Metadata = openllm.MetadataOutput( model_name=llm_config['model_name'], backend=llm.__llm_backend__, model_id=llm.model_id, - configuration=llm_config.model_dump_json(flatten=True).decode(), + configuration=llm_config.model_dump_json().decode(), prompt_template=llm.runner.prompt_template, system_message=llm.runner.system_message, ) @@ -67,6 +67,5 @@ def metadata_v1(_: str) -> openllm.MetadataOutput: return _Metadata -openllm.mount_entrypoints( - svc, llm -) # HACK: This must always be the last line in this file, as we will do some MK for OpenAPI schema. +# HACK: This must always be the last line in this file, as we will do some MK for OpenAPI schema. +openllm.mount_entrypoints(svc, llm)